15 resultados para testing method
em Universidad Politécnica de Madrid
Resumo:
Thin film photovoltaic (TF) modules have gained importance in the photovoltaic (PV) market. New PV plants increasingly use TF technologies. In order to have a reliable sample of a PV module population, a huge number of modules must be measured. There is a big variety of materials used in TF technology. Some of these modules are made of amorphous or microcrystalline silicon. Other are made of CIS or CdTe. Not all these materials respond the same under standard test conditions (STC) of power measurement. Power rates of the modules may vary depending on both the extent and the history of sunlight exposure. Thus, it is necessary a testing method adapted to each TF technology. This test must guarantee repeatability of measurements of generated power. This paper shows responses of different commercial TF PV modules to sunlight exposure. Several test procedures were performed in order to find the best methodology to obtain measurements of TF PV modules at STC in the easiest way. A methodology for indoor measurements adapted to these technologies is described.
Resumo:
Photovoltaic modules based on thin film technology are gaining importance in the photovoltaic market, and module installers and plant owners have increasingly begun to request methods of performing module quality control. These modules pose additional problems for measuring power under standard test conditions (STC), beyond problems caused by the temperature of the module and the ambient variables. The main difficulty is that the modules’ power rates may vary depending both on the amount of time they have been exposed to the sun during recent hours and on their history of sunlight exposure. In order to assess the current state of the module, it is necessary to know its sunlight exposure history. Thus, an easily accomplishable testing method that ensures the repeatability of the measurements of the power generated is needed. This paper examines different tests performed on commercial thin film PV modules of CIS, a-Si and CdTe technologies in order to find the best way to obtain measurements. A method for obtaining indoor measurements of these technologies that takes into account periods of sunlight exposure is proposed. Special attention is paid to CdTe as a fast growing technology in the market.
Resumo:
El acero es, junto con el hormigón, el material más ampliamente empleado en la construcción de obra civil y de edificación. Además de su elevada resistencia, su carácter dúctil resulta un aspecto de particular interés desde el punto de vista de la seguridad estructural, ya que permite redistribuir esfuerzos a elementos adyacentes y, por tanto, almacenar una mayor energía antes del colapso final de la estructura. No obstante, a pesar de su extendida utilización, todavía existen aspectos relacionados con su comportamiento en rotura que necesitan una mayor clarificación y que permitirían un mejor aprovechamiento de sus propiedades. Cuando un elemento de acero es ensayado a tracción y alcanza la carga máxima, sufre la aparición de un cuello de estricción que plantea dificultades para conocer el comportamiento del material desde dicho instante hasta la rotura. La norma ISO 6892-1, que define el método a emplear en un ensayo de tracción con materiales metálicos, establece procedimientos para determinar los parámetros relacionados con este tramo último de la curva F − E. No obstante, la definición de dichos parámetros resulta controvertida, ya que éstos presentan una baja reproducibilidad y una baja repetibilidad que resultan difíciles de explicar. En esta Tesis se busca profundizar en el conocimiento del último tramo de la curva F − E de los aceros de construcción. Para ello se ha realizado una amplia campaña experimental sobre dos aceros representativos en el campo de la construcción civil: el alambrón de partida empleado en la fabricación de alambres de pretensado y un acero empleado como refuerzo en hormigón armado. Los dos materiales analizados presentan formas de rotura diferentes: mientras el primero de ellos presenta una superficie de rotura plana con una región oscura claramente apreciable en su interior, el segundo rompe según la clásica superficie en forma de copa y cono. La rotura en forma de copa y cono ha sido ampliamente estudiada en el pasado y existen modelos de rotura que han logrado reproducirla con éxito, en especial el modelo de Gurson- Tvergaard-Needleman (GTN). En cuanto a la rotura exhibida por el primer material, en principio nada impide abordar su reproducción numérica con un modelo GTN, sin embargo, las diferencias observadas entre ambos materiales en los ensayos experimentales permiten pensar en otro criterio de rotura. En la presente Tesis se realiza una amplia campaña experimental con probetas cilíndricas fabricadas con dos aceros representativos de los empleados en construcción con comportamientos en rotura diferentes. Por un lado se analiza el alambrón de partida empleado en la fabricación de alambres de pretensado, cuyo frente de rotura es plano y perpendicular a la dirección de aplicación de la carga con una región oscura en su interior. Por otro lado, se estudian barras de acero empleadas como armadura pasiva tipo B 500 SD, cuyo frente de rotura presenta la clásica superficie en forma de copa y cono. Estos trabajos experimentales han permitido distinguir dos comportamientos en rotura claramente diferenciados entre ambos materiales y, en el caso del primer material, se ha identificado un comportamiento asemejable al exhibido por materiales frágiles. En este trabajo se plantea la hipótesis de que el primer material, cuya rotura provoca un frente de rotura plano y perpendicular a la dirección de aplicación de la carga, rompe de manera cuasifrágil como consecuencia de un proceso de decohesión, de manera que la región oscura que se observa en el centro del frente de rotura se asemeja a una entalla circular perpendicular a la dirección de aplicación de la carga. Para la reproducción numérica de la rotura exhibida por el primer material, se plantea un criterio de rotura basado en un modelo cohesivo que, como aspecto novedoso, se hace depender de la triaxialidad de tensiones, parámetro determinante en el fallo de este tipo de materiales. Este tipo de modelos presenta varias ventajas respecto a los modelos GTN habitualmente empleados. Mientras los modelos GTN precisan de numerosos parámetros para su calibración, los modelos cohesivos precisan fundamentalmente de dos parámetros para definir su curva de ablandamiento: la tensión de decohesión ft y la energía de fractura GF . Además, los parámetros de los modelos GTN no son medibles de manera experimental, mientras que GF sí lo es. En cuanto a ft, aunque no existe un método para su determinación experimental, sí resulta un parámetro más fácilmente interpretable que los empleados por los modelos GTN, que utilizan valores como el porcentaje de huecos presentes en el material para iniciar el fenómeno de coalescencia o el porcentaje de poros que provoca una pérdida total de la capacidad resistente. Para implementar este criterio de rotura se ha desarrollado un elemento de intercara cohesivo dependiente de la triaxialidad de tensiones. Se han reproducido con éxito los ensayos de tracción llevados a cabo en la campaña experimental empleando dicho elemento de intercara. Además, en estos modelos la rotura se produce fenomenológicamente de la misma manera observada en los ensayos experimentales: produciéndose una decohesión circular en torno al eje de la probeta. En definitiva, los trabajos desarrollados en esta Tesis, tanto experimentales como numéricos, contribuyen a clarificar el comportamiento de los aceros de construcción en el último tramo de la curva F − E y los mecanismos desencadenantes de la rotura final del material, aspecto que puede contribuir a un mejor aprovechamiento de las propiedades de estos aceros en el futuro y a mejorar la seguridad de las estructuras construidas con ellos. Steel is, together with concrete, the most widely used material in civil engineering works. Not only its high strength, but also its ductility is of special interest from the point of view of the structural safety, since it enables stress distribution with adjacent elements and, therefore, more energy can be stored before reaching the structural failure. However, despite of being extensively used, there are still some aspects related to its fracture behaviour that need to be clarified and that will allow for a better use of its properties. When a steel item is tested under tension and reaches the maximum load point, necking process begins, which makes difficult to define the material behaviour from that moment onward. The ISO standard 6892-1, which defines the tensile testing method for metallic materials, describes the procedures to obtain some parameters related to this last section of the F − E curve. Nevertheless, these parameters have proved to be controversial, since they have low reproducibility and repeatibility rates that are difficult to explain. This Thesis tries to deepen the knowledge of the last section of the F − E curve for construction steels. An extensive experimental campaign has been carried out with two representative steels used in civil engineering works: a steel rod used for manufacturing prestressing steel wires, before the cold-drawing process is applied, and steel bars used in reinforced concrete structures. Both materials have different fracture surfaces: while the first of them shows a flat fracture surface, perpendicular to the loading direction with a dark region in the centre of it, the second one shows the classical cup-cone fracture surface. The cup-cone fracture surface has been deeply studied in the past and different numerical models have been able to reproduce it with success, with a special mention to the Gurson-Tvergaard-Needleman model (GTN). Regarding the failure surface shown by the first material, in principle it can be numerically reproduced by a GTN model, but the differences observed between both materials in the experimental campaign suggest thinking of a different failure criterium. In the present Thesis, an extensive experimental campaign has been carried out using cylindrical specimens made of two representative construction steels with different fracture behaviours. On one hand, the initial eutectoid steel rod used for manufacturing prestressing steel wires is analysed, which presents a flat fracture surface, perpendicular to the loading direction, and with a dark region in the centre of it. On the other hand, B 500 SD steel bars, typically used in reinforced concrete structures and with the typical cup-cone fracture surface, are studied. These experimental works have allowed distinguishing two clearly different fracture behaviours between both materials and, in the case of the first one, a fragile-like behaviour has been identified. For the first material, which shows a flat fracture surface perpendicular to the loading direction, the following hypothesis is proposed in this study: a quasi-brittle fracture is developed as a consequence of a decohesion process, with the dark region acting as a circular crack perpendicular to the loading direction. To reproduce numerically the fracture behaviour shown by the first material, a failure criterium based on a cohesive model is proposed in this Thesis. As an innovative contribution, this failure criterium depends on the stress triaxiality state of the material, which is a key parameter when studying fracture in this kind of materials. This type of models have some advantages when compared to the widely used GTN models. While GTN models need a high number of parameters to be defined, cohesive models need basically two parameters to define the softening curve: the decohesion stress ft and the fracture energy GF . In addition to this, GTN models parameters cannot be measured experimentally, while GF is indeed. Regarding ft, although no experimental procedure is defined for its obtention, it has an easier interpretation than the parameters used by the GTN models like, for instance, the void volume needed for the coalescence process to start or the void volume that leads to a total loss of the bearing capacity. In order to implement this failure criterium, a triaxiality-dependent cohesive interface element has been developed. The experimental results obtained in the experimental campaign have been successfully reproduced by using this interface element. Furthermore, in these models the failure mechanism is developed in the same way as observed experimentally: with a circular decohesive process taking place around the longitudinal axis of the specimen. In summary, the works developed in this Thesis, both experimental and numerical, contribute to clarify the behaviour of construction steels in the last section of the F − E curve and the mechanisms responsible for the eventual material failure, an aspect that can lead to a better use of the properties of these steels in the future and a safety improvement in the structures built with them.
Resumo:
La artroplastia de cadera se considera uno de los mayores avances quirúrgicos de la Medicina. La aplicación de esta técnica de Traumatología se ha incrementado notablemente en los últimos anos, a causa principalmente del progresivo incremento de la esperanza de vida. En efecto, con la edad aumentan los problemas de artrosis y osteoporosis, enfermedades típicas de las articulaciones y de los huesos que requieren en muchos casos la sustitución protésica total o parcial de la articulación. El buen comportamiento funcional de una prótesis depende en gran medida de la estabilidad primaria, es decir, el correcto anclaje de la prótesis en el momento de su implantación. Las prótesis no cementadas basan su éxito a largo plazo en la osteointegración que tiene lugar entre el material protésico y el tejido óseo, y para lograrla es imprescindible conseguir unas buenas condiciones de estabilidad primaria. El aflojamiento aséptico es la principal causa de fallo de artroplastia total de cadera. Este es un fenómeno en el que, debido a complejas interacciones de factores mecánicos y biológicos, se producen movimientos relativos que comprometen la funcionalidad del implante. La minimización de los correspondientes danos depende en gran medida de la detección precoz del aflojamiento. Para lograr la detección temprana del aflojamiento aséptico del vástago femoral se han ensayado diferentes técnicas, tanto in vivo como in vitro: análisis numéricos y técnicas experimentales basadas en sensores de movimientos provocados por cargas transmitidas natural o artificialmente, tales como impactos o vibraciones de distintas frecuencias. Los montajes y procedimientos aplicados son heterogéneos y, en muchas ocasiones, complejos y costosos, no existiendo acuerdo sobre una técnica simple y eficaz de aplicación general. Asimismo, en la normativa vigente que regula las condiciones que debe cumplir una prótesis previamente a su comercialización, no hay ningún apartado referido específicamente a la evaluación de la bondad del diseño del vástago femoral con respecto a la estabilidad primaria. El objetivo de esta tesis es desarrollar una metodología para el análisis, in vitro, de la estabilidad de un vástago femoral implantado, a fin de poder evaluar las técnicas de implantación y los diferentes diseños de prótesis previamente a su oferta en el mercado. Además se plantea como requisito fundamental que el método desarrollado sea sencillo, reversible, repetible, no destructivo, con control riguroso de parámetros (condiciones de contorno de cargas y desplazamientos) y con un sistema de registro e interpretación de resultados rápido, fiable y asequible. Como paso previo, se ha realizado un análisis cualitativo del problema de contacto en la interfaz hueso-vástago aplicando una técnica optomecánica del campo continuo (fotoelasticidad). Para ello se han fabricado tres modelos en 2D del conjunto hueso-vástago, simulando tres tipos de contactos en la interfaz: contacto sin adherencia y con holgura, contacto sin adherencia y sin holgura, y contacto con adherencia y homogéneo. Aplicando la misma carga a cada modelo, y empleando la técnica de congelación de tensiones, se han visualizado los correspondientes estados tensionales, siendo estos más severos en el modelo de unión sin adherencia, como cabía esperar. En todo caso, los resultados son ilustrativos de la complejidad del problema de contacto y confirman la conveniencia y necesidad de la vía experimental para el estudio del problema. Seguidamente se ha planteado un ensayo dinámico de oscilaciones libres con instrumentación de sensores resistivos tipo galga extensométrica. Las muestras de ensayo han sido huesos fémur en todas sus posibles variantes: modelos simplificados, hueso sintético normalizado y hueso de cadáver, seco y fresco. Se ha diseñado un sistema de empotramiento del extremo distal de la muestra (fémur) con control riguroso de las condiciones de anclaje. La oscilación libre de la muestra se ha obtenido mediante la liberación instantánea de una carga estética determinada y aplicada previamente, bien con una maquina de ensayo o bien por gravedad. Cada muestra se ha instrumentado con galgas extensométricas convencionales cuya señal se ha registrado con un equipo dinámico comercial. Se ha aplicado un procedimiento de tratamiento de señal para acotar, filtrar y presentar las respuestas de los sensores en el dominio del tiempo y de la frecuencia. La interpretación de resultados es de tipo comparativo: se aplica el ensayo a una muestra de fémur Intacto que se toma de referencia, y a continuación se repite el ensayo sobre la misma muestra con una prótesis implantada; la comparación de resultados permite establecer conclusiones inmediatas sobre los efectos de la implantación de la prótesis. La implantación ha sido realizada por un cirujano traumatólogo utilizando las mismas técnicas e instrumental empleadas en el quirófano durante la práctica clínica real, y se ha trabajado con tres vástagos femorales comerciales. Con los resultados en el dominio del tiempo y de la frecuencia de las distintas aplicaciones se han establecido conclusiones sobre los siguientes aspectos: Viabilidad de los distintos tipos de muestras sintéticas: modelos simplificados y fémur sintético normalizado. Repetibilidad, linealidad y reversibilidad del ensayo. Congruencia de resultados con los valores teóricos deducidos de la teoría de oscilaciones libres de barras. Efectos de la implantación de tallos femorales en la amplitud de las oscilaciones, amortiguamiento y frecuencias de oscilación. Detección de armónicos asociados a la micromovilidad. La metodología se ha demostrado apta para ser incorporada a la normativa de prótesis, es de aplicación universal y abre vías para el análisis de la detección y caracterización de la micromovilidad de una prótesis frente a las cargas de servicio. ABSTRACT Total hip arthroplasty is considered as one of the greatest surgical advances in medicine. The application of this technique on Traumatology has increased significantly in recent years, mainly due to the progressive increase in life expectancy. In fact, advanced age increases osteoarthritis and osteoporosis problems, which are typical diseases of joints and bones, and in many cases require full or partial prosthetic replacement on the joint. Right functional behavior of prosthesis is highly dependent on the primary stability; this means it depends on the correct anchoring of the prosthesis at the time of implantation. Uncemented prosthesis base their long-term success on the quality of osseointegration that takes place between the prosthetic material and bone tissue, and to achieve this good primary stability conditions is mandatory. Aseptic loosening is the main cause of failure in total hip arthroplasty. This is a phenomenon in which relative movements occur, due to complex interactions of mechanical and biological factors, and these micromovements put the implant functionality at risk. To minimize possible damage, it greatly depends on the early detection of loosening. For this purpose, various techniques have been tested both in vivo and in vitro: numerical analysis and experimental techniques based on sensors for movements caused by naturally or artificially transmitted loads, such as impacts or vibrations at different frequencies. The assemblies and methods applied are heterogeneous and, in many cases, they are complex and expensive, with no agreement on the use of a simple and effective technique for general purposes. Likewise, in current regulations for governing the conditions to be fulfilled by the prosthesis before going to market, there is no specific section related to the evaluation of the femoral stem design in relation to primary stability. The aim of this thesis is to develop a in vitro methodology for analyzing the stability of an implanted femoral stem, in order to assess the implantation techniques and the different prosthesis designs prior to its offer in the market. We also propose as a fundamental requirement that the developed testing method should be simple, reversible, repeatable, non-destructive, with close monitoring of parameters (boundary conditions of loads and displacements) and with the availability of a register system to record and interpret results in a fast, reliable and affordable manner. As a preliminary step, we have performed a qualitative analysis of the contact problems in the bone-stem interface, through the application of a continuous field optomechanical technique (photoelasticity). For this proposal three 2D models of bone–stem set, has been built simulating three interface contact types: loosened an unbounded contact, unbounded and fixed contact, and bounded homogeneous contact. By means of applying the same load to each model, and using the stress freezing technique, it has displayed the corresponding stress states, being more severe as expected, in the unbounded union model. In any case, the results clearly show the complexity of the interface contact problem, and they confirm the need for experimental studies about this problem. Afterward a free oscillation dynamic test has been done using resistive strain gauge sensors. Test samples have been femur bones in all possible variants: simplified models, standardized synthetic bone, and dry and cool cadaveric bones. An embedding system at the distal end of the sample with strong control of the anchoring conditions has been designed. The free oscillation of the sample has been obtained by the instantaneous release of a static load, which was previously determined and applied to the sample through a testing machine or using the gravity force. Each sample was equipped with conventional strain gauges whose signal is registered with a marketed dynamic equipment. Then, it has applied a signal processing procedure to delimit, filter and present the time and frequency response signals from the sensors. Results are interpreted by comparing different trials: the test is applied to an intact femur sample which is taken as a reference, and then this test is repeated over the same sample with an implanted prosthesis. From comparison between results, immediate conclusions about the effects of the implantation of the prosthesis can be obtained. It must be said that the implementation has been made by an expert orthopedic surgeon using the same techniques and instruments as those used in clinical surgery. He has worked with three commercial femoral stems. From the results obtained in the time and frequency domains for the different applications the following conclusions have been established: Feasibility of the different types of synthetic samples: simplified models and standardized synthetic femur. Repeatability, linearity and reversibility of the testing method. Consistency of results with theoretical values deduced from the bars free oscillations theory. Effects of introduction of femoral stems in the amplitude, damping and frequencies of oscillations Detection of micromobility associated harmonics. This methodology has been proved suitable to be included in the standardization process of arthroplasty prosthesis, it is universally applicable and it allows establishing new methods for the analysis, detection and characterization of prosthesis micromobility due to functional loads.
Resumo:
Actualmente la optimization de la calidad de experiencia (Quality of Experience- QoE) de HTTP Adaptive Streaming (HAS) de video recibe una atención creciente. Este incremento de interés proviene fundamentalmente de las carencias de las soluciones actuales HAS, que, al no ser QoE-driven, no incluyen la percepción de la calidad de los usuarios finales como una parte integral de la lógica de adaptación. Por lo tanto, la obtención de información de referencia fiable en QoE en HAS presenta retos importantes, ya que las metodologías de evaluación subjetiva de la calidad de vídeo propuestas en las normas actuales no son adecuadas para tratar con la variación temporal de la calidad que es consustancial de HAS. Esta tesis investiga la influencia de la adaptación dinámica en la calidad de la transmisión de vídeo considerando métodos de evaluación subjetiva. Tras un estudio exhaustivo del estado del arte en la evaluación subjetiva de QoE en HAS, se han resaltado los retos asociados y las líneas de investigación abiertas. Como resultado, se han seleccionado dos líneas principales de investigación: el análisis del impacto en la QoE de los parámetros de las técnicas de adaptación y la investigación de las metodologías de prueba subjetiva adecuada para evaluación de QoE en HAS. Se han llevado a cabo un conjunto de experimentos de laboratorio para investigar las cuestiones planteadas mediante la utilización de diferentes metodologáas para pruebas subjetivas. El análisis estadístico muestra que no son robustas todas las suposiciones y reivindicaciones de las referencias analizadas, en particular en lo que respecta al impacto en la QoE de la frecuencia de las variaciones de calidad, de las adaptaciones suaves o abruptas y de las oscilaciones de calidad. Por otra parte, nuestros resultados confirman la influencia de otros parámetros, como la longitud de los segmentos de vídeo y la amplitud de las oscilaciones de calidad. Los resultados también muestran que tomar en consideración las características objetivas de los contenidos puede ser beneficioso para la mejora de la QoE en HAS. Además, todos los resultados han sido validados mediante extensos análisis experimentales que han incluido estudio tanto en otros laboratorios como en crowdsourcing Por último, sobre los aspectos metodológicos de las pruebas subjetivas de QoE, se ha realizado la comparación entre los resultados experimentales obtenidos a partir de un método estandarizado basado en estímulos cortos (ACR) y un método semi continuo (desarrollado para la evaluación de secuencias prolongadas de vídeo). A pesar de algunas diferencias, el resultado de los análisis estadísticos no muestra ningún efecto significativo de la metodología de prueba. Asimismo, aunque se percibe la influencia de la presencia de audio en la evaluación de degradaciones del vídeo, no se han encontrado efectos estadísticamente significativos de dicha presencia. A partir de la ausencia de influencia del método de prueba y de la presencia de audio, se ha realizado un análisis adicional sobre el impacto de realizar comparaciones estadísticas múltiples en niveles estadísticos de importancia que aumentan la probabilidad de los errores de tipo-I (falsos positivos). Nuestros resultados muestran que, para obtener un efectos sólido en el análisis estadístico de los resultados subjetivos, es necesario aumentar el número de sujetos de las pruebas claramente por encima de los tamaños de muestras propuestos por las normas y recomendaciones actuales. ABSTRACT Optimizing the Quality of Experience (QoE) of HTTP adaptive video streaming (HAS) is receiving increasing attention nowadays. The growth of interest is mainly caused by the fact that current HAS solutions are not QoE-driven, i.e. end-user quality perception is not integral part of the adaptation logic. However, obtaining the necessary reliable ground truths on HAS QoE faces substantial challenges, since the subjective video quality assessment methodologies as proposed by current standards are not well-suited for dealing with the time-varying quality properties that are characteristic for HAS. This thesis investigates the influence of dynamic quality adaptation on the QoE of streaming video by means of subjective evaluation approaches. Based on a comprehensive survey of related work on subjective HAS QoE assessment, the related challenges and open research questions are highlighted and discussed. As a result, two main research directions are selected for further investigation: analysis of the QoE impact of different technical adaptation parameters, and investigation of testing methodologies suitable for HAS QoE evaluation. In order to investigate related research issues and questions, a set of laboratory experiments have been conducted using different subjective testing methodologies. Our statistical analysis demonstrates that not all assumptions and claims reported in the literature are robust, particularly as regards the QoE impact of switching frequency, smooth vs. abrupt switching, and quality oscillation. On the other hand, our results confirm the influence of some other parameters such as chunk length and switching amplitude on perceived quality. We also show that taking the objective characteristics of the content into account can be beneficial to improve the adaptation viewing experience. In addition, all aforementioned findings are validated by means of an extensive cross-experimental analysis that involves external laboratory and crowdsourcing studies. Finally, to address the methodological aspects of subjective QoE testing, a comparison between the experimental results obtained from a (short stimuli-based) ACR standardized method and a semi-continuous method (developed for assessment of long video sequences) has been performed. In spite of observation of some differences, the result of statistical analysis does not show any significant effect of testing methodology. Similarly, although the influence of audio presence on evaluation of video-related degradations is perceived, no statistically significant effect of audio presence could be found. Motivating by this finding (no effect of testing method and audio presence), a subsequent analysis has been performed investigating the impact of performing multiple statistical comparisons on statistical levels of significance which increase the likelihood of Type-I errors (false positives). Our results show that in order to obtain a strong effect from the statistical analysis of the subjective results, it is necessary to increase the number of test subjects well beyond the sample sizes proposed by current quality assessment standards and recommendations.
Resumo:
The usage of HTTP adaptive streaming (HAS) has become widely spread in multimedia services. Because it allows the service providers to improve the network resource utilization and user׳s Quality of Experience (QoE). Using this technology, the video playback interruption is reduced since the network and server status in addition to capability of user device, all are taken into account by HAS client to adapt the quality to the current condition. Adaptation can be done using different strategies. In order to provide optimal QoE, the perceptual impact of adaptation strategies from point of view of the user should be studied. However, the time-varying video quality due to the adaptation which usually takes place in a long interval introduces a new type of impairment making the subjective evaluation of adaptive streaming system challenging. The contribution of this paper is two-fold: first, it investigates the testing methodology to evaluate HAS QoE by comparing the subjective experimental outcomes obtained from ACR standardized method and a semi-continuous method developed to evaluate the long sequences. In addition, influence of using audiovisual stimuli to evaluate the video-related impairment is inquired. Second, impact of some of the adaptation technical factors including the quality switching amplitude and chunk size in combination with high range of commercial content type is investigated. The results of this study provide a good insight toward achieving appropriate testing method to evaluate HAS QoE, in addition to designing switching strategies with optimal visual quality.
Resumo:
Belief propagation (BP) is a technique for distributed inference in wireless networks and is often used even when the underlying graphical model contains cycles. In this paper, we propose a uniformly reweighted BP scheme that reduces the impact of cycles by weighting messages by a constant ?edge appearance probability? rho ? 1. We apply this algorithm to distributed binary hypothesis testing problems (e.g., distributed detection) in wireless networks with Markov random field models. We demonstrate that in the considered setting the proposed method outperforms standard BP, while maintaining similar complexity. We then show that the optimal ? can be approximated as a simple function of the average node degree, and can hence be computed in a distributed fashion through a consensus algorithm.
Resumo:
Software testing is a key aspect of software reliability and quality assurance in a context where software development constantly has to overcome mammoth challenges in a continuously changing environment. One of the characteristics of software testing is that it has a large intellectual capital component and can thus benefit from the use of the experience gained from past projects. Software testing can, then, potentially benefit from solutions provided by the knowledge management discipline. There are in fact a number of proposals concerning effective knowledge management related to several software engineering processes. Objective: We defend the use of a lesson learned system for software testing. The reason is that such a system is an effective knowledge management resource enabling testers and managers to take advantage of the experience locked away in the brains of the testers. To do this, the experience has to be gathered, disseminated and reused. Method: After analyzing the proposals for managing software testing experience, significant weaknesses have been detected in the current systems of this type. The architectural model proposed here for lesson learned systems is designed to try to avoid these weaknesses. This model (i) defines the structure of the software testing lessons learned; (ii) sets up procedures for lesson learned management; and (iii) supports the design of software tools to manage the lessons learned. Results: A different approach, based on the management of the lessons learned that software testing engineers gather from everyday experience, with two basic goals: usefulness and applicability. Conclusion: The architectural model proposed here lays the groundwork to overcome the obstacles to sharing and reusing experience gained in the software testing and test management. As such, it provides guidance for developing software testing lesson learned systems.
Resumo:
The Pseudo-Dynamic Test Method (PDTM) is being developped currently as an alternative to the shaking table testing of large size models. However, the stepped slow execution of the former type of test has been found to be the source of important errors arising from the stress relaxation. A new continuous test method, wich allows the selection of a suitable time-scale factor in the response in order to control these errors, es proposed here. Such scaled-time response is theoretically obtained by simply augmenting the mass of the structure for wich some practical solutions are proposed.
Resumo:
One of the most significant aspects of a building?s acoustic behavior is the airborne sound insulation of the room façades, since this determines the protection of its inhabitants against environmental noise. For this reason, authorities in most countries have established in their acoustic regulations for buildings the minimum value of sound insulation that must be respected for façades. In order to verify compliance with legal requirements it is usual to perform acoustic measurements in the finished buildings and then compare the measurement results with the established limits. Since there is always a certain measurement uncertainty, this uncertainty must be calculated and taken into account in order to ensure compliance with specifications. The most commonly used method for measuring sound insulation on façades is the so-called Global Loudspeaker Method, specified in ISO 140-5:1998. This method uses a loudspeaker placed outside the building as a sound source. The loudspeaker directivity has a significant influence on the measurement results, and these results may change noticeably by choosing different loudspeakers, even though they all fulfill the directivity requirements of ISO 140-5. This work analyzes the influence of the loudspeaker directivity on the results of façade sound insulation measurement, and determines its contribution to measurement uncertainty. The theoretical analysis is experimentally validated by means of an intermediate precision test according to ISO 5725-3:1994, which compares the values of sound insulation obtained for a façade using various loudspeakers with different directivities. Keywords: Uncertainty, Façade, Insulation
Resumo:
Europe needs to restructure its energy system. The aim to decrease the reliance on fossil fuels to a higher dependence on renewable energy has now been imposed by The European Commission. In order to achieve this goal there is a great interest in Norway to become "The Green Battery of Europe". In the pursuit of this goal a GIS-tool was created to investigate the pump storage potential in Norway. The tool searches for possible connections between existing reservoirs and dams with the criteria selected by the user. The aim of this thesis was to test the tool and see if the results suggested were plausible, develop a cost calculation method for the PSH lines, and make suggestions for further development of the tool. During the process the tool presented many non-feasible pumped storage hydropower (PSH) connections. The area of Telemark was chosen for the more detailed study. The results were discussed and some improvements were suggested for further development of the tool. Also a sensitivity test was done to see which of the parameters set by the user are the most relevant for the PSH connection suggestion. From a range of the most promising PSH plants suggested by the tool, the one between Songavatn and Totak was chosen for a case study, where there already exists a power plant between both reservoirs. A new Pumped Storage Plant was designed with a power production of 1200 MW. There are still many topics open to discussion, such as how to deal with environmental restrictions, or how to deal with inflows and outflows of the reservoirs from the existing power plants. Consequently the GIS-tool can be a very useful tool to establish the best possible connections between existing reservoirs and dams, but it still needs a deep study and the creation of new parameters for the user.
Resumo:
At present, photovoltaic energy is one of the most important renewable energy sources. The demand for solar panels has been continuously growing, both in the industrial electric sector and in the private sector. In both cases the analysis of the solar panel efficiency is extremely important in order to maximize the energy production. In order to have a more efficient photovoltaic system, the most accurate understanding of this system is required. However, in most of the cases the only information available in this matter is reduced, the experimental testing of the photovoltaic device being out of consideration, normally for budget reasons. Several methods, normally based on an equivalent circuit model, have been developed to extract the I-V curve of a photovoltaic device from the small amount of data provided by the manufacturer. The aim of this paper is to present a fast, easy, and accurate analytical method, developed to calculate the equivalent circuit parameters of a solar panel from the only data that manufacturers usually provide. The calculated circuit accurately reproduces the solar panel behavior, that is, the I-V curve. This fact being extremely important for practical reasons such as selecting the best solar panel in the market for a particular purpose, or maximize the energy extraction with MPPT (Maximum Peak Power Tracking) methods.
Resumo:
The concept of service oriented architecture has been extensively explored in software engineering, due to the fact that it produces architectures made up of several interconnected modules, easy to reuse when building new systems. This approach to design would be impossible without interconnection mechanisms such as REST (Representationa State Transfer) services, which allow module communication while minimizing coupling. . However, this low coupling brings disadvantages, such as the lack of transparency, which makes it difficult to sistematically create tests without knowledge of the inner working of a system. In this article, we present an automatic error detection system for REST services, based on a statistical analysis over responses produced at multiple service invocations. Thus, a service can be systematically tested without knowing its full specification. The method can find errors in REST services which could not be identified by means of traditional testing methods, and provides limited testing coverage for services whose response format is unknown. It can be also useful as a complement to other testing mechanisms.
Resumo:
The cup anemometer rotor aerodynamics is analytically studied based on the aerodynamics of a single cup. The effect of the rotation on the aerodynamic force is included in the analytical model, together with the displacement of the aerodynamic center during one turn of the cup. The model can be fitted to the testing results, indicating the presence of both the aforementioned effects
Resumo:
Hoy en día, el proceso de un proyecto sostenible persigue realizar edificios de elevadas prestaciones que son, energéticamente eficientes, saludables y económicamente viables utilizando sabiamente recursos renovables para minimizar el impacto sobre el medio ambiente reduciendo, en lo posible, la demanda de energía, lo que se ha convertido, en la última década, en una prioridad. La Directiva 2002/91/CE "Eficiencia Energética de los Edificios" (y actualizaciones posteriores) ha establecido el marco regulatorio general para el cálculo de los requerimientos energéticos mínimos. Desde esa fecha, el objetivo de cumplir con las nuevas directivas y protocolos ha conducido las políticas energéticas de los distintos países en la misma dirección, centrándose en la necesidad de aumentar la eficiencia energética en los edificios, la adopción de medidas para reducir el consumo, y el fomento de la generación de energía a través de fuentes renovables. Los edificios de energía nula o casi nula (ZEB, Zero Energy Buildings ó NZEB, Net Zero Energy Buildings) deberán convertirse en un estándar de la construcción en Europa y con el fin de equilibrar el consumo de energía, además de reducirlo al mínimo, los edificios necesariamente deberán ser autoproductores de energía. Por esta razón, la envolvente del edifico y en particular las fachadas son importantes para el logro de estos objetivos y la tecnología fotovoltaica puede tener un papel preponderante en este reto. Para promover el uso de la tecnología fotovoltaica, diferentes programas de investigación internacionales fomentan y apoyan soluciones para favorecer la integración completa de éstos sistemas como elementos arquitectónicos y constructivos, los sistemas BIPV (Building Integrated Photovoltaic), sobre todo considerando el próximo futuro hacia edificios NZEB. Se ha constatado en este estudio que todavía hay una falta de información útil disponible sobre los sistemas BIPV, a pesar de que el mercado ofrece una interesante gama de soluciones, en algunos aspectos comparables a los sistemas tradicionales de construcción. Pero por el momento, la falta estandarización y de una regulación armonizada, además de la falta de información en las hojas de datos técnicos (todavía no comparables con las mismas que están disponibles para los materiales de construcción), hacen difícil evaluar adecuadamente la conveniencia y factibilidad de utilizar los componentes BIPV como parte integrante de la envolvente del edificio. Organizaciones internacionales están trabajando para establecer las normas adecuadas y procedimientos de prueba y ensayo para comprobar la seguridad, viabilidad y fiabilidad estos sistemas. Sin embargo, hoy en día, no hay reglas específicas para la evaluación y caracterización completa de un componente fotovoltaico de integración arquitectónica de acuerdo con el Reglamento Europeo de Productos de la Construcción, CPR 305/2011. Los productos BIPV, como elementos de construcción, deben cumplir con diferentes aspectos prácticos como resistencia mecánica y la estabilidad; integridad estructural; seguridad de utilización; protección contra el clima (lluvia, nieve, viento, granizo), el fuego y el ruido, aspectos que se han convertido en requisitos esenciales, en la perspectiva de obtener productos ambientalmente sostenibles, saludables, eficientes energéticamente y económicamente asequibles. Por lo tanto, el módulo / sistema BIPV se convierte en una parte multifuncional del edificio no sólo para ser física y técnicamente "integrado", además de ser una oportunidad innovadora del diseño. Las normas IEC, de uso común en Europa para certificar módulos fotovoltaicos -IEC 61215 e IEC 61646 cualificación de diseño y homologación del tipo para módulos fotovoltaicos de uso terrestre, respectivamente para módulos fotovoltaicos de silicio cristalino y de lámina delgada- atestan únicamente la potencia del módulo fotovoltaico y dan fe de su fiabilidad por un período de tiempo definido, certificando una disminución de potencia dentro de unos límites. Existe también un estándar, en parte en desarrollo, el IEC 61853 (“Ensayos de rendimiento de módulos fotovoltaicos y evaluación energética") cuyo objetivo es la búsqueda de procedimientos y metodologías de prueba apropiados para calcular el rendimiento energético de los módulos fotovoltaicos en diferentes condiciones climáticas. Sin embargo, no existen ensayos normalizados en las condiciones específicas de la instalación (p. ej. sistemas BIPV de fachada). Eso significa que es imposible conocer las efectivas prestaciones de estos sistemas y las condiciones ambientales que se generan en el interior del edificio. La potencia nominal de pico Wp, de un módulo fotovoltaico identifica la máxima potencia eléctrica que éste puede generar bajo condiciones estándares de medida (STC: irradición 1000 W/m2, 25 °C de temperatura del módulo y distribución espectral, AM 1,5) caracterizando eléctricamente el módulo PV en condiciones específicas con el fin de poder comparar los diferentes módulos y tecnologías. El vatio pico (Wp por su abreviatura en inglés) es la medida de la potencia nominal del módulo PV y no es suficiente para evaluar el comportamiento y producción del panel en términos de vatios hora en las diferentes condiciones de operación, y tampoco permite predecir con convicción la eficiencia y el comportamiento energético de un determinado módulo en condiciones ambientales y de instalación reales. Un adecuado elemento de integración arquitectónica de fachada, por ejemplo, debería tener en cuenta propiedades térmicas y de aislamiento, factores como la transparencia para permitir ganancias solares o un buen control solar si es necesario, aspectos vinculados y dependientes en gran medida de las condiciones climáticas y del nivel de confort requerido en el edificio, lo que implica una necesidad de adaptación a cada contexto específico para obtener el mejor resultado. Sin embargo, la influencia en condiciones reales de operación de las diferentes soluciones fotovoltaicas de integración, en el consumo de energía del edificio no es fácil de evaluar. Los aspectos térmicos del interior del ambiente o de iluminación, al utilizar módulos BIPV semitransparentes por ejemplo, son aún desconocidos. Como se dijo antes, la utilización de componentes de integración arquitectónica fotovoltaicos y el uso de energía renovable ya es un hecho para producir energía limpia, pero también sería importante conocer su posible contribución para mejorar el confort y la salud de los ocupantes del edificio. Aspectos como el confort, la protección o transmisión de luz natural, el aislamiento térmico, el consumo energético o la generación de energía son aspectos que suelen considerarse independientemente, mientras que todos juntos contribuyen, sin embargo, al balance energético global del edificio. Además, la necesidad de dar prioridad a una orientación determinada del edificio, para alcanzar el mayor beneficio de la producción de energía eléctrica o térmica, en el caso de sistemas activos y pasivos, respectivamente, podría hacer estos últimos incompatibles, pero no necesariamente. Se necesita un enfoque holístico que permita arquitectos e ingenieros implementar sistemas tecnológicos que trabajen en sinergia. Se ha planteado por ello un nuevo concepto: "C-BIPV, elemento fotovoltaico consciente integrado", esto significa necesariamente conocer los efectos positivos o negativos (en términos de confort y de energía) en condiciones reales de funcionamiento e instalación. Propósito de la tesis, método y resultados Los sistemas fotovoltaicos integrados en fachada son a menudo soluciones de vidrio fácilmente integrables, ya que por lo general están hechos a medida. Estos componentes BIPV semitransparentes, integrados en el cerramiento proporcionan iluminación natural y también sombra, lo que evita el sobrecalentamiento en los momentos de excesivo calor, aunque como componente estático, asimismo evitan las posibles contribuciones pasivas de ganancias solares en los meses fríos. Además, la temperatura del módulo varía considerablemente en ciertas circunstancias influenciada por la tecnología fotovoltaica instalada, la radiación solar, el sistema de montaje, la tipología de instalación, falta de ventilación, etc. Este factor, puede suponer un aumento adicional de la carga térmica en el edificio, altamente variable y difícil de cuantificar. Se necesitan, en relación con esto, más conocimientos sobre el confort ambiental interior en los edificios que utilizan tecnologías fotovoltaicas integradas, para abrir de ese modo, una nueva perspectiva de la investigación. Con este fin, se ha diseñado, proyectado y construido una instalación de pruebas al aire libre, el BIPV Env-lab "BIPV Test Laboratory", para la caracterización integral de los diferentes módulos semitransparentes BIPV. Se han definido también el método y el protocolo de ensayos de caracterización en el contexto de un edificio y en condiciones climáticas y de funcionamiento reales. Esto ha sido posible una vez evaluado el estado de la técnica y la investigación, los aspectos que influyen en la integración arquitectónica y los diferentes tipos de integración, después de haber examinado los métodos de ensayo para los componentes de construcción y fotovoltaicos, en condiciones de operación utilizadas hasta ahora. El laboratorio de pruebas experimentales, que consiste en dos habitaciones idénticas a escala real, 1:1, ha sido equipado con sensores y todos los sistemas de monitorización gracias a los cuales es posible obtener datos fiables para evaluar las prestaciones térmicas, de iluminación y el rendimiento eléctrico de los módulos fotovoltaicos. Este laboratorio permite el estudio de tres diferentes aspectos que influencian el confort y consumo de energía del edificio: el confort térmico, lumínico, y el rendimiento energético global (demanda/producción de energía) de los módulos BIPV. Conociendo el balance de energía para cada tecnología solar fotovoltaica experimentada, es posible determinar cuál funciona mejor en cada caso específico. Se ha propuesto una metodología teórica para la evaluación de estos parámetros, definidos en esta tesis como índices o indicadores que consideran cuestiones relacionados con el bienestar, la energía y el rendimiento energético global de los componentes BIPV. Esta metodología considera y tiene en cuenta las normas reglamentarias y estándares existentes para cada aspecto, relacionándolos entre sí. Diferentes módulos BIPV de doble vidrio aislante, semitransparentes, representativos de diferentes tecnologías fotovoltaicas (tecnología de silicio monocristalino, m-Si; de capa fina en silicio amorfo unión simple, a-Si y de capa fina en diseleniuro de cobre e indio, CIS) fueron seleccionados para llevar a cabo una serie de pruebas experimentales al objeto de demostrar la validez del método de caracterización propuesto. Como resultado final, se ha desarrollado y generado el Diagrama Caracterización Integral DCI, un sistema gráfico y visual para representar los resultados y gestionar la información, una herramienta operativa útil para la toma de decisiones con respecto a las instalaciones fotovoltaicas. Este diagrama muestra todos los conceptos y parámetros estudiados en relación con los demás y ofrece visualmente toda la información cualitativa y cuantitativa sobre la eficiencia energética de los componentes BIPV, por caracterizarlos de manera integral. ABSTRACT A sustainable design process today is intended to produce high-performance buildings that are energy-efficient, healthy and economically feasible, by wisely using renewable resources to minimize the impact on the environment and to reduce, as much as possible, the energy demand. In the last decade, the reduction of energy needs in buildings has become a top priority. The Directive 2002/91/EC “Energy Performance of Buildings” (and its subsequent updates) established a general regulatory framework’s methodology for calculation of minimum energy requirements. Since then, the aim of fulfilling new directives and protocols has led the energy policies in several countries in a similar direction that is, focusing on the need of increasing energy efficiency in buildings, taking measures to reduce energy consumption, and fostering the use of renewable sources. Zero Energy Buildings or Net Zero Energy Buildings will become a standard in the European building industry and in order to balance energy consumption, buildings, in addition to reduce the end-use consumption should necessarily become selfenergy producers. For this reason, the façade system plays an important role for achieving these energy and environmental goals and Photovoltaic can play a leading role in this challenge. To promote the use of photovoltaic technology in buildings, international research programs encourage and support solutions, which favors the complete integration of photovoltaic devices as an architectural element, the so-called BIPV (Building Integrated Photovoltaic), furthermore facing to next future towards net-zero energy buildings. Therefore, the BIPV module/system becomes a multifunctional building layer, not only physically and functionally “integrated” in the building, but also used as an innovative chance for the building envelope design. It has been found in this study that there is still a lack of useful information about BIPV for architects and designers even though the market is providing more and more interesting solutions, sometimes comparable to the existing traditional building systems. However at the moment, the lack of an harmonized regulation and standardization besides to the non-accuracy in the technical BIPV datasheets (not yet comparable with the same ones available for building materials), makes difficult for a designer to properly evaluate the fesibility of this BIPV components when used as a technological system of the building skin. International organizations are working to establish the most suitable standards and test procedures to check the safety, feasibility and reliability of BIPV systems. Anyway, nowadays, there are no specific rules for a complete characterization and evaluation of a BIPV component according to the European Construction Product Regulation, CPR 305/2011. BIPV products, as building components, must comply with different practical aspects such as mechanical resistance and stability; structural integrity; safety in use; protection against weather (rain, snow, wind, hail); fire and noise: aspects that have become essential requirements in the perspective of more and more environmentally sustainable, healthy, energy efficient and economically affordable products. IEC standards, commonly used in Europe to certify PV modules (IEC 61215 and IEC 61646 respectively crystalline and thin-film ‘Terrestrial PV Modules-Design Qualification and Type Approval’), attest the feasibility and reliability of PV modules for a defined period of time with a limited power decrease. There is also a standard (IEC 61853, ‘Performance Testing and Energy Rating of Terrestrial PV Modules’) still under preparation, whose aim is finding appropriate test procedures and methodologies to calculate the energy yield of PV modules under different climate conditions. Furthermore, the lack of tests in specific conditions of installation (e.g. façade BIPV devices) means that it is difficult knowing the exact effective performance of these systems and the environmental conditions in which the building will operate. The nominal PV power at Standard Test Conditions, STC (1.000 W/m2, 25 °C temperature and AM 1.5) is usually measured in indoor laboratories, and it characterizes the PV module at specific conditions in order to be able to compare different modules and technologies on a first step. The “Watt-peak” is not enough to evaluate the panel performance in terms of Watt-hours of various modules under different operating conditions, and it gives no assurance of being able to predict the energy performance of a certain module at given environmental conditions. A proper BIPV element for façade should take into account thermal and insulation properties, factors as transparency to allow solar gains if possible or a good solar control if necessary, aspects that are linked and high dependent on climate conditions and on the level of comfort to be reached. However, the influence of different façade integrated photovoltaic solutions on the building energy consumption is not easy to assess under real operating conditions. Thermal aspects, indoor temperatures or luminance level that can be expected using building integrated PV (BIPV) modules are not well known. As said before, integrated photovoltaic BIPV components and the use of renewable energy is already a standard for green energy production, but would also be important to know the possible contribution to improve the comfort and health of building occupants. Comfort, light transmission or protection, thermal insulation or thermal/electricity power production are aspects that are usually considered alone, while all together contribute to the building global energy balance. Besides, the need to prioritize a particular building envelope orientation to harvest the most benefit from the electrical or thermal energy production, in the case of active and passive systems respectively might be not compatible, but also not necessary. A holistic approach is needed to enable architects and engineers implementing technological systems working in synergy. A new concept have been suggested: “C-BIPV, conscious integrated BIPV”. BIPV systems have to be “consciously integrated” which means that it is essential to know the positive and negative effects in terms of comfort and energy under real operating conditions. Purpose of the work, method and results The façade-integrated photovoltaic systems are often glass solutions easily integrable, as they usually are custommade. These BIPV semi-transparent components integrated as a window element provides natural lighting and shade that prevents overheating at times of excessive heat, but as static component, likewise avoid the possible solar gains contributions in the cold months. In addition, the temperature of the module varies considerably in certain circumstances influenced by the PV technology installed, solar radiation, mounting system, lack of ventilation, etc. This factor may result in additional heat input in the building highly variable and difficult to quantify. In addition, further insights into the indoor environmental comfort in buildings using integrated photovoltaic technologies are needed to open up thereby, a new research perspective. This research aims to study their behaviour through a series of experiments in order to define the real influence on comfort aspects and on global energy building consumption, as well as, electrical and thermal characteristics of these devices. The final objective was to analyze a whole set of issues that influence the global energy consumption/production in a building using BIPV modules by quantifying the global energy balance and the BIPV system real performances. Other qualitative issues to be studied were comfort aspect (thermal and lighting aspects) and the electrical behaviour of different BIPV technologies for vertical integration, aspects that influence both energy consumption and electricity production. Thus, it will be possible to obtain a comprehensive global characterization of BIPV systems. A specific design of an outdoor test facility, the BIPV Env-lab “BIPV Test Laboratory”, for the integral characterization of different BIPV semi-transparent modules was developed and built. The method and test protocol for the BIPV characterization was also defined in a real building context and weather conditions. This has been possible once assessed the state of the art and research, the aspects that influence the architectural integration and the different possibilities and types of integration for PV and after having examined the test methods for building and photovoltaic components, under operation conditions heretofore used. The test laboratory that consists in two equivalent test rooms (1:1) has a monitoring system in which reliable data of thermal, daylighting and electrical performances can be obtained for the evaluation of PV modules. The experimental set-up facility (testing room) allows studying three different aspects that affect building energy consumption and comfort issues: the thermal indoor comfort, the lighting comfort and the energy performance of BIPV modules tested under real environmental conditions. Knowing the energy balance for each experimented solar technology, it is possible to determine which one performs best. A theoretical methodology has been proposed for evaluating these parameters, as defined in this thesis as indices or indicators, which regard comfort issues, energy and the overall performance of BIPV components. This methodology considers the existing regulatory standards for each aspect, relating them to one another. A set of insulated glass BIPV modules see-through and light-through, representative of different PV technologies (mono-crystalline silicon technology, mc-Si, amorphous silicon thin film single junction, a-Si and copper indium selenide thin film technology CIS) were selected for a series of experimental tests in order to demonstrate the validity of the proposed characterization method. As result, it has been developed and generated the ICD Integral Characterization Diagram, a graphic and visual system to represent the results and manage information, a useful operational tool for decision-making regarding to photovoltaic installations. This diagram shows all concepts and parameters studied in relation to each other and visually provides access to all the results obtained during the experimental phase to make available all the qualitative and quantitative information on the energy performance of the BIPV components by characterizing them in a comprehensive way.