31 resultados para maximum power point tracking(MPPT)
Resumo:
La capacidad de transporte es uno de los baremos fundamentales para evaluar la progresión que puede llegar a tener un área económica y social. Es un sector de elevada importancia para la sociedad actual. Englobado en los distintos tipos de transporte, uno de los medios de transporte que se encuentra más en alza en la actualidad, es el ferroviario. Tanto para movilidad de pasajeros como para mercancías, el tren se ha convertido en un medio de transporte muy útil. Se encuentra dentro de las ciudades, entre ciudades con un radio pequeño entre ellas e incluso cada vez más, gracias a la alta velocidad, entre ciudades con gran distancia entre ellas. Esta Tesis pretende ayudar en el diseño de una de las etapas más importantes de los Proyectos de instalación de un sistema ferroviario: el sistema eléctrico de tracción. La fase de diseño de un sistema eléctrico de tracción ferroviaria se enfrenta a muchas dudas que deben ser resueltas con precisión. Del éxito de esta fase dependerá la capacidad de afrontar las demandas de energía de la explotación ferroviaria. También se debe atender a los costes de instalación y de operación, tanto costes directos como indirectos. Con la Metodología que se presenta en esta Tesis se ofrecerá al diseñador la opción de manejar un sistema experto que como soluciones le plantee un conjunto de escenarios de sistemas eléctricos correctos, comprobados por resolución de modelos de ecuaciones. Correctos desde el punto de vista de validez de distintos parámetros eléctrico, como de costes presupuestarios e impacto de costes indirectos. Por tanto, el diseñador al haber hecho uso de esta Metodología, tendría en un espacio de tiempo relativamente corto, un conjunto de soluciones factibles con las que poder elegir cuál convendría más según sus intereses finales. Esta Tesis se ha desarrollado en una vía de investigación integrada dentro del Centro de Investigaciones Ferroviarias CITEF-UPM. Entre otros proyectos y vías de investigación, en CITEF se ha venido trabajando en estudios de validación y dimensionamiento de sistemas eléctricos ferroviarios con diversos y variados clientes y sistemas ferroviarios. A lo largo de los proyectos realizados, el interés siempre ha girado mayoritariamente sobre los siguientes parámetros del sistema eléctrico: - Calcular número y posición de subestaciones de tracción. Potencia de cada subestación. - Tipo de catenaria a lo largo del recorrido. Conductores que componen la catenaria. Características. - Calcular número y posición de autotransformadores para sistemas funcionando en alterna bitensión o 2x25kV. - Posición Zonas Neutras. - Validación según normativa de: o Caídas de tensión en la línea o Tensiones máximas en el retorno de la línea o Sobrecalentamiento de conductores o Sobrecalentamiento de los transformadores de las subestaciones de tracción La idea es que las soluciones aportadas por la Metodología sugieran escenarios donde de estos parámetros estén dentro de los límites que marca la normativa. Tener la posibilidad de tener un repositorio de posibles escenarios donde los parámetros y elementos eléctricos estén calculados como correctos, aporta un avance en tiempos y en pruebas, que mejoraría ostensiblemente el proceso habitual de diseño para los sistemas eléctricos ferroviarios. Los costes directos referidos a elementos como subestaciones de tracción, autotransformadores, zonas neutras, ocupan un gran volumen dentro del presupuesto de un sistema ferroviario. En esta Tesis se ha querido profundizar también en el efecto de los costes indirectos provocados en la instalación y operación de sistemas eléctricos. Aquellos derivados del impacto medioambiental, los costes que se generan al mantener los equipos eléctricos y la instalación de la catenaria, los costes que implican la conexión entre las subestaciones de tracción con la red general o de distribución y por último, los costes de instalación propios de cada elemento compondrían los costes indirectos que, según experiencia, se han pensado relevantes para ejercer un cierto control sobre ellos. La Metodología cubrirá la posibilidad de que los diseños eléctricos propuestos tengan en cuenta variaciones de coste inasumibles o directamente, proponer en igualdad de condiciones de parámetros eléctricos, los más baratos en función de los costes comentados. Analizando los costes directos e indirectos, se ha pensado dividir su impacto entre los que se computan en la instalación y los que suceden posteriormente, durante la operación de la línea ferroviaria. Estos costes normalmente suelen ser contrapuestos, cuánto mejor es uno peor suele ser el otro y viceversa, por lo que hace falta un sistema que trate ambos objetivos por separado. Para conseguir los objetivos comentados, se ha construido la Metodología sobre tres pilares básicos: - Simulador ferroviario Hamlet: Este simulador integra módulos para construir esquemas de vías ferroviarios completos; módulo de simulación mecánica y de la tracción de material rodante; módulo de señalización ferroviaria; módulo de sistema eléctrico. Software realizado en C++ y Matlab. - Análisis y estudio de cómo focalizar los distintos posibles escenarios eléctricos, para que puedan ser examinados rápidamente. Pico de demanda máxima de potencia por el tráfico ferroviario. - Algoritmos de optimización: A partir de un estudio de los posibles algoritmos adaptables a un sistema tan complejo como el que se plantea, se decidió que los algoritmos genéticos serían los elegidos. Se han escogido 3 algoritmos genéticos, permitiendo recabar información acerca del comportamiento y resultados de cada uno de ellos. Los elegidos por motivos de tiempos de respuesta, multiobjetividad, facilidad de adaptación y buena y amplia aplicación en proyectos de ingeniería fueron: NSGA-II, AMGA-II y ɛ-MOEA. - Diseño de funciones y modelo preparado para trabajar con los costes directos e indirectos y las restricciones básicas que los escenarios eléctricos no deberían violar. Estas restricciones vigilan el comportamiento eléctrico y la estabilidad presupuestaria. Las pruebas realizadas utilizando el sistema han tratado o bien de copiar situaciones que se puedan dar en la realidad o directamente sistemas y problemas reales. Esto ha proporcionado además de la posibilidad de validar la Metodología, también se ha posibilitado la comparación entre los algoritmos genéticos, comparar sistemas eléctricos escogidos con los reales y llegar a conclusiones muy satisfactorias. La Metodología sugiere una vía de trabajo muy interesante, tanto por los resultados ya obtenidos como por las oportunidades que puede llegar a crear con la evolución de la misma. Esta Tesis se ha desarrollado con esta idea, por lo que se espera pueda servir como otro factor para trabajar con la validación y diseño de sistemas eléctricos ferroviarios. ABSTRACT Transport capacity is one of the critical points to evaluate the progress than a specific social and economical area is able to reach. This is a sector of high significance for the actual society. Included inside the most common types of transport, one of the means of transport which is elevating its use nowadays is the railway. Such as for passenger transport of weight movements, the train is being consolidated like a very useful mean of transport. Railways are installed in many geography areas. Everyone know train in cities, or connecting cities inside a surrounding area or even more often, taking into account the high-speed, there are railways infrastructure between cities separated with a long distance. This Ph.D work aims to help in the process to design one of the most essential steps in Installation Projects belonging to a railway system: Power Supply System. Design step of the railway power supply, usually confronts to several doubts and uncertainties, which must be solved with high accuracy. Capacity to supply power to the railway traffic depends on the success of this step. On the other hand is very important to manage the direct and indirect costs derived from Installation and Operation. With the Methodology is presented in this Thesis, it will be offered to the designer the possibility to handle an expert system that finally will fill a set of possible solutions. These solutions must be ready to work properly in the railway system, and they were tested using complex equation models. This Thesis has been developed through a research way, integrated inside Citef (Railway Research Centre of Technical University of Madrid). Among other projects and research ways, in Citef has been working in several validation studies and dimensioning of railway power supplies. It is been working by a large range of clients and railways systems. Along the accomplished Projects, the main goal has been rounded mostly about the next list of parameters of the electrical system: - Calculating number and location of traction substations. Power of each substation. - Type of Overhead contact line or catenary through the railway line. The wires which set up the catenary. Main Characteristics. - Calculating number and position of autotransformers for systems working in alternating current bi-voltage of called 2x25 kV. - Location of Neutral Zones. - Validating upon regulation of: o Drop voltages along the line o Maximum return voltages in the line o Overheating/overcurrent of the wires of the catenary o Avoiding overheating in the transformers of the traction substations. Main objective is that the solutions given by the Methodology, could be suggest scenarios where all of these parameters from above, would be between the limits established in the regulation. Having the choice to achieve a repository of possible good scenarios, where the parameters and electrical elements will be assigned like ready to work, that gives a great advance in terms of times and avoiding several tests. All of this would improve evidently the regular railway electrical systems process design. Direct costs referred to elements like traction substations, autotransformers, neutral zones, usually take up a great volume inside the general budget in railway systems. In this Thesis has been thought to bear in mind another kind of costs related to railway systems, also called indirect costs. These could be enveloped by those enmarked during installation and operation of electrical systems. Those derived from environmental impact; costs generated during the maintenance of the electrical elements and catenary; costs involved in the connection between traction substations and general electric grid; finally costs linked with the own installation of the whole electrical elements needed for the correct performance of the railway system. These are integrated inside the set has been collected taking into account own experience and research works. They are relevant to be controlled for our Methodology, just in case for the designers of this type of systems. The Methodology will cover the possibility that the final proposed power supply systems will be hold non-acceptable variations of costs, comparing with initial expected budgets, or directly assuming a threshold of budget for electrical elements in actual scenario, and achieving the cheapest in terms of commented costs from above. Analyzing direct and indirect costs, has been thought to divide their impact between two main categories. First one will be inside the Installation and the other category will comply with the costs often happens during Railway Operation time. These costs normally are opposed, that means when one is better the other turn into worse, in costs meaning. For this reason is necessary treating both objectives separately, in order to evaluate correctly the impact of each one into the final system. The objectives detailed before build the Methodology under three basic pillars: - Railway simulator Hamlet: This software has modules to configure many railway type of lines; mechanical and traction module to simulate the movement of rolling stock; signaling module; power supply module. This software has been developed using C++ and Matlab R13a - Previously has been mandatory to study how would be possible to work properly with a great number of feasible electrical systems. The target comprised the quick examination of these set of scenarios in terms of time. This point is talking about Maximum power demand peaks by railway operation plans. - Optimization algorithms. A railway infrastructure is a very complex system. At the beginning it was necessary to search about techniques and optimization algorithms, which could be adaptable to this complex system. Finally three genetic multiobjective algorithms were the chosen. Final decision was taken attending to reasons such as time complexity, able to multiobjective, easy to integrate in our problem and with a large application in engineering tasks. They are: NSGA-II, AMGA-II and ɛ-MOEA. - Designing objectives functions and equation model ready to work with the direct and indirect costs. The basic restrictions are not able to avoid, like budgetary or electrical, connected hardly with the recommended performance of elements, catenary and safety in a electrical railway systems. The battery of tests launched to the Methodology has been designed to be as real as possible. In fact, due to our work in Citef and with real Projects, has been integrated and configured three real railway lines, in order to evaluate correctly the final results collected by the Methodology. Another topic of our tests has been the comparison between the performances of the three algorithms chosen. Final step has been the comparison again with different possible good solutions, it means power supply system designs, provided by the Methodology, testing the validity of them. Once this work has been finished, the conclusions have been very satisfactory. Therefore this Thesis suggest a very interesting way of research and work, in terms of the results obtained and for the future opportunities can be created with the evolution of this. This Thesis has been developed with this idea in mind, so is expected this work could adhere another factor to work in the difficult task of validation and design of railway power supply systems.
Resumo:
The aim of this study was to examine the acute effects of endurance exercise on jumping and kicking performance in young soccer players. Twenty-one top-class young soccer players (16.1±0.2 years) performed a countermovement jump test and a maximal instep soccer kick test before and after running for 20 min on a treadmill at 80% of their individual maximum heart rate. Two force platforms were used to obtain the following parameters during the countermovement jump: jump height, maximum power, maximum power relative to body mass, maximum vertical ground reaction force, maximum vertical ground reaction force relative to body mass, and maximum vertical ground reaction force applied to each leg. Maximum vertical ground reaction force and maximum vertical ground reaction force relative to body mass applied to the support leg during the kicks were also calculated with a force platform. The kicking motion was recorded using a three-dimensional motion-capture system. Maximum velocity of the ball, maximum linear velocity of the toe, ankle, knee and hip, and linear velocity of the toe at ball contact during the kicks were calculated. Non-significant differences were found in the parameters measured during the countermovement jump and the maximal instep soccer kick test before and after running, suggesting that the jumping and kicking performances of top-class young soccer players were not significantly affected after 20 min treadmill running at 80% of their individual maximum heart rate.
Resumo:
Este proyecto continua con el estudio iniciado en los proyectos: Alimentación “Energy Harvesting” basada en fuentes piezoeléctricas, Alimentación autónoma: aplicación a fuente piezoeléctrica y Banco de pruebas para sistemas autoalimentados en aplicaciones viales, que se realizaron anteriormente. Este estudio se basa en la búsqueda de soluciones de alimentación Energy Harvesting centrados en fuentes piezoeléctricas. El objetivo de este estudio es conocer las posibilidades que pueden ofrecer los materiales piezoeléctricos para alimentación autónoma cuando son excitados por vehículos en diferentes circunstancias y entornos viales. Para la realización de este estudio se han utilizado señales provenientes del banco de ensayos, así como señales obtenidas de una instalación construida en un vial del Campus. Para poder realizar el estudio de los diversos materiales piezoeléctricos utilizados en este proyecto, se necesitan conocer las características de los modelos eléctricos. Para obtener estas características se ha diseñado un programa basado en LabView, este programa además puede modificar la señal para que simule las diversas circunstancias a las que pueden estar sometidos los diferentes materiales piezoeléctricos. En este estudio se ha tenido en cuenta la distancia de seguridad entre vehículos para diferentes velocidades y la posibilidad de poder amplificar la excitación que pueden recibir los diversos materiales. Utilizando el programa PSpice OrCAD® se realizan los diferentes modelos con el objetivo de observar el comportamiento de los materiales bajo diversas circunstancias. También se han utilizado varias topologías rectificadoras diferentes para poder comprobar en que situación se puede obtener la máxima tensión acumulada posible. Además también se ha evaluado la máxima potencia que puede entregar cada material en cada situación. Por ultimo se han realizado varias pruebas prácticas con el banco de ensayos para comprobar la acumulación energética real que se puede obtener de los materiales instalados en él. También se ha realizado el estudio de una aplicación práctica que sea capaz de alimentar pequeños circuitos o sensores de poco consumo. Abstract This Project goes on with the study started in the projects Feeding “Energy Harvesting” based on piezoelectric sources, independent feeding: application to piezoelectric source and testing bench to self fed systems in roads applications that were previously done. This study is based on the search of feeding solutions Energy Harvesting focus on piezoelectric sources. The aim of this study is know the different piezoelectric materials possibilities that can offer to independent feeding when different circumstances and road surroundings are excited by vehicles in. To develop this study several signals of the testing bench have been used and also signals get from an installation constructed on a Campus road. As far as the study of the different piezoelectric materials used in this project is concerned we should know the electrical models features. In order to obtain these characteristics a program based on LabView was designed, this program can modify the signal so that it simulates the diverse circumstances to which they can be put under the different piezoelectric materials. In this study I have bearded in mind the safe distance between vehicles for different speeds and the possibility of being able to amplify the excitation that can receive the diverse materials. Using the program PSpice OrCAD® the different models with the objective are made to observe the behaviour of the materials under diverse circumstances. Also several different rectifying topologies have been used to be able to verify the maximum overvoltage accumulated. In addition I have assessed the maximum power that can give each material in each situation. Last but not least different practical proofs with the testing bench have been done to check the actual energy accumulation that can be obtained from the materials installed on it. Also the study of a practical application has been made that is able to feed small circuits or low consumption sensor.
Resumo:
El acero es, junto con el hormigón, el material más ampliamente empleado en la construcción de obra civil y de edificación. Además de su elevada resistencia, su carácter dúctil resulta un aspecto de particular interés desde el punto de vista de la seguridad estructural, ya que permite redistribuir esfuerzos a elementos adyacentes y, por tanto, almacenar una mayor energía antes del colapso final de la estructura. No obstante, a pesar de su extendida utilización, todavía existen aspectos relacionados con su comportamiento en rotura que necesitan una mayor clarificación y que permitirían un mejor aprovechamiento de sus propiedades. Cuando un elemento de acero es ensayado a tracción y alcanza la carga máxima, sufre la aparición de un cuello de estricción que plantea dificultades para conocer el comportamiento del material desde dicho instante hasta la rotura. La norma ISO 6892-1, que define el método a emplear en un ensayo de tracción con materiales metálicos, establece procedimientos para determinar los parámetros relacionados con este tramo último de la curva F − E. No obstante, la definición de dichos parámetros resulta controvertida, ya que éstos presentan una baja reproducibilidad y una baja repetibilidad que resultan difíciles de explicar. En esta Tesis se busca profundizar en el conocimiento del último tramo de la curva F − E de los aceros de construcción. Para ello se ha realizado una amplia campaña experimental sobre dos aceros representativos en el campo de la construcción civil: el alambrón de partida empleado en la fabricación de alambres de pretensado y un acero empleado como refuerzo en hormigón armado. Los dos materiales analizados presentan formas de rotura diferentes: mientras el primero de ellos presenta una superficie de rotura plana con una región oscura claramente apreciable en su interior, el segundo rompe según la clásica superficie en forma de copa y cono. La rotura en forma de copa y cono ha sido ampliamente estudiada en el pasado y existen modelos de rotura que han logrado reproducirla con éxito, en especial el modelo de Gurson- Tvergaard-Needleman (GTN). En cuanto a la rotura exhibida por el primer material, en principio nada impide abordar su reproducción numérica con un modelo GTN, sin embargo, las diferencias observadas entre ambos materiales en los ensayos experimentales permiten pensar en otro criterio de rotura. En la presente Tesis se realiza una amplia campaña experimental con probetas cilíndricas fabricadas con dos aceros representativos de los empleados en construcción con comportamientos en rotura diferentes. Por un lado se analiza el alambrón de partida empleado en la fabricación de alambres de pretensado, cuyo frente de rotura es plano y perpendicular a la dirección de aplicación de la carga con una región oscura en su interior. Por otro lado, se estudian barras de acero empleadas como armadura pasiva tipo B 500 SD, cuyo frente de rotura presenta la clásica superficie en forma de copa y cono. Estos trabajos experimentales han permitido distinguir dos comportamientos en rotura claramente diferenciados entre ambos materiales y, en el caso del primer material, se ha identificado un comportamiento asemejable al exhibido por materiales frágiles. En este trabajo se plantea la hipótesis de que el primer material, cuya rotura provoca un frente de rotura plano y perpendicular a la dirección de aplicación de la carga, rompe de manera cuasifrágil como consecuencia de un proceso de decohesión, de manera que la región oscura que se observa en el centro del frente de rotura se asemeja a una entalla circular perpendicular a la dirección de aplicación de la carga. Para la reproducción numérica de la rotura exhibida por el primer material, se plantea un criterio de rotura basado en un modelo cohesivo que, como aspecto novedoso, se hace depender de la triaxialidad de tensiones, parámetro determinante en el fallo de este tipo de materiales. Este tipo de modelos presenta varias ventajas respecto a los modelos GTN habitualmente empleados. Mientras los modelos GTN precisan de numerosos parámetros para su calibración, los modelos cohesivos precisan fundamentalmente de dos parámetros para definir su curva de ablandamiento: la tensión de decohesión ft y la energía de fractura GF . Además, los parámetros de los modelos GTN no son medibles de manera experimental, mientras que GF sí lo es. En cuanto a ft, aunque no existe un método para su determinación experimental, sí resulta un parámetro más fácilmente interpretable que los empleados por los modelos GTN, que utilizan valores como el porcentaje de huecos presentes en el material para iniciar el fenómeno de coalescencia o el porcentaje de poros que provoca una pérdida total de la capacidad resistente. Para implementar este criterio de rotura se ha desarrollado un elemento de intercara cohesivo dependiente de la triaxialidad de tensiones. Se han reproducido con éxito los ensayos de tracción llevados a cabo en la campaña experimental empleando dicho elemento de intercara. Además, en estos modelos la rotura se produce fenomenológicamente de la misma manera observada en los ensayos experimentales: produciéndose una decohesión circular en torno al eje de la probeta. En definitiva, los trabajos desarrollados en esta Tesis, tanto experimentales como numéricos, contribuyen a clarificar el comportamiento de los aceros de construcción en el último tramo de la curva F − E y los mecanismos desencadenantes de la rotura final del material, aspecto que puede contribuir a un mejor aprovechamiento de las propiedades de estos aceros en el futuro y a mejorar la seguridad de las estructuras construidas con ellos. Steel is, together with concrete, the most widely used material in civil engineering works. Not only its high strength, but also its ductility is of special interest from the point of view of the structural safety, since it enables stress distribution with adjacent elements and, therefore, more energy can be stored before reaching the structural failure. However, despite of being extensively used, there are still some aspects related to its fracture behaviour that need to be clarified and that will allow for a better use of its properties. When a steel item is tested under tension and reaches the maximum load point, necking process begins, which makes difficult to define the material behaviour from that moment onward. The ISO standard 6892-1, which defines the tensile testing method for metallic materials, describes the procedures to obtain some parameters related to this last section of the F − E curve. Nevertheless, these parameters have proved to be controversial, since they have low reproducibility and repeatibility rates that are difficult to explain. This Thesis tries to deepen the knowledge of the last section of the F − E curve for construction steels. An extensive experimental campaign has been carried out with two representative steels used in civil engineering works: a steel rod used for manufacturing prestressing steel wires, before the cold-drawing process is applied, and steel bars used in reinforced concrete structures. Both materials have different fracture surfaces: while the first of them shows a flat fracture surface, perpendicular to the loading direction with a dark region in the centre of it, the second one shows the classical cup-cone fracture surface. The cup-cone fracture surface has been deeply studied in the past and different numerical models have been able to reproduce it with success, with a special mention to the Gurson-Tvergaard-Needleman model (GTN). Regarding the failure surface shown by the first material, in principle it can be numerically reproduced by a GTN model, but the differences observed between both materials in the experimental campaign suggest thinking of a different failure criterium. In the present Thesis, an extensive experimental campaign has been carried out using cylindrical specimens made of two representative construction steels with different fracture behaviours. On one hand, the initial eutectoid steel rod used for manufacturing prestressing steel wires is analysed, which presents a flat fracture surface, perpendicular to the loading direction, and with a dark region in the centre of it. On the other hand, B 500 SD steel bars, typically used in reinforced concrete structures and with the typical cup-cone fracture surface, are studied. These experimental works have allowed distinguishing two clearly different fracture behaviours between both materials and, in the case of the first one, a fragile-like behaviour has been identified. For the first material, which shows a flat fracture surface perpendicular to the loading direction, the following hypothesis is proposed in this study: a quasi-brittle fracture is developed as a consequence of a decohesion process, with the dark region acting as a circular crack perpendicular to the loading direction. To reproduce numerically the fracture behaviour shown by the first material, a failure criterium based on a cohesive model is proposed in this Thesis. As an innovative contribution, this failure criterium depends on the stress triaxiality state of the material, which is a key parameter when studying fracture in this kind of materials. This type of models have some advantages when compared to the widely used GTN models. While GTN models need a high number of parameters to be defined, cohesive models need basically two parameters to define the softening curve: the decohesion stress ft and the fracture energy GF . In addition to this, GTN models parameters cannot be measured experimentally, while GF is indeed. Regarding ft, although no experimental procedure is defined for its obtention, it has an easier interpretation than the parameters used by the GTN models like, for instance, the void volume needed for the coalescence process to start or the void volume that leads to a total loss of the bearing capacity. In order to implement this failure criterium, a triaxiality-dependent cohesive interface element has been developed. The experimental results obtained in the experimental campaign have been successfully reproduced by using this interface element. Furthermore, in these models the failure mechanism is developed in the same way as observed experimentally: with a circular decohesive process taking place around the longitudinal axis of the specimen. In summary, the works developed in this Thesis, both experimental and numerical, contribute to clarify the behaviour of construction steels in the last section of the F − E curve and the mechanisms responsible for the eventual material failure, an aspect that can lead to a better use of the properties of these steels in the future and a safety improvement in the structures built with them.
Resumo:
Modern transmitters usually have to amplify and transmit signals with simultaneous envelope and phase modulation. Due to this property of the transmitted signal, linear power amplifiers (class A, B, or AB) are usually used as a solution for the power amplifier stage. These amplifiers have high linearity, but suffer from low efficiency when the transmitted signal has high peak-to-average power ratio. The Kahn envelope elimination and restoration technique is used to enhance the efficiency of RF transmitters, by combining highly efficient, nonlinear RF amplifier (class E) with a highly efficient envelope amplifier in order to obtain a linear and highly efficient RF amplifier. This paper presents a solution for the envelope amplifier based on a multilevel converter in series with a linear regulator. The multilevel converter is implemented by employing voltage dividers based on switching capacitors. The implemented envelope amplifier can reproduce any signal with a maximum spectral component of 2 MHz and give instantaneous maximum power of 50 W. The efficiency measurements show that when the signals with low average value are transmitted, the implemented prototypes have up to 20% higher efficiency than linear regulators used as a conventional solution.
Resumo:
This paper proposes a first study in-depth of solar-fossil hybridization from a general perspective. It develops a set of useful parameters for analyzing and comparing hybrid plants, it studies the case of hybridizing Brayton cycles with current solar technologies and shows a tentative extrapolation of the results to integrated combined cycle systems (ISCSS). In particular, three points have been analyzed: the technical requirements for solar technologies to be hybridized with Brayton cycles, the temperatures and pressures at which hybridization would produce maximum power per unit of fossil fuel, and their mapping to current solar technologies and Brayton cycles. Major conclusions are that a hybrid plant works in optimum conditions which are not equal to those of the solar or power blocks considered independently, and that hybridizing at the Brayton cycle of a combined cycle could be energetically advantageous.
Resumo:
El proyecto que se presenta a continuación recoge la adaptación de una Central Térmica de carbón al cumplimiento de la DIRECTIVA 2010/75/UE DEL PARLAMENTO EUROPEO Y DEL CONSEJO de 24 de noviembre de 2010 sobre las emisiones industriales. La Central sobre la que se realiza el proyecto tiene un grupo térmico de carbón suscritico refrigerado por agua, con una potencia a plena carga de 350 MWe y de 190 MWe a mínimo técnico. Genera 1 090 t/h de vapor a 540 °C y 168 kg/cm2 funcionando a plena carga. Actualmente las emisiones de NOx son de 650 mg/m3, (condiciones normales, seco, 6 % O2). El objeto del proyecto es reducir estas emisiones a un valor máximo de 200 mg/m3 en las mismas condiciones. El proyecto analiza detalladamente las condiciones actuales de operación de la instalación en cuanto a combustible utilizado, horas de funcionamiento, condiciones climáticas y producción. Se analiza así mismo, todas las técnicas disponibles en mercado para la reducción del NOx, diferenciando entre medidas primarias (actúan sobre los efectos de formación) y secundarias (limpieza de gases). Las medidas primarias ya están implementadas en la central, por tanto, el proyecto plantea la reducción con medidas secundarias. De las medidas secundarias analizadas se ha seleccionado la instalación de un Reactor de Reducción Selectiva Catalítica (Reactor SCR). Tras un análisis de los diferentes reactores y catalizadores disponibles se ha seleccionado un reactor de configuración High-dust, una disposición de catalizador en 3 capas más 1, cuyos componentes están basados en óxidos metálicos (TiO2, V2O5, WO3) y estructura laminar. Se ha buscado la instalación del reactor para operar a una temperatura inferior a 450 °C. Como agente reductor se ha seleccionado NH3 a una dilución del 24,5 %. El proyecto recoge también el diseño de todo el sistema de almacenamiento, evaporación, dilución e inyección de amoniaco. El resultado del proyecto garantiza una concentración en los gases de salida por la chimenea inferior 180 mg/m3(n) de NOx. La reducción del NOx a los límites establecidos, tienen un coste por MWh neto generado para la central, trabajando 60 % a plena carga y 40 % a mínimo técnico y una amortización de 10 años, de 4,10 €/MWh. ABSTRACT The following project shows the compliance adjustment of a coal-fired power station to the 2010/75/EU Directive of the European Parliament and Council 24th November 2010 on industrial emissions. The project is based on a power station with a subcritical thermal coal unit, cooled with water, with a maximum power of 350 MWe and a technical minimum of 190 MWe. It produces 1 090 t/h of steam at 540 ° C and 168 kg/cm2 operating under full load. Currently, NOx emissions are 650 mg / m3 (normal conditions, dry, 6% O2). The project aims to reduce these emissions to a maximum value of 200 mg / m3 under the same conditions. The project analyses in detail the current operating conditions of the system in terms of fuel used, hours of operation, climatic conditions and production. In addition, it also analyses every available technique of NOx reduction on the market, distinguishing between primary (acting on the effects of formation) and secondary measures (gas cleaning). Primary measures are already implemented in the plant, thus proposing reduction with secondary measures. Among the secondary measures analyzed, it has been selected to install a Selective Catalytic Reduction Reactor (SCR Reactor). Having researched the different reactors and catalysts available, for the reactor has been selected High-dust configuration, an arrangement of catalyst in 3 layers plus 1, whose components are based on metal oxides (TiO2, V2O5, WO3) and laminar structure. The reactor has been sought facility to operate at a temperature below 450 ° C. NH3 diluted to 24,5 % has been selected as reducing agent. The project also includes the design of the entire storage system, evaporation, dilution and ammonia injection. The results of the project ensure a gas concentration in the lower chimney exit below 180 mg / m3(n) NOx. The reduction of NOx to the established limits has a cost per net MWh generated in the plant, working at 60% of full load and at 40% of technical minimum, with an amortization of 10 years, 4,10 € / MWh.
Resumo:
El presente trabajo se enmarca en el ámbito de la eficiencia energética y contempla la gestión del consumo eléctrico en hogares. Concretamente, para este proyecto fin de grado se propone el desarrollo de un sistema informático que permita el análisis y monitorización del consumo eléctrico y optimización en la contratación del suministro eléctrico en el hogar. El sistema desarrollado permite la monitorización del consumo eléctrico, expresado en kilovatios-hora (kWh), y la monitorización del coste real de dicho consumo, expresado en euros, en función del tipo de tarifa que se tenga contratada en la modalidad del PVPC1 (Precio Voluntario para el Pequeño Consumidor). También se ha desarrollado una interfaz web a través de la cual el usuario tiene acceso a la información y datos del sistema. En dicha web se muestran gráficas de consumo, potencia, voltaje, corriente y coste de la energía por días. Además, se ha dotado al sistema de un generador de alertas que notifica al usuario, vía web y vía correo electrónico, cuando el consumo sobrepasa los límites fijados por él mismo. El usuario, por tanto, podrá definir los valores de alerta de sobreconsumo y visualizar tanto un histórico de las alertas generadas en el pasado como las alertas activas en ese momento. Las alertas se muestran en la gráfica correspondiente dentro de la aplicación web. Por último, se dispone de la opción de exportar las gráficas que son visualizadas en la aplicación web en formato PNG, JPEG, PDF y SVG, además de la posibilidad de imprimirla.---ABSTRACT---This project belongs to the Energy Efficiency field and is aimed at home energy management. Specifically, for this thesis the development of a computer system that allows monitoring and analysis of energy consumption and contracted power optimization is proposed. The developed system allows energy consumption management within households (expressed in kilowatts per hour, kWh) and real cost monitoring (in euros) according to the contract tariff. A web interface has been developed in order to provide the user with power consumption information and control energy tools. In this web application, electric consumption, power, voltage, current and energy cost by day are shown. Besides, an alert generation system has been implemented so that the user can define maximum power consumption values and be informed through email or web when these values are exceeded. The user will be able to check older power alerts as well as the currently active ones. These alerts are shown in a specific graph within the web application. Finally, the user generated graphs can be exported from the web using PNG, JPEG, PDF or SVG image formats as well as be printed from the web.
Resumo:
Systems used for target localization, such as goods, individuals, or animals, commonly rely on operational means to meet the final application demands. However, what would happen if some means were powered up randomly by harvesting systems? And what if those devices not randomly powered had their duty cycles restricted? Under what conditions would such an operation be tolerable in localization services? What if the references provided by nodes in a tracking problem were distorted? Moreover, there is an underlying topic common to the previous questions regarding the transfer of conceptual models to reality in field tests: what challenges are faced upon deploying a localization network that integrates energy harvesting modules? The application scenario of the system studied is a traditional herding environment of semi domesticated reindeer (Rangifer tarandus tarandus) in northern Scandinavia. In these conditions, information on approximate locations of reindeer is as important as environmental preservation. Herders also need cost-effective devices capable of operating unattended in, sometimes, extreme weather conditions. The analyses developed are worthy not only for the specific application environment presented, but also because they may serve as an approach to performance of navigation systems in absence of reasonably accurate references like the ones of the Global Positioning System (GPS). A number of energy-harvesting solutions, like thermal and radio-frequency harvesting, do not commonly provide power beyond one milliwatt. When they do, battery buffers may be needed (as it happens with solar energy) which may raise costs and make systems more dependent on environmental temperatures. In general, given our problem, a harvesting system is needed that be capable of providing energy bursts of, at least, some milliwatts. Many works on localization problems assume that devices have certain capabilities to determine unknown locations based on range-based techniques or fingerprinting which cannot be assumed in the approach considered herein. The system presented is akin to range-free techniques, but goes to the extent of considering very low node densities: most range-free techniques are, therefore, not applicable. Animal localization, in particular, uses to be supported by accurate devices such as GPS collars which deplete batteries in, maximum, a few days. Such short-life solutions are not particularly desirable in the framework considered. In tracking, the challenge may times addressed aims at attaining high precision levels from complex reliable hardware and thorough processing techniques. One of the challenges in this Thesis is the use of equipment with just part of its facilities in permanent operation, which may yield high input noise levels in the form of distorted reference points. The solution presented integrates a kinetic harvesting module in some nodes which are expected to be a majority in the network. These modules are capable of providing power bursts of some milliwatts which suffice to meet node energy demands. The usage of harvesting modules in the aforementioned conditions makes the system less dependent on environmental temperatures as no batteries are used in nodes with harvesters--it may be also an advantage in economic terms. There is a second kind of nodes. They are battery powered (without kinetic energy harvesters), and are, therefore, dependent on temperature and battery replacements. In addition, their operation is constrained by duty cycles in order to extend node lifetime and, consequently, their autonomy. There is, in turn, a third type of nodes (hotspots) which can be static or mobile. They are also battery-powered, and are used to retrieve information from the network so that it is presented to users. The system operational chain starts at the kinetic-powered nodes broadcasting their own identifier. If an identifier is received at a battery-powered node, the latter stores it for its records. Later, as the recording node meets a hotspot, its full record of detections is transferred to the hotspot. Every detection registry comprises, at least, a node identifier and the position read from its GPS module by the battery-operated node previously to detection. The characteristics of the system presented make the aforementioned operation own certain particularities which are also studied. First, identifier transmissions are random as they depend on movements at kinetic modules--reindeer movements in our application. Not every movement suffices since it must overcome a certain energy threshold. Second, identifier transmissions may not be heard unless there is a battery-powered node in the surroundings. Third, battery-powered nodes do not poll continuously their GPS module, hence localization errors rise even more. Let's recall at this point that such behavior is tight to the aforementioned power saving policies to extend node lifetime. Last, some time is elapsed between the instant an identifier random transmission is detected and the moment the user is aware of such a detection: it takes some time to find a hotspot. Tracking is posed as a problem of a single kinetically-powered target and a population of battery-operated nodes with higher densities than before in localization. Since the latter provide their approximate positions as reference locations, the study is again focused on assessing the impact of such distorted references on performance. Unlike in localization, distance-estimation capabilities based on signal parameters are assumed in this problem. Three variants of the Kalman filter family are applied in this context: the regular Kalman filter, the alpha-beta filter, and the unscented Kalman filter. The study enclosed hereafter comprises both field tests and simulations. Field tests were used mainly to assess the challenges related to power supply and operation in extreme conditions as well as to model nodes and some aspects of their operation in the application scenario. These models are the basics of the simulations developed later. The overall system performance is analyzed according to three metrics: number of detections per kinetic node, accuracy, and latency. The links between these metrics and the operational conditions are also discussed and characterized statistically. Subsequently, such statistical characterization is used to forecast performance figures given specific operational parameters. In tracking, also studied via simulations, nonlinear relationships are found between accuracy and duty cycles and cluster sizes of battery-operated nodes. The solution presented may be more complex in terms of network structure than existing solutions based on GPS collars. However, its main gain lies on taking advantage of users' error tolerance to reduce costs and become more environmentally friendly by diminishing the potential amount of batteries that can be lost. Whether it is applicable or not depends ultimately on the conditions and requirements imposed by users' needs and operational environments, which is, as it has been explained, one of the topics of this Thesis.
Resumo:
El requerimiento de proveer alta frecuencia de datos en los modernos sistema de comunicación inalámbricos resulta en complejas señales moduladas de radio-frequencia (RF) con un gran ancho de banda y alto ratio pico-promedio (PAPR). Para garantizar la linealidad del comportamiento, los amplificadores lineales de potencia comunes funcionan típicamente entre 4 y 10 dB de back-o_ desde la máxima potencia de salida, ocasionando una baja eficiencia del sistema. La eliminación y restauración de la evolvente (EER) y el seguimiento de la evolvente (ET) son dos prometedoras técnicas para resolver el problema de la eficiencia. Tanto en EER como en ET, es complicado diseñar un amplificador de potencia que sea eficiente para señales de RF de alto ancho de banda y alto PAPR. Una propuesta común para los amplificadores de potencia es incluir un convertidor de potencia de muy alta eficiencia operando a frecuencias más altas que el ancho de banda de la señal RF. En este caso, la potencia perdida del convertidor ocasionado por la alta frecuencia desaconseja su práctica cuando el ancho de banda es muy alto. La solución a este problema es el enfoque de esta disertación que presenta dos arquitecturas de amplificador evolvente: convertidor híbrido-serie con una técnica de evolvente lenta y un convertidor multinivel basado en un convertidor reductor multifase con control de tiempo mínimo. En la primera arquitectura, una topología híbrida está compuesta de una convertidor reductor conmutado y un regulador lineal en serie que trabajan juntos para ajustar la tensión de salida para seguir a la evolvente con precisión. Un algoritmo de generación de una evolvente lenta crea una forma de onda con una pendiente limitada que es menor que la pendiente máxima de la evolvente original. La salida del convertidor reductor sigue esa forma de onda en vez de la evolvente original usando una menor frecuencia de conmutación, porque la forma de onda no sólo tiene una pendiente reducida sino también un menor ancho de banda. De esta forma, el regulador lineal se usa para filtrar la forma de onda tiene una pérdida de potencia adicional. Dependiendo de cuánto se puede reducir la pendiente de la evolvente para producir la forma de onda, existe un trade-off entre la pérdida de potencia del convertidor reductor relacionada con la frecuencia de conmutación y el regulador lineal. El punto óptimo referido a la menor pérdida de potencia total del amplificador de evolvente es capaz de identificarse con la ayuda de modelo preciso de pérdidas que es una combinación de modelos comportamentales y analíticos de pérdidas. Además, se analiza el efecto en la respuesta del filtro de salida del convertidor reductor. Un filtro de dampeo paralelo extra es necesario para eliminar la oscilación resonante del filtro de salida porque el convertidor reductor opera en lazo abierto. La segunda arquitectura es un amplificador de evolvente de seguimiento de tensión multinivel. Al contrario que los convertidores que usan multi-fuentes, un convertidor reductor multifase se emplea para generar la tensión multinivel. En régimen permanente, el convertidor reductor opera en puntos del ciclo de trabajo con cancelación completa del rizado. El número de niveles de tensión es igual al número de fases de acuerdo a las características del entrelazamiento del convertidor reductor. En la transición, un control de tiempo mínimo (MTC) para convertidores multifase es novedosamente propuesto y desarrollado para cambiar la tensión de salida del convertidor reductor entre diferentes niveles. A diferencia de controles convencionales de tiempo mínimo para convertidores multifase con inductancia equivalente, el propuesto MTC considera el rizado de corriente por cada fase basado en un desfase fijo que resulta en diferentes esquemas de control entre las fases. La ventaja de este control es que todas las corrientes vuelven a su fase en régimen permanente después de la transición para que la siguiente transición pueda empezar muy pronto, lo que es muy favorable para la aplicación de seguimiento de tensión multinivel. Además, el control es independiente de la carga y no es afectado por corrientes de fase desbalanceadas. Al igual que en la primera arquitectura, hay una etapa lineal con la misma función, conectada en serie con el convertidor reductor multifase. Dado que tanto el régimen permanente como el estado de transición del convertidor no están fuertemente relacionados con la frecuencia de conmutación, la frecuencia de conmutación puede ser reducida para el alto ancho de banda de la evolvente, la cual es la principal consideración de esta arquitectura. La optimización de la segunda arquitectura para más alto anchos de banda de la evolvente es presentada incluyendo el diseño del filtro de salida, la frecuencia de conmutación y el número de fases. El área de diseño del filtro está restringido por la transición rápida y el mínimo pulso del hardware. La rápida transición necesita un filtro pequeño pero la limitación del pulso mínimo del hardware lleva el diseño en el sentido contrario. La frecuencia de conmutación del convertidor afecta principalmente a la limitación del mínimo pulso y a las pérdidas de potencia. Con una menor frecuencia de conmutación, el ancho de pulso en la transición es más pequeño. El número de fases relativo a la aplicación específica puede ser optimizado en términos de la eficiencia global. Otro aspecto de la optimización es mejorar la estrategia de control. La transición permite seguir algunas partes de la evolvente que son más rápidas de lo que el hardware puede soportar al precio de complejidad. El nuevo método de sincronización de la transición incrementa la frecuencia de la transición, permitiendo que la tensión multinivel esté más cerca de la evolvente. Ambas estrategias permiten que el convertidor pueda seguir una evolvente con un ancho de banda más alto que la limitación de la etapa de potencia. El modelo de pérdidas del amplificador de evolvente se ha detallado y validado mediante medidas. El mecanismo de pérdidas de potencia del convertidor reductor tiene que incluir las transiciones en tiempo real, lo cual es diferente del clásico modelos de pérdidas de un convertidor reductor síncrono. Este modelo estima la eficiencia del sistema y juega un papel muy importante en el proceso de optimización. Finalmente, la segunda arquitectura del amplificador de evolvente se integra con el amplificador de clase F. La medida del sistema EER prueba el ahorro de energía con el amplificador de evolvente propuesto sin perjudicar la linealidad del sistema. ABSTRACT The requirement of delivering high data rates in modern wireless communication systems results in complex modulated RF signals with wide bandwidth and high peak-to-average ratio (PAPR). In order to guarantee the linearity performance, the conventional linear power amplifiers typically work at 4 to 10 dB back-off from the maximum output power, leading to low system efficiency. The envelope elimination and restoration (EER) and envelope tracking (ET) are two promising techniques to overcome the efficiency problem. In both EER and ET, it is challenging to design efficient envelope amplifier for wide bandwidth and high PAPR RF signals. An usual approach for envelope amplifier includes a high-efficiency switching power converter operating at a frequency higher than the RF signal's bandwidth. In this case, the power loss of converter caused by high switching operation becomes unbearable for system efficiency when signal bandwidth is very wide. The solution of this problem is the focus of this dissertation that presents two architectures of envelope amplifier: a hybrid series converter with slow-envelope technique and a multilevel converter based on a multiphase buck converter with the minimum time control. In the first architecture, a hybrid topology is composed of a switched buck converter and a linear regulator in series that work together to adjust the output voltage to track the envelope with accuracy. A slow envelope generation algorithm yields a waveform with limited slew rate that is lower than the maximum slew rate of the original envelope. The buck converter's output follows this waveform instead of the original envelope using lower switching frequency, because the waveform has not only reduced slew rate but also reduced bandwidth. In this way, the linear regulator used to filter the waveform has additional power loss. Depending on how much reduction of the slew rate of envelope in order to obtain that waveform, there is a trade-off between the power loss of buck converter related to the switching frequency and the power loss of linear regulator. The optimal point referring to the lowest total power loss of this envelope amplifier is identified with the help of a precise power loss model that is a combination of behavioral and analytic loss model. In addition, the output filter's effect on the response is analyzed. An extra parallel damping filter is needed to eliminate the resonant oscillation of output filter L and C, because the buck converter operates in open loop. The second architecture is a multilevel voltage tracking envelope amplifier. Unlike the converters using multi-sources, a multiphase buck converter is employed to generate the multilevel voltage. In the steady state, the buck converter operates at complete ripple cancellation points of duty cycle. The number of the voltage levels is equal to the number of phases according the characteristics of interleaved buck converter. In the transition, a minimum time control (MTC) for multiphase converter is originally proposed and developed for changing the output voltage of buck converter between different levels. As opposed to conventional minimum time control for multiphase converter with equivalent inductance, the proposed MTC considers the current ripple of each phase based on the fixed phase shift resulting in different control schemes among the phases. The advantage of this control is that all the phase current return to the steady state after the transition so that the next transition can be triggered very soon, which is very favorable for the application of multilevel voltage tracking. Besides, the control is independent on the load condition and not affected by the unbalance of phase current. Like the first architecture, there is also a linear stage with the same function, connected in series with the multiphase buck converter. Since both steady state and transition state of the converter are not strongly related to the switching frequency, it can be reduced for wide bandwidth envelope which is the main consideration of this architecture. The optimization of the second architecture for wider bandwidth envelope is presented including the output filter design, switching frequency and the number of phases. The filter design area is restrained by fast transition and the minimum pulse of hardware. The fast transition needs small filter but the minimum pulse of hardware limitation pushes the filter in opposite way. The converter switching frequency mainly affects the minimum pulse limitation and the power loss. With lower switching frequency, the pulse width in the transition is smaller. The number of phases related to specific application can be optimized in terms of overall efficiency. Another aspect of optimization is improving control strategy. Transition shift allows tracking some parts of envelope that are faster than the hardware can support at the price of complexity. The new transition synchronization method increases the frequency of transition, allowing the multilevel voltage to be closer to the envelope. Both control strategies push the converter to track wider bandwidth envelope than the limitation of power stage. The power loss model of envelope amplifier is detailed and validated by measurements. The power loss mechanism of buck converter has to include the transitions in real time operation, which is different from classical power loss model of synchronous buck converter. This model estimates the system efficiency and play a very important role in optimization process. Finally, the second envelope amplifier architecture is integrated with a Class F amplifier. EER system measurement proves the power saving with the proposed envelope amplifier without disrupting the linearity performance.
Resumo:
Considering the measurement procedures recommended by the ICNIRP, this communication is a proposal for a measurement procedure based in the maximum peak values of equivalent plane wave power density. This procedure has been included in a project being developed in Leganés, Spain. The project plans to deploy a real time monitoring system for RF to provide the city with a useful tool to adapt the environmental EM conditions to the new regulations approved. A first stage consisting of 105 measurement points has been finished and all the values are under the threshold of the regulation.
Resumo:
In this paper, filter design methodology and application of GaN HEMTs for high efficiency Envelope Amplifier in RF transmitters are proposed. The main objectives of the filter design are generation of the envelope reference with the minimum possible distortion and high efficiency of the amplifier obtained by the optimum trade-off between conduction and switching losses. This optimum point was determined using power losses model for synchronous buck with sinusoidal output voltage and experimental results showed good correspondence with the model and verified the proposed methodology. On the other hand, comparing to Si MOSFETs, GaN HEMTs can provide higher efficiency of the envelope amplifier, due to superior conductivity and switching characteristics. Experimental results verified benefits of GaN devices comparing to the appliance of Si switching devices with very good Figure Of Merit, for this particular application
Resumo:
Helium Brayton cycles have been studied as power cycles for both fission and fusion reactors obtaining high thermal efficiency. This paper studies several technological schemes of helium Brayton cycles applied for the HiPER reactor proposal. Since HiPER integrates technologies available at short term, its working conditions results in a very low maximum temperature of the energy sources, something that limits the thermal performance of the cycle. The aim of this work is to analyze the potential of the helium Brayton cycles as power cycles for HiPER. Several helium Brayton cycle configurations have been investigated with the purpose of raising the cycle thermal efficiency under the working conditions of HiPER. The effects of inter-cooling and reheating have specifically been studied. Sensitivity analyses of the key cycle parameters and component performances on the maximum thermal efficiency have also been carried out. The addition of several inter-cooling stages in a helium Brayton cycle has allowed obtaining a maximum thermal efficiency of over 36%, and the inclusion of a reheating process may also yield an added increase of nearly 1 percentage point to reach 37%. These results confirm that helium Brayton cycles are to be considered among the power cycle candidates for HiPER.
Resumo:
The objective of this paper is to provide performance metrics for small-signal stability assessment of a given system architecture. The stability margins are stated utilizing a concept of maximum peak criteria (MPC) derived from the behavior of an impedance-based sensitivity function. For each minor-loop gain defined at every system interface, a single number to state the robustness of stability is provided based on the computed maximum value of the corresponding sensitivity function. In order to compare various power-architecture solutions in terms of stability, a parameter providing an overall measure of the whole system stability is required. The selected figure of merit is geometric average of each maximum peak value within the system. It provides a meaningful metrics for system comparisons: the best system in terms of robust stability is the one that minimizes this index. In addition, the largest peak value within the system interfaces is given thus detecting the weakest point of the system in terms of robustness.
Resumo:
At present, photovoltaic energy is one of the most important renewable energy sources. The demand for solar panels has been continuously growing, both in the industrial electric sector and in the private sector. In both cases the analysis of the solar panel efficiency is extremely important in order to maximize the energy production. In order to have a more efficient photovoltaic system, the most accurate understanding of this system is required. However, in most of the cases the only information available in this matter is reduced, the experimental testing of the photovoltaic device being out of consideration, normally for budget reasons. Several methods, normally based on an equivalent circuit model, have been developed to extract the I-V curve of a photovoltaic device from the small amount of data provided by the manufacturer. The aim of this paper is to present a fast, easy, and accurate analytical method, developed to calculate the equivalent circuit parameters of a solar panel from the only data that manufacturers usually provide. The calculated circuit accurately reproduces the solar panel behavior, that is, the I-V curve. This fact being extremely important for practical reasons such as selecting the best solar panel in the market for a particular purpose, or maximize the energy extraction with MPPT (Maximum Peak Power Tracking) methods.