35 resultados para figure of merit

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

There are many studies related with airport surface routing algorithms, based on different approaches and with different evaluation methods and metrics. So, the need of performing a balanced analysis and comparison using a common framework is evident. This paper presents an implementation of an evaluation tool for airport surface routing algorithms. The routing evaluation tool presented here is based in three basic pillars composed by the airport model, the model and generation of traffic and a comprehensive figure of merit function. The paper includes some example evaluations performed over Barajas Airport with representative traffic samples using several simple routing methods.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper, filter design methodology and application of GaN HEMTs for high efficiency Envelope Amplifier in RF transmitters are proposed. The main objectives of the filter design are generation of the envelope reference with the minimum possible distortion and high efficiency of the amplifier obtained by the optimum trade-off between conduction and switching losses. This optimum point was determined using power losses model for synchronous buck with sinusoidal output voltage and experimental results showed good correspondence with the model and verified the proposed methodology. On the other hand, comparing to Si MOSFETs, GaN HEMTs can provide higher efficiency of the envelope amplifier, due to superior conductivity and switching characteristics. Experimental results verified benefits of GaN devices comparing to the appliance of Si switching devices with very good Figure Of Merit, for this particular application

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the Laser-Fired Contact (LFC) process, a laser beam fires a metallic layer through a dielectric passivating layer into the silicon wafer to form an electrical contact with the silicon bulk [1]. This laser technique is an interesting alternative for the fabrication of both laboratory and industrial scale high efficiency passivated emitter and rear cell (PERC). One of the principal characteristics of this promising technique is the capability to reduce the recombination losses at the rear surface in crystalline silicon solar cells. Therefore, it is crucial to optimize LFC because this process is one of the most promising concepts to produce rear side point contacts at process speeds compatible with the final industrial application. In that sense, this work investigates the optimization of LFC processing to improve the back contact in silicon solar cells using fully commercial solid state lasers with pulse width in the ns range, thus studying the influence of the wavelength using the three first harmonics (corresponding to wavelengths of 1064 nm, 532 nm and 355 nm). Previous studies of our group focused their attention in other processing parameters as laser fluence, number of pulses, passivating material [2, 3] thickness of the rear metallic contact [4], etc. In addition, the present work completes the parametric optimization by assessing the influence of the laser wavelength on the contact property. In particular we report results on the morphology and electrical behaviour of samples specifically designed to assess the quality of the process. In order to study the influence of the laser wavelength on the contact feature we used as figure of merit the specific contact resistance. In all processes the best results have been obtained using green (532 nm) and UV (355 nm), with excellent values for this magnitude far below 1 mΩcm2.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The objective of this paper is to provide performance metrics for small-signal stability assessment of a given system architecture. The stability margins are stated utilizing a concept of maximum peak criteria (MPC) derived from the behavior of an impedance-based sensitivity function. For each minor-loop gain defined at every system interface, a single number to state the robustness of stability is provided based on the computed maximum value of the corresponding sensitivity function. In order to compare various power-architecture solutions in terms of stability, a parameter providing an overall measure of the whole system stability is required. The selected figure of merit is geometric average of each maximum peak value within the system. It provides a meaningful metrics for system comparisons: the best system in terms of robust stability is the one that minimizes this index. In addition, the largest peak value within the system interfaces is given thus detecting the weakest point of the system in terms of robustness.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The objective of this paper is to present a simplified method to analyze small-signal stability of a power system and provide performance metrics for stability assessment of a given power-system-architecture. The stability margins are stated utilizing a concept of maximum peak criteria (MPC), derived from the behavior of an impedance-based sensitivity function that provides a single number to state the robustness of the stability of a well-defined minor-loop gain. For each minor-loop gain, defined at every system interface, the robustness of the stability is provided as a maximum value of the corresponding sensitivity function. Typically power systems comprise of various interfaces and, therefore, in order to compare different architecture solutions in terms of stability, a single number providing an overall measure of the whole system stability is required. The selected figure of merit is geometric average of each maximum peak value within the system, combined with the worst case value of system interfaces.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper the capabilities of ultra low power FPGAs to implement Wake-up Radios (WuR) for ultra low energy Wireless Sensor Networks (WSNs) are analyzed. The main goal is to evaluate the utilization of very low power configurable devices to take advantage of their speed, flexibility and low power consumption instead of the more common approaches based on ASICs or microcontrollers. In this context, energy efficiency is a key aspect, considering that usually the instant power consumption is considered a figure of merit, more than the total energy consumed by the application.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

—In this paper, application of a new technological solution for power switches based on Gallium Nitride and a filter design methodology for high efficiency Envelope Amplifier in RF transmitters are proposed. Comparing to Si MOSFETs, GaN HEMTs can provide higher efficiency of the Envelope Amplifier, due to better Figure Of Merit (lower product of on- resistance and gate charge). Benefits of their application were verified through the experimental results. The goal of the filter design is to generate the envelope reference with the minimum possible distortion and to improve the efficiency of the Amplifier, obtaining the optimum trade-off between conduction and switching losses.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quantum Key Distribution is carving its place among the tools used to secure communications. While a difficult technology, it enjoys benefits that set it apart from the rest, the most prominent is its provable security based on the laws of physics. QKD requires not only the mastering of signals at the quantum level, but also a classical processing to extract a secret-key from them. This postprocessing has been customarily studied in terms of the efficiency, a figure of merit that offers a biased view of the performance of real devices. Here we argue that it is the throughput the significant magnitude in practical QKD, specially in the case of high speed devices, where the differences are more marked, and give some examples contrasting the usual postprocessing schemes with new ones from modern coding theory. A good understanding of its implications is very important for the design of modern QKD devices.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The figure of protection "micro-reserves" was created in the Region of Valencia (ANONYMOUS, 1994) with the aim of protecting endangered plant species. This is one of the areas of greatest floristic richness and uniqueness of the western Mediterranean. In this area rare, endemic or threatened vascular flora has a peculiar distribution: they appear to form small fragments spread over the entire region (LAGUNA, 1994; LAGUNA, 2001) The protection of every these small populations of great scientific value has significant challenges. It doesn´t try to protect every species that set out in Annex IV of the by then existing Law 4 / 1989 (repealed in 2007), or to protect to the most ecological level with the creation of Natural Protected Area but an intermediate level: the plant community of small size. According to the decree: “as Micro-Reserve will be declared the natural parcels of land under 20 hectares that contain a high concentration of rare plants, endemic, threatened or of high scientific interest” (ANONYMOUS, 1994) . Of course, the statement of an area as micro-reserve carries certain prohibitions that are harmful to the vegetal community

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In previous works we demonstrated the benefits of using micro–nano patterning materials to be used as bio-photonic sensing cells (BICELLs), referred as micro–nano photonic structures having immobilized bioreceptors on its surface with the capability of recognizing the molecular binding by optical transduction. Gestrinone/anti-gestrinone and BSA/anti-BSA pairs were proven under different optical configurations to experimentally validate the biosensing capability of these bio-sensitive photonic architectures. Moreover, Three-Dimensional Finite Difference Time Domain (FDTD) models were employed for simulating the optical response of these structures. For this article, we have developed an effective analytical simulation methodology capable of simulating complex biophotonic sensing architectures. This simulation method has been tested and compared with previous experimental results and FDTD models. Moreover, this effective simulation methodology can be used for efficiently design and optimize any structure as BICELL. In particular for this article, six different BICELL's types have been optimized. To carry out this optimization we have considered three figures of merit: optical sensitivity, Q-factor and signal amplitude. The final objective of this paper is not only validating a suitable and efficient optical simulation methodology but also demonstrating the capability of this method for analyzing the performance of a given number of BICELLs for label-free biosensing.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

La fiabilidad está pasando a ser el principal problema de los circuitos integrados según la tecnología desciende por debajo de los 22nm. Pequeñas imperfecciones en la fabricación de los dispositivos dan lugar ahora a importantes diferencias aleatorias en sus características eléctricas, que han de ser tenidas en cuenta durante la fase de diseño. Los nuevos procesos y materiales requeridos para la fabricación de dispositivos de dimensiones tan reducidas están dando lugar a diferentes efectos que resultan finalmente en un incremento del consumo estático, o una mayor vulnerabilidad frente a radiación. Las memorias SRAM son ya la parte más vulnerable de un sistema electrónico, no solo por representar más de la mitad del área de los SoCs y microprocesadores actuales, sino también porque las variaciones de proceso les afectan de forma crítica, donde el fallo de una única célula afecta a la memoria entera. Esta tesis aborda los diferentes retos que presenta el diseño de memorias SRAM en las tecnologías más pequeñas. En un escenario de aumento de la variabilidad, se consideran problemas como el consumo de energía, el diseño teniendo en cuenta efectos de la tecnología a bajo nivel o el endurecimiento frente a radiación. En primer lugar, dado el aumento de la variabilidad de los dispositivos pertenecientes a los nodos tecnológicos más pequeños, así como a la aparición de nuevas fuentes de variabilidad por la inclusión de nuevos dispositivos y la reducción de sus dimensiones, la precisión del modelado de dicha variabilidad es crucial. Se propone en la tesis extender el método de inyectores, que modela la variabilidad a nivel de circuito, abstrayendo sus causas físicas, añadiendo dos nuevas fuentes para modelar la pendiente sub-umbral y el DIBL, de creciente importancia en la tecnología FinFET. Los dos nuevos inyectores propuestos incrementan la exactitud de figuras de mérito a diferentes niveles de abstracción del diseño electrónico: a nivel de transistor, de puerta y de circuito. El error cuadrático medio al simular métricas de estabilidad y prestaciones de células SRAM se reduce un mínimo de 1,5 veces y hasta un máximo de 7,5 a la vez que la estimación de la probabilidad de fallo se mejora en varios ordenes de magnitud. El diseño para bajo consumo es una de las principales aplicaciones actuales dada la creciente importancia de los dispositivos móviles dependientes de baterías. Es igualmente necesario debido a las importantes densidades de potencia en los sistemas actuales, con el fin de reducir su disipación térmica y sus consecuencias en cuanto al envejecimiento. El método tradicional de reducir la tensión de alimentación para reducir el consumo es problemático en el caso de las memorias SRAM dado el creciente impacto de la variabilidad a bajas tensiones. Se propone el diseño de una célula que usa valores negativos en la bit-line para reducir los fallos de escritura según se reduce la tensión de alimentación principal. A pesar de usar una segunda fuente de alimentación para la tensión negativa en la bit-line, el diseño propuesto consigue reducir el consumo hasta en un 20 % comparado con una célula convencional. Una nueva métrica, el hold trip point se ha propuesto para prevenir nuevos tipos de fallo debidos al uso de tensiones negativas, así como un método alternativo para estimar la velocidad de lectura, reduciendo el número de simulaciones necesarias. Según continúa la reducción del tamaño de los dispositivos electrónicos, se incluyen nuevos mecanismos que permiten facilitar el proceso de fabricación, o alcanzar las prestaciones requeridas para cada nueva generación tecnológica. Se puede citar como ejemplo el estrés compresivo o extensivo aplicado a los fins en tecnologías FinFET, que altera la movilidad de los transistores fabricados a partir de dichos fins. Los efectos de estos mecanismos dependen mucho del layout, la posición de unos transistores afecta a los transistores colindantes y pudiendo ser el efecto diferente en diferentes tipos de transistores. Se propone el uso de una célula SRAM complementaria que utiliza dispositivos pMOS en los transistores de paso, así reduciendo la longitud de los fins de los transistores nMOS y alargando los de los pMOS, extendiéndolos a las células vecinas y hasta los límites de la matriz de células. Considerando los efectos del STI y estresores de SiGe, el diseño propuesto mejora los dos tipos de transistores, mejorando las prestaciones de la célula SRAM complementaria en más de un 10% para una misma probabilidad de fallo y un mismo consumo estático, sin que se requiera aumentar el área. Finalmente, la radiación ha sido un problema recurrente en la electrónica para aplicaciones espaciales, pero la reducción de las corrientes y tensiones de los dispositivos actuales los está volviendo vulnerables al ruido generado por radiación, incluso a nivel de suelo. Pese a que tecnologías como SOI o FinFET reducen la cantidad de energía colectada por el circuito durante el impacto de una partícula, las importantes variaciones de proceso en los nodos más pequeños va a afectar su inmunidad frente a la radiación. Se demuestra que los errores inducidos por radiación pueden aumentar hasta en un 40 % en el nodo de 7nm cuando se consideran las variaciones de proceso, comparado con el caso nominal. Este incremento es de una magnitud mayor que la mejora obtenida mediante el diseño de células de memoria específicamente endurecidas frente a radiación, sugiriendo que la reducción de la variabilidad representaría una mayor mejora. ABSTRACT Reliability is becoming the main concern on integrated circuit as the technology goes beyond 22nm. Small imperfections in the device manufacturing result now in important random differences of the devices at electrical level which must be dealt with during the design. New processes and materials, required to allow the fabrication of the extremely short devices, are making new effects appear resulting ultimately on increased static power consumption, or higher vulnerability to radiation SRAMs have become the most vulnerable part of electronic systems, not only they account for more than half of the chip area of nowadays SoCs and microprocessors, but they are critical as soon as different variation sources are regarded, with failures in a single cell making the whole memory fail. This thesis addresses the different challenges that SRAM design has in the smallest technologies. In a common scenario of increasing variability, issues like energy consumption, design aware of the technology and radiation hardening are considered. First, given the increasing magnitude of device variability in the smallest nodes, as well as new sources of variability appearing as a consequence of new devices and shortened lengths, an accurate modeling of the variability is crucial. We propose to extend the injectors method that models variability at circuit level, abstracting its physical sources, to better model sub-threshold slope and drain induced barrier lowering that are gaining importance in FinFET technology. The two new proposed injectors bring an increased accuracy of figures of merit at different abstraction levels of electronic design, at transistor, gate and circuit levels. The mean square error estimating performance and stability metrics of SRAM cells is reduced by at least 1.5 and up to 7.5 while the yield estimation is improved by orders of magnitude. Low power design is a major constraint given the high-growing market of mobile devices that run on battery. It is also relevant because of the increased power densities of nowadays systems, in order to reduce the thermal dissipation and its impact on aging. The traditional approach of reducing the voltage to lower the energy consumption if challenging in the case of SRAMs given the increased impact of process variations at low voltage supplies. We propose a cell design that makes use of negative bit-line write-assist to overcome write failures as the main supply voltage is lowered. Despite using a second power source for the negative bit-line, the design achieves an energy reduction up to 20% compared to a conventional cell. A new metric, the hold trip point has been introduced to deal with new sources of failures to cells using a negative bit-line voltage, as well as an alternative method to estimate cell speed, requiring less simulations. With the continuous reduction of device sizes, new mechanisms need to be included to ease the fabrication process and to meet the performance targets of the successive nodes. As example we can consider the compressive or tensile strains included in FinFET technology, that alter the mobility of the transistors made out of the concerned fins. The effects of these mechanisms are very dependent on the layout, with transistor being affected by their neighbors, and different types of transistors being affected in a different way. We propose to use complementary SRAM cells with pMOS pass-gates in order to reduce the fin length of nMOS devices and achieve long uncut fins for the pMOS devices when the cell is included in its corresponding array. Once Shallow Trench isolation and SiGe stressors are considered the proposed design improves both kinds of transistor, boosting the performance of complementary SRAM cells by more than 10% for a same failure probability and static power consumption, with no area overhead. While radiation has been a traditional concern in space electronics, the small currents and voltages used in the latest nodes are making them more vulnerable to radiation-induced transient noise, even at ground level. Even if SOI or FinFET technologies reduce the amount of energy transferred from the striking particle to the circuit, the important process variation that the smallest nodes will present will affect their radiation hardening capabilities. We demonstrate that process variations can increase the radiation-induced error rate by up to 40% in the 7nm node compared to the nominal case. This increase is higher than the improvement achieved by radiation-hardened cells suggesting that the reduction of process variations would bring a higher improvement.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

La electrónica digital moderna presenta un desafío a los diseñadores de sistemas de potencia. El creciente alto rendimiento de microprocesadores, FPGAs y ASICs necesitan sistemas de alimentación que cumplan con requirimientos dinámicos y estáticos muy estrictos. Específicamente, estas alimentaciones son convertidores DC-DC de baja tensión y alta corriente que necesitan ser diseñados para tener un pequeño rizado de tensión y una pequeña desviación de tensión de salida bajo transitorios de carga de una alta pendiente. Además, dependiendo de la aplicación, se necesita cumplir con otros requerimientos tal y como proveer a la carga con ”Escalado dinámico de tensión”, donde el convertidor necesitar cambiar su tensión de salida tan rápidamente posible sin sobreoscilaciones, o ”Posicionado Adaptativo de la Tensión” donde la tensión de salida se reduce ligeramente cuanto más grande sea la potencia de salida. Por supuesto, desde el punto de vista de la industria, las figuras de mérito de estos convertidores son el coste, la eficiencia y el tamaño/peso. Idealmente, la industria necesita un convertidor que es más barato, más eficiente, más pequeño y que aún así cumpla con los requerimienos dinámicos de la aplicación. En este contexto, varios enfoques para mejorar la figuras de mérito de estos convertidores se han seguido por la industria y la academia tales como mejorar la topología del convertidor, mejorar la tecnología de semiconducores y mejorar el control. En efecto, el control es una parte fundamental en estas aplicaciones ya que un control muy rápido hace que sea más fácil que una determinada topología cumpla con los estrictos requerimientos dinámicos y, consecuentemente, le da al diseñador un margen de libertar más amplio para mejorar el coste, la eficiencia y/o el tamaño del sistema de potencia. En esta tesis, se investiga cómo diseñar e implementar controles muy rápidos para el convertidor tipo Buck. En esta tesis se demuestra que medir la tensión de salida es todo lo que se necesita para lograr una respuesta casi óptima y se propone una guía de diseño unificada para controles que sólo miden la tensión de salida Luego, para asegurar robustez en controles muy rápidos, se proponen un modelado y un análisis de estabilidad muy precisos de convertidores DC-DC que tienen en cuenta circuitería para sensado y elementos parásitos críticos. También, usando este modelado, se propone una algoritmo de optimización que tiene en cuenta las tolerancias de los componentes y sensados distorsionados. Us ando este algoritmo, se comparan controles muy rápidos del estado del arte y su capacidad para lograr una rápida respuesta dinámica se posiciona según el condensador de salida utilizado. Además, se propone una técnica para mejorar la respuesta dinámica de los controladores. Todas las propuestas se han corroborado por extensas simulaciones y prototipos experimentales. Con todo, esta tesis sirve como una metodología para ingenieros para diseñar e implementar controles rápidos y robustos de convertidores tipo Buck. ABSTRACT Modern digital electronics present a challenge to designers of power systems. The increasingly high-performance of microprocessors, FPGAs (Field Programmable Gate Array) and ASICs (Application-Specific Integrated Circuit) require power supplies to comply with very demanding static and dynamic requirements. Specifically, these power supplies are low-voltage/high-current DC-DC converters that need to be designed to exhibit low voltage ripple and low voltage deviation under high slew-rate load transients. Additionally, depending on the application, other requirements need to be met such as to provide to the load ”Dynamic Voltage Scaling” (DVS), where the converter needs to change the output voltage as fast as possible without underdamping, or ”Adaptive Voltage Positioning” (AVP) where the output voltage is slightly reduced the greater the output power. Of course, from the point of view of the industry, the figures of merit of these converters are the cost, efficiency and size/weight. Ideally, the industry needs a converter that is cheaper, more efficient, smaller and that can still meet the dynamic requirements of the application. In this context, several approaches to improve the figures of merit of these power supplies are followed in the industry and academia such as improving the topology of the converter, improving the semiconductor technology and improving the control. Indeed, the control is a fundamental part in these applications as a very fast control makes it easier for the topology to comply with the strict dynamic requirements and, consequently, gives the designer a larger margin of freedom to improve the cost, efficiency and/or size of the power supply. In this thesis, how to design and implement very fast controls for the Buck converter is investigated. This thesis proves that sensing the output voltage is all that is needed to achieve an almost time-optimal response and a unified design guideline for controls that only sense the output voltage is proposed. Then, in order to assure robustness in very fast controls, a very accurate modeling and stability analysis of DC-DC converters is proposed that takes into account sensing networks and critical parasitic elements. Also, using this modeling approach, an optimization algorithm that takes into account tolerances of components and distorted measurements is proposed. With the use of the algorithm, very fast analog controls of the state-of-the-art are compared and their capabilities to achieve a fast dynamic response are positioned de pending on the output capacitor. Additionally, a technique to improve the dynamic response of controllers is also proposed. All the proposals are corroborated by extensive simulations and experimental prototypes. Overall, this thesis serves as a methodology for engineers to design and implement fast and robust controls for Buck-type converters.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Cuando se trata de Rem Koolhaas, su espejo no refleja una sola imagen sino múltiples, es un prisma poliédrico. Su espejo nos devuelve el Rem mediático, el intelectual, el conceptualizador, el constructor, el analista, el periodista, el actor... En el caso de esta investigación, fijamos el punto de mira en el Rem COMUNICADOR. “Rem a los dos lados del espejo” se enmarca en una investigación sobre los medios de comunicación de arquitectura, su reflejo en la producción arquitectónica y viceversa. Se trata de llegar a discernir si comunicación y producción arquitectónica colisionan y confluyen en el caso de grandes comunicadores como Rem Koolhaas, si el mensaje y el medio transmisor adquieren las mismas cualidades. Centrándose en la figura de Rem Koolhaas, la tesis aborda la evolución de su faceta comunicativa y las transformaciones sucesivas en el campo de la comunicación arquitectónica, en paralelo a su evolución conceptual a lo largo de su trayectoria. La investigación, por tanto, no se centra tanto en su componente teórica o en la práctica arquitectónica de OMA, sino en la exposición de su producción al mundo, especialmente a través de sus ensayos y libros. “Delirious New York” y “SMLXL” son un reflejo del momento conceptual en que se inscriben, y contienen mucha información sobre los referentes gráficos que irremediablemente han influido en su composición. Especialmente, la aparición de “SMLXL” supuso un revulsivo para el mundo de la comunicación arquitectónica, porque puso el foco sobre la importancia de dejar atrás un discurso narrativo linea y unifocal, para afrontar la comunicación barajando múltiples variables, y aproximaciones, en un proceso similar al desarrollo de un proyecto de arquitectura. Presenta un diseño muy novedoso y una edición extremadamente cuidada, que atiende a parámetros mucho más ambiciosos que los meramente narrativos. Profundiza en la necesidad de una temática global, planteando cuál es la aproximación más apropiada para cada uno de los proyectos que describe, transmitiendo al lector una percepción más allá de lo estrictamente visual, más próximo a lo sensorial. Además, su enorme repercusión a nivel internacional y el gran interés que despertó (no solamente entre los arquitectos, sino también entre diseñadores gráficos, publicistas, personas provenientes de todo tipo de tendencias artísticas y público en general), provocó la globalización del fenómeno de las publicaciones arquitectónicas y puso de manifiesto la importancia de la comunicación como una disciplina en sí misma, dentro de la producción arquitectónica en la era actual. A pesar de la importancia de “SMLXL” a todos los niveles, la presente tesis plantea que, donde realmente se culmina esa experiencia comunicativa, es en “Content”, al incluir nuevos parámetros relacionados con la fusión conceptual de continente y contenido. Es en esta publicación donde el objeto de la comunicación y la expresión de la misma se convierten en un único elemento, que se rige por leyes similares. En este caso, la ley fundamental es la aplicación hasta sus máximas consecuencias de la “cultura de la congestión”, tanto en el mensaje como en el medio, generando lo que hemos convenido en denominar “comunicación congestiva”. Esta concepción deviene en que necesariamente se materialice como un producto efímero, desechable, casi virtual, porque responde a las condiciones de un momento muy concreto y específico y fuera de ese contexto pierde su significación.. La “cultura de la congestión” empieza a surgir en los planteamientos de Koolhaas en la Architectural Association School of Architecture de Londres, bajo la tutela de Elia Zenghelis. Posteriormente se desarrolla en su manifiesto retroactivo sobre Manhattan, “Delirious New York”, donde declara la guerra abierta al urbanismo del movimiento moderno y afirma que la ciudad realmente contemporánea es aquella que es fruto de un desarrollo no planificado, hiperdensa y posible gracias a los avances tecnológicos de su era. Finalmente comienza a materializarse en la Diploma Unit 9 de la AA, donde entra como profesor en 1975, dejando una huella indeleble en las generaciones posteriores de arquitectos que pasaron dicha unidad. Rem Koolhaas es ante todo un intelectual y por ello, todo el constructo teórico sobre la metrópolis comienza a reflejarse en su obra a través de OMA desde el comienzo de su producción. Podemos decir a grandes rasgos que su carrera está marcada por dos hitos históricos fundamentales que determinan tres etapas diferenciadas en su producción. En sus primeros años de profesión, Koolhaas sigue fascinado por la metrópolis urbana y la aplicación del método paranoico crítico a su producción arquitectónica. Es un arquitecto profundamente surrealista. Entiende este método como una estrategia de conocimiento y aproximación al mundo que le rodea: “dejar salir el inconsciente pero sostenerlo con las muletas de la racionalidad”. Pero lo que en realidad le interesa es su aplicación a la gran escala, el “Bigness”, y por ello, participa en proyectos muy ambiciosos de los que surgen conceptos que, más allá de resultar premiados o no, han dejado una huella ideológica en el devenir de la arquitectura. Entre estos proyectos, cabe destacar su propuesta para el Parque de la Villette o la Très Grande Bibliotèque de París. Sus proyectos de esta época destilan una gran carga conceptual, que devienen en unos interiores sorprendentes pero una apariencia exterior sobria o incluso podríamos decir "povera", por el uso de materiales efímeros, poco habituales en la macro-arquitectura hasta ese momento. Súbitamente, en 1997, explotó el denominado “Efecto Bilbao”, de la mano de Frank Gehry (1). El Museo Guggenheim de Bilbao, con su espectacularidad, sus formas pregnantes e imposibles, impacta al mundo. Nace la era de la “Arquitectura del Espectáculo”; la transformación de la ciudad a través de ICONOS que actúen como nodos de atracción y concentración en torno a los cuales supuestamente se revitaliza la actividad económica, cultural y sociopolítica de la ciudad, como si a través de un único gesto se pudieran regenerar todos los tejidos internos de la urbe. Rem Koolhaas comprende rápidamente que la aproximación a la ciudad ha cambiado y, sobre todo, el mercado. En el mundo de la globalización, la única manera de llegar a materializar el “Bigness”, es encerrando sus ejercicios intelectuales en formas pregnantes, bellas, icónicas, espectaculares. Koolhaas encuentra su marca personal en la estética “Stealth”, proveniente de los aviones de combate facetados para evitar los radares, elaborados en los años 80. De esta época surgen proyectos como la Casa da Música de Oporto o la Biblioteca de Seattle; ambos edificios son iconos facetados, de belleza pregnante, que dejan una huella indeleble en la ciudad y provocan, al igual que el Guggenheim, un cierto efecto de recuperación y revitalización en el entorno en que se asientan, al menos de manera temporal. En cualquier caso, Koolhaas nunca abandona los ejercicios meramente teóricos, pero segrega su actividad en dos: OMA produce aquello que tiene vocación de ser construido y se rige por los parámetros del mercado global y AMO, la otra cara del espejo de Rem, aplica el pensamiento arquitectónico a campos no explorados, sin la dependencia de agentes externos, pudiendo permitirse ser un laboratorio puramente experimental. En este escenario, llega el 11 de septiembre de 2001 y el ataque a las Torres Gemelas de Nueva York tiene efectos devastadores a todos los niveles, significando, en un período de tiempo sorprendentemente corto, un cambio en el orden mundial. Rem Koolhaas da entonces un giro de 180 grados, dirige su mirada hacia China, donde entiende que sus aportaciones tienen un beneficio social más directo que en occidente. (2) Para presentar al mundo su nuevo cambio de rumbo y la creación del “Think Tank” AMO, plantea una gran exposición en la NeueGallerie de Berlín bajo el título de “Content”, experiencia paralela a la edición del libro con el mismo título, que inicialmente nace como “catálogo de la exposición, pero que internamente siempre se concibió como el documento más trascendente en el estudio desde “SMLXL”. Sin embargo, en muchos aspectos se trata de su opuesto: una publicación con formato revista, de tapa blanda, con paginado muy fino, formato de "folleto de supermercado" y contenido hiperdenso. Es un experimento efímero, fugaz, ligero, barato, de “usar y tirar”. De hecho, está fuera de stock, ya no se edita. Probablemente Rem Koolhaas desaprobaría que se hiciera una investigación que pusiera el foco sobre el mismo, porque diez años después de su publicación seguramente opine que su vigencia ha caducado. Sin embargo, muestra con una claridad meridiana el estado conceptual y vital de OMA en el momento de su publicación y representa, además un verdadero hito en la comunicación arquitectónica, un punto de no retorno, el máximo exponente de lo que hemos denominado “comunicación congestiva”. La presente tesis plantea que “Content” contiene la esencia de la mayor aportación de Rem Koolhaas al mundo de la arquitectura: la transformación profunda y definitiva de la comunicación arquitectónica mediante la convergencia del estado conceptual y la transmisión del mismo. Su legado arquitectónico y conceptual ha marcado a todas las generaciones posteriores de manera indeleble. Sus ensayos, sus teorías, sus proyectos y sus edificaciones ya pertenecen a la historia de la arquitectura, sin ninguna duda. Pero es su revisión del concepto de la comunicación en arquitectura lo que ha tenido y tendrá un reflejo inmediato en las generaciones futuras, no solamente en la comunicación sino en su arquitectura, a través de un intercambio biyectivo. El planteamiento a futuro sería determinar qué sucede tras “Content”, tras la hiperdensidad máxima, tras la cultura de la congestión visual; qué es lo que propone Koolhaas y qué se va a plantear también en el mundo de la comunicación arquitectónica. Para ello, estudiaremos en profundidad sus últimos proyectos relacionados con la comunicación, como su propuesta para la Biennale de Arquitectura de Venecia de 2014, su intensa investigación sobre el “Metabolismo” en “Project Japan: Metabolism Talks...”, o la dirección de sus últimos planteamientos territoriales. En los últimos tiempos Rem Koolhaas habla de “Preservación”, de “Sobriedad”, de “Esencialismo”, de “Performance”... El autor intelectual de la cultura de la congestión habla ahora de la “low density”...como no podía ser de otra manera en la otra cara del espejo. En definitiva, el color blanco como suma de todos los colores, todas las longitudes de onda del espectro visible recibidas al tiempo. ABSTRACT When talking about Rem Koolhaas, the mirror does not only reflect one but numerous images: it is nothing but a polyhedral prism. His mirror gives us the image of Rem the media celebrity, the intellectual, the conceptualizer, the builder, the analyst, the journalist, the actor... This research sets the spotlight on Rem the COMMUNICATOR. "Rem on both sides of the mirror" belongs to a research on architectural media, its influence on the architectural production and vice versa. It is aimed at getting to discern whether communication and architectural production collide and converge in the case of great communicators such as Rem Koolhaas, and whether the message and transmission media acquire the same features. Focusing on the figure of Rem Koolhaas, this thesis addresses the evolution of his communicative facet and the successive transformations in the field of architectural communication, parallel to the conceptual evolution he underwent throughout his career. Therefore, this research is not so much focused on his theoretical component or on the OMA’s architectural practice, but on the exhibition of his production to the world, especially through his essays and books. "Delirious New York" and "SMLXL" hold up a mirror to the conceptual moment they are part of, and contain a great deal of information about the graphic references that have inevitably influenced his work. Specially, the launch of "SMLXL" was a salutary shock for the architectural communication world, since it set the spotlight on the importance of leaving a linear and unifocal narrative behind in order to face communication considering multiple variables and approaches, based on a process similar to the development of an architectural project. It offers a very innovative design and an extremely careful editing, which deals with parameters much more ambitious than those merely narrative. It explores the need for a global subject and suggests the most appropriate approach for each of the projects described, giving the reader a closer insight to the sensory that goes beyond what’s strictly visual. In addition, its huge international impact and the great interest shown, not only by architects but also by graphic designers, publishers, people from all kinds of artistic trends and the general public, led to the globalisation of the architectural publications phenomenon and brought the importance of communication as a discipline in itself, within the architectural production in the age at hand, to light. Despite the importance of "SMLXL" at all levels, this thesis suggests that the communication experience really culminates in "Content", for it includes new conceptual parameters associated with the container-content conceptual fusion. It is in this book where the purpose of communication and the expression of such become a single element, ruled by similar laws. In this particular case, the fundamental law is to implement the "culture of congestion" to its extreme consequences in both the message and the media, leading to what we have agreed to refer to as "congestive communication”. This concept leads to its inevitable materialisation into an ephemeral, disposable, almost virtual product, because it meets the conditions of a very concrete and specific time, and outside that context it loses its significance. The "culture of congestion" emerged in Koolhaas’ approaches under the guidance of Elia Zenghelis, in the Architectural Association School of Architecture of London. Subsequently, his retroactive manifesto on Manhattan, "Delirious New York" developed it, waging an all-out war against the modern movement urbanism and maintaining that the really contemporary cities are those hyperdense ones that rise as a result of an unplanned development and thanks to the typical technological advances of their time. Finally it began to materialise in the Diploma Unit 9 of the AA, in which he started lecturing in 1975, leaving an indelible mark on subsequent generations of architects who passed that unit. First and foremost, Rem Koolhaas is an intellectual and, therefore, all the theoretical construct in the metropolis began to be reflected in his work through OMA since the beginnings of his production. Broadly speaking, we can say that his career is influenced by two essential historic events, which determine three different stages in his production. In the early years of his career, Koolhaas was still fascinated by the urban metropolis and the implementation of the paranoiac-critical method to his architectural production. He was then a deeply surreal architect. He understood this method as a knowledge strategy and an approach to the world around him: "let the subconscious out but hold it with the crutches of reasonableness”. However, he was actually interested in its implementation on a broad scale, the "Bigness", and therefore, he took part in ambitious projects that led to the accrual of concepts that, beyond being rewarded, left an ideological impression on the evolution of architecture. These projects included his proposal for the Parc de la Villette or the Très Grande Bibliotèque in Paris. The projects he carried out during this period showed a great conceptual background, which evolved into surprising interiors but a sober, or even "povera", exterior appearance, thanks to the use of ephemeral materials that were atypical in the macro-architecture field until that moment. Suddenly, in 1997, the so-called "Bilbao effect" boomed thanks to Frank Gehry (1). The Guggenheim Museum of Bilbao amazed the world with its spectacular nature and its pregnant and impossible shapes. It was the beginning of the era of “The architecture of spectacle”: the transformation of the city through ICONS that would act as nodes of attraction and gathering, around which the economic, cultural and socio-political activity of the city was supposed to be revitalized, as if through a single gesture all internal tissues of the city could be rebuilt. Rem Koolhaas quickly realized that the approach to the city, and especially to the global market, had changed. In the world of globalisation, the only way to get to materialise such "Bigness" was by keeping his intellectual exercises in pregnant, beautiful, iconic and spectacular shapes. Koolhaas found his personal brand in the Stealth aesthetic, resulting from the eighties American combat aircrafts whose shape was faceted in order to avoid radars. Projects such as the Casa da Música in Porto or the Seattle Library date from this period; both buildings are faceted icons of pregnant beauty that left an indelible mark on the city and caused, like the Guggenheim, some degree of recovery and revitalization on the environment in which they were based, at least temporarily. In any case, Koolhaas never gave the merely theoretical exercises up, but he segregated his work in two: OMA produced what was destined to be built and ruled by the parameters of the global market and AMO, Rem’s other side of the mirror, applied the architectural thought in unexplored fields, notwithstanding external agents and being able to work as a purely experimental laboratory. In light of this backdrop, September 11th 2001 came and the attacks on the Twin Towers in New York had devastating effects at all levels, leading to a change in the world order, in a surprisingly short period of time. Rem Koolhaas made a 180° turn directing his vision towards China, where he believed his contributions would have a more direct social benefit than in the Western world. (2) In order to introduce his new course of direction and the creation of the AMO "Think Tank", he planned a major exhibition in the Neue Nationalgalerie of Berlin under the title "Content", in parallel with edition of the book with the same title, which was at first the "exhibition catalog” but, deep down, was always conceived as the most important document of the Office since "SMLXL". However, in many ways it was just the opposite: a publication characterised by its magazine format, soft cover, very fine paging, "supermarket brochure" form and hyperdense content. It was an ephemeral, brief, light, cheap and "disposable" experiment. In fact, it is currently out of stock and out of print. Rem Koolhaas would probably disapprove of a research that sets the spotlight on him, for he would probably say that his validity has expired given that it has been ten years since its publication. However, it shows OMA’s conceptual and vital status at the time of its publication with crystalline clarity and it is also a true milestone in architectural communication. A point of no return. The epitome of the so-called "congestive communication ". This thesis suggests that "Content" contains the essence of Rem Koolhaas’ greatest contribution to the world of architecture: the deep and definitive transformation of architectural communication through the convergence of the conceptual state and the transmission thereof. His architectural and conceptual legacy has left an indelible mark on all subsequent generations. There is no doubt his essays, theories, projects and buildings already belong to the history of architecture. But it is his review on the concept of communication in architecture that has had and shall have an immediate influence on future generations, not only in their communication but also in their architecture, through a bijective exchange. Future approaches should try to determine what happens after "Content", after the maximum hyperdensity, after the visual culture of congestion; what shall Koolhaas suggest as well as what shall happen in the world of architectural communication. To this end, we shall study his latest communication-related projects, such as the design of the Venetian Architecture Biennale in 2014, his intensive research on the "Metabolism" in "Project Japan: Metabolism Talks ...", or the course of his latest territorial approaches in depth. Most recently, Rem Koolhaas has talked about "Preservation", "Sobriety" of "Essentialism", "Performance", etc. The mastermind of the culture of congestion now speaks of the "low density"... as it could not be otherwise, on the other side of the mirror. Summarizing, the white color as the sum of all colors; all wavelengths of the visible spectrum received at the same time.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A compact planar array with parasitic elements is studied to be used in MIMO systems. Classical compact arrays suffer from high coupling which makes correlation and matching efficiency to be worse. A proper matching network improves these lacks although its bandwidth is low and may increase the antenna size. The proposed antenna makes use of parasitic elements to improve both correlation and efficiency. A specific software based on MoM has been developed to analyze radiating structures with several feed points. The array is optimized through a Genetic Algorithm to determine parasitic elements position in order to fulfill different figures of merit. The proposed design provides the required correlation and matching efficiency to have a good performance over a significant bandwidth.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The authors are from UPM and are relatively grouped, and all have intervened in different academic or real cases on the subject, at different times as being of different age. With precedent from E. Torroja and A. Páez in Madrid Spain Safety Probabilistic models for concrete about 1957, now in ICOSSAR conferences, author J.M. Antón involved since autumn 1967 for euro-steel construction in CECM produced a math model for independent load superposition reductions, and using it a load coefficient pattern for codes in Rome Feb. 1969, practically adopted for European constructions, giving in JCSS Lisbon Feb. 1974 suggestion of union for concrete-steel-al.. That model uses model for loads like Gumbel type I, for 50 years for one type of load, reduced to 1 year to be added to other independent loads, the sum set in Gumbel theories to 50 years return period, there are parallel models. A complete reliability system was produced, including non linear effects as from buckling, phenomena considered somehow in actual Construction Eurocodes produced from Model Codes. The system was considered by author in CEB in presence of Hydraulic effects from rivers, floods, sea, in reference with actual practice. When redacting a Road Drainage Norm in MOPU Spain an optimization model was realized by authors giving a way to determine the figure of Return Period, 10 to 50 years, for the cases of hydraulic flows to be considered in road drainage. Satisfactory examples were a stream in SE of Spain with Gumbel Type I model and a paper of Ven Te Chow with Mississippi in Keokuk using Gumbel type II, and the model can be modernized with more varied extreme laws. In fact in the MOPU drainage norm the redacting commission acted also as expert to set a table of return periods for elements of road drainage, in fact as a multi-criteria complex decision system. These precedent ideas were used e.g. in wide Codes, indicated in symposia or meetings, but not published in journals in English, and a condensate of contributions of authors is presented. The authors are somehow involved in optimization for hydraulic and agro planning, and give modest hints of intended applications in presence of agro and environment planning as a selection of the criteria and utility functions involved in bayesian, multi-criteria or mixed decision systems. Modest consideration is made of changing in climate, and on the production and commercial systems, and on others as social and financial.