14 resultados para musique expérimentale, noise, post-punk, science-fiction, underground
em Universidad Politécnica de Madrid
Resumo:
En esta Tesis se estudian las Imágenes Arquitectónicas que aparecen descritas en textos literarios de Ciencia Ficción dentro del periodo temporal que limitan los años 1.926 a 1.976. Consta de tres partes fundamentales: En la primera, se plantean las condiciones y objetivos del trabajo y se define el espacio conceptual que ocupa la Arquitectura descrita en Ciencia Ficción. En la segunda, se realizan análisis sobre las formas y temas arquitectónicos, así como de problemas relacionados con ellos que se encuentran en la narrativa fantacientífica, y se desarrollan ideas e hipótesis especulativas, sobre sugerencias referidas a situaciones reales o imaginarias, partiendo de posiciones implícitas en los relatos. En la tercera, se elabora una antología que data y ordena una amplia muestra de textos representativos. SUMMARY. This thesis studies the Architectural Images described in the Science Fiction stories between 1.926 and 1.976. It consist of three main parts: The first one, explains the conditions and objectives of this work and defines the conceptual space of Described Architecture in Science Fiction. The second one, analyses the architectural forms and items, as well as the problems related to the latter, that can be found in the scientific-fantastic narrative. It also develops ideas and speculative hipothesis about suggestions related to real and imaginary situations from implicit positions in the stories. The third one, builds up an anthology, dating and ordering a wide variety of representative excerpts of the S. F. texts o
Resumo:
Desde el comienzo de la Historia, el ser humano ha pensado en su futuro, ya fuera por el hecho de preocuparse por su supervivencia o por razones más filosóficas, como el porvenir como especie. Pensar en el futuro es anticipación. El género de la ciencia ficción es el que más se basa en la anticipación. En muchos subgéneros de la ciencia ficción se tratan temas futuristas, viajes espaciales a otros planetas, inteligencia artificial, etc. Por otra parte, en la actualidad vivimos un período donde los cambios en las TIC (Tecnologías de la Información y Comunicación) son prácticamente diarios, aunque podría llegar a decirse que los cambios se producen hora a hora. A veces, muchos de estos avances nos parecen producto de la fantasía o la ciencia ficción, o salidos de alguna película futurista. Sin embargo, dichos avances son perfectamente posibles y explicables a través de la ciencia actual. A través de este proyecto se pretenden analizar estos avances, relacionándolos con distintas obras de ciencia ficción. Una gran cantidad de los avances tecnológicos tienen su origen en alguna obra de la ciencia ficción, ya sea literatura o cine. Otro objetivo es realizar un estudio de diferentes propuestas tecnológicas de diferentes obras de esta rama que no se hayan realizado todavía, y analizar su viabilidad, su utilidad y los posibles cambios sociológicos que produciría en el mundo en el que vivimos. El tercer objetivo es evaluar la aptitud y actitud de los ingenieros de telecomunicación en cuanto a la innovación y la proyección hacia el futuro de los estos posibles cambios tecnológicos. Abstract Since the beginning of History, humans have thought about their future, either concerned about survival or by philosophical reasons like the future as species. Thinking about the future is speculation. The science fiction genre is the one that is based on speculation. Sub-genres of science fiction covers futuristic themes like space travel to other planets, artificial intelligence, etc. Today we live in a period where changes in ICT (Information and Communication Technologies) are almost daily, although we should say that changes occur in a matter of hours. Sometimes, many of these advances seem a product of fantasy or science fiction, or coming out of a futuristic movie. However, these advances are perfectly possible and explainable by current science. Through this project the intention is to analyze these developments, relating them to various works of science fiction. A great number of this technological advancements have their origin in a work of science fiction, either literature or film. Another objective is to study different technological proposals of this genre that have not been done yet, and analyze their feasibility, usefulness and potential sociological changes that occur in the world we live in. The third objective is to evaluate the ability and attitude of telecommunication engineers in terms of innovation and future projection of these potential technological changes.
Resumo:
El siguiente Trabajo Fin de Master se divide en tres partes, siempre, alrededor de la identificación biométrica. Esta ciencia aprovecha que existen ciertas características biológicas o conductuales singulares e inalterables, por lo que pueden ser analizados y medidos para crear una huella biométrica. Estas características son difíciles de perder, transferir u olvidar y son perdurables en el tiempo. En la primera de las partes se ofrecerá una visión sobre la importancia histórica de esta ciencia, desde los primeros vestigios humanos, en la zona de Mesopotamia, pasando por los grandes innovadores y precursores de la identificación biométrica, como Bertillon, Galton, Vucetich, etc. auspiciados por una finalidad mayoritariamente criminalística o policiaca, hasta la gran revolución a finales del siglo pasado, en las que técnicas que parecían ciencia ficción se hicieron realidad. En el siguiente apartado se analizaran las 6 principales técnicas que se usan actualmente, realizando una mirada más profunda en los principios naturales, fisiológicos y físicos sobre los que se basan se expondrán las tendencias futuras sobre las que trabajara la industria para el desarrollo de técnicas más seguras, menos invasivas y más transparentes para los usuarios finales. Estas como ha pasado a lo largo de la historia sonaran increíbles, pero una vez más la raza humana conseguirá materializarlas e introducirlas en su día a día. Para finalizar y después de este estudio en detalle, se intentará realizar una comparación y análisis basados en las más importantes características para las técnicas biométricas, fiabilidad, facilidad, usurpación, aceptación y estabilidad, etc. ABSTRACT The following Master's Thesis is divided into three parts, always, about biometric identification. This science fail that certain biological or behavioural characteristics unique and unchangeable, so it can be analysed and measured to create a biometric fingerprint. These features are hard to miss, or forget to transfer and are enduring in time. In the first part a vision of the historical importance of this science are offered, from the earliest human remains in the area of Mesopotamia, to the great innovators and pioneers of biometric identification, such as Bertillon, Galton, Vucetich, etc. . sponsored a largely forensic or detective purpose, until the great revolution in the late nineteenth century, in which techniques that seemed science fiction became reality. The following section will analyse the 6 main techniques currently in use, making a deeper look at the natural, physiological and physical principles on which future trends are based on the industry to work for the development of techniques will be discussed more safer, less invasive and more transparent to end users. Such as has happened throughout history sounded amazing, but once again the human race get materialize and introduce them in their day to day. Finally and after the study in detail, and try to make a comparison based on the most important features for biometric technologies, reliability, ease, alienation, acceptance and stability analysis etc..
Resumo:
La obra de Emilio Pérez Piñero que se desarrolla entre los años 1961 y 1972 año en el que muere en un accidente de tráfico volviendo de Figueras, se centra principalmente en artefactos desplegables y desmontables, ejecutando prototipos que en el presente trabajo se han dividido en dos grupos; la cúpula reticular y la infraestructura. No pudo por tanto acudir al Congreso de 1972 de la UIA a recoger el premio Auguste Perret a la innovación tecnológica, que en años anteriores habían recibido Félix Candela, Jean Prouvé, Hans Scharoun o Frei Otto, y que en aquella ocasión tuvo que recoger su viuda. Parámetros como el de la movilidad, indeterminación, intercambiabilidad, obsolescencia y otros que se analizan en el presente trabajo, aparecen a lo largo de toda su obra ya que muchos de sus artefactos están ubicados en no-lugares y tienen un carácter itinerante y por tanto se hace indispensable su rápido montaje y desmontaje, que unas veces se resuelve mediante la desmontabilidad y otras con la plegabilidad de éstos. Aunque pueda parecer Piñero una figura autárquica, lo cierto es que durante la década donde concentra su trabajo se produce una explosión en torno a al arquetipo que será denominado de forma genérica `artefacto´, ligado conceptualmente a los parámetros que aquí se definen. Entendemos artefacto como objeto material realizado por una o varias personas para cumplir una función, es sinónimo de máquina y aparato y deriva de las palabras latinas ars o artis (técnica) y facto (hecho), para designar a los objetos cuya fabricación requiere alguna destreza. El término latino `ars´ engloba a las técnicas y a las artes, lo que no ocurre con el término castellano arte que deriva de él. Los movimientos utópicos que comparte la década con Piñero, utilizan el arquetipo infraestructural, ligero y high tech, para a través de una arquitectura más ligada a la ciencia ficción, realizar una crítica al Movimiento Moderno institucionalizado, todos ellos comparten cierta obsesión por la movilidad, ligada ésta a la idea de espacio flexible, dinámico, nómada. Este concepto de neo-nomadismo, que representa un habitar dinámico, aglutina las nuevas formas de vivir donde la movilidad social y geográfica son habituales. El nomadismo, por otra parte se entiende como sinónimo de democracia y libertad. La arquitectura pasa de ser pesada, estática, permanente a ser un elemento dinámico en continuo movimiento. A veces con connotaciones biológicas se asimilan los artefactos a organismos vivos y les confieren dichas propiedades de crecimiento y autonomía energética, acumulándose en torno a megaestructuras, donde quedan `enchufados´. En este intento de dotar movilidad a lo inmueble, se buscan estructuras vivas y modificables que crecen en una asimilación de las leyes naturales utilizando los parámetros de metamorfosis, simbiosis y cambio. Estos movimientos de vanguardia tienen también ciertas connotaciones políticas y sociales basadas en la libertad y la movilidad y reniegan del consumismo institucionalizado, de la arquitectura como instrumento de consumo, como objeto de usar en la cultura de masas. El carácter político de la autogestión, de la customización como parámetro proyectual, de la autosuficiencia energética, que anticipa la llegada de la crisis energética del año 1973. Objeto de este trabajo será relacionar los conceptos que aparecen fuertemente en el entorno de la década de los años sesenta del siglo XX, en el trabajo de Emilio Pérez Piñero. Parámetros encontrados como conceptos en los grupos de vanguardia y utopía a su vez fuertemente influenciados por las figuras del ingeniero Richard Buckminster Fuller y del arquitecto Konrad Wachsmann. Se analizará que posible influencia tiene la obra de Fuller, principalmente el prototipo denominado cúpula reticular, en la obra de Pérez Piñero y sus coetáneos analizando sus pensamientos teóricos en torno a parámetros como la energía, principalmente en las teorías relativas a Synergetics. El término inventado por Richard B. Fuller es una contracción de otro más largo que en inglés agrupa tres palabras; synergetic-energetic geometry. La definición de sinergia es la cooperación, es decir es el resultado de la acción conjunta de dos o más causas, pero con un efecto superior a la suma de estas causas. El segundo término, energetics geometry, que traducido sería geometría energética hace referencia en primer lugar a la geometría; ya que desarrolla el sistema de referencia que utiliza la naturaleza para construir sus sistemas y en segundo lugar a la energía; ya que además debe ser el sistema que establezca las relaciones más económicas utilizando el mínimo de energía. Por otro lado se analiza la repercusión del prototipo denominado Infraestructura, término acuñado por Yona Friedman y basado estructuralmente y conceptualmente en los desarrollos sobre grandes estructuras de Konrad Wachsmann. El arquitecto alemán divulga su conocimiento en seminarios impartidos por todo el mundo como el que imparte en Tokio y se denomina Wachsmann´s Seminar donde participan algunos de los componentes del grupo Metabolista que sorprenderán al mundo con sus realizaciones en la exposición de Osaka de 1970. El intervalo temporal entre 1961 hasta 1972 hace referencia a la horquilla donde Pérez Piñero realiza su obra arquitectónica, que comienza en 1961 cuando gana el concurso convocado en Londres por la UIA (Unión Internacional de Arquitectos) con el proyecto conocido como Teatro Ambulante, hasta 1972 que es cuando fallece volviendo de Figueras donde está realizando dos encargos de Salvador Dalí; la cubrición del escenario del futuro Teatro-Museo Salvador Dalí y la Vidriera Hipercúbica que debía cerrar la boca de tal escenario. Bajo el título de `Artefactos energéticos. De Fuller a Piñero (1961-1972)´, se presenta esta Tesis doctoral, que tiene la intención de vincular la obra de Emilio Pérez Piñero con la de las neo vanguardias producidas por una serie de arquitectos que operan en el ámbito internacional. Estas vinculaciones se producen de una forma general, donde a través de una serie de estrategias según la metodología que posteriormente se describe se buscan relaciones de la obra del autor español con algunos de los movimientos más significativos que aparecen en dicha década y de manera específica estableciendo relaciones con las obras y pensamientos de los autores que pertenecen a estos movimientos y donde estas relaciones se hacen más evidentes. El objeto del presente trabajo es analizar y explicar la obra del arquitecto Emilio Pérez Piñero, que espacialmente se localiza en el territorio español, desde el punto de vista de estos movimientos para posteriormente poder determinar si existen puntos en común y si el arquitecto español no solo comparte la década temporalmente sino también conceptualmente y por tanto utiliza el ideario que utilizan sus coetáneos que forman parte de las neovanguardias de los años sesenta de siglo XX. ABSTRACT ABSTRACT The Work of Emilio Perez Piñero was developed between the years 1961 and 1972 when he died in a car accident coming back from Figueres, where he was building a geodesic dome to close the building that enclose the Dali’s museum. All his Work is mainly centered in artifact that could be collapsible and removable, taking the two prototypes that are described in this work as a recurrent element in all his creation. These are the reticular dome and the infrastructure that are very influenced by the work from Richard B. Fuller and Konrad Wachsmann. Emilio Pérez Piñero could not receive the Auguste Perret Prize in 1972 awarded by the UIA that years before have received architects as Felix Candela, Jean Prouvé, Hans Scharoun or Frei Otto, and this time Pérez Piñero´s wife will accept it because of his death. Parameters like mobility, changeability, expendability, indetermination and others appear currently in his Work. All the inventions that Piñero had been patented and all of the artifacts that he created are usually located in no-places, because they do have a shifting identity. This kind of building has to be quickly set on site, and this problem has to be solved in term of foldability or demounting. In the decade where his work focuses, an explosion has occurred around this archetype to be generally called artifact that is usually linked to mobility. We understand artifact as a material object made by one or more people to work in a particular way. It is sometimes equated with the terms machinery and apparatus and it is derived from the Latin word `ars´ or `artis´, what means techniques and `facto´ (fact). And we use this term to refer to objects whose manufacture requires the same skill, in fact the Latin word `ars´ covers the techniques and arts, which does not occur with the Castillan term `arte´ that derives from it and means only art. The term neo-nomadic is a relatively new name used for a dynamic life, commonly referred to new forms of life where social and geographical mobility are common. On the other hand nomadic could be understood as a synonymous for democracy and freedom. The architecture is not going to be hard and static anymore but a dynamic element in the move. The Neo-avant-garde movement that shares the decade with Piñero uses this infrastructural archetype, which is light and high-tech, to criticize the institutionalized Modern Movement through architecture linked to science fiction. They all share an obsession with mobility, a concept that is connected to the terms `dynamic´, `nomadic´, `flexibile´, etc. Sometimes, with biological connotations, the utopian assimilate the artifacts to living organisms and give them these properties of growth and energy autonomy, and they apparently grow around megastructures where they are plugged. In this attempt to provide mobility to the inertness, living structures and possibility of change are sought in order to make them grow like a living organism and to assimilate the natural laws of growth. According to a definition from architecture provided by Fernández- Galiano who calls it `exosomatic artifact´, he understand architecture as artifact of the human environment that regulates natural energy flows and channels the energy stored in fuels for the benefit of living beings that inhabit. It is also true that during the sixties a new environmental awareness in public opinion is formed and that is due to the exploitation and disproportionate use of energy resources, acceleration of technological processes and mass consumption. Consequently a new concept is born: energy autonomy, it is very close to rational use of natural energy. Such a concept will be culturally assimilated with the requirement of independence not only in the management but also in the building construction until we arrive at energy autonomy. The individuals become energy consumer, which in turn can enter the energy produced in the system to `life in an eco-mode way´. The objectives of this research are analyzing all of these parameters and concepts that are coming into view in the surrounding of the decade and relate them with the Work of Pérez Piñero. Terms strongly present in the avant-garde movements around the decade, a young architect’s generation strongly influenced by Richard B. Fuller and Konrad Wachsmann. However, it will be analyzed how important the influence of Buckminster Fuller's Work was and his theoretical text about energy on the Work of Pérez Piñero and his fellows of the decade. The term Synergetic was invented by Fuller and came from the words synergy and energetic geometry. Synergy is the cooperation or interaction of two or more agents to produce a greater effect than the sum of their separate effects. Energetic geometry is related to the geometries that the Nature is using to build their construction but always using low energy consumption. On the other hand, the influences from Wachsmann around the prototype called Infrastructure have been analyzed. The German architect has developed knowledge around huge structures that he has spread all around the world through seminars that he has been conducted. One of these was the Wachsmann´s seminar in Tokyo, where same of the members of the Metabolist group were taking part of. Later these young architects will surprise the world with his artifacts at the World Exposition in Osaka in 1970. Between 1961 and 1972 Pérez Piñero produced his architectural work. It began in 1961 when he received the first prize with his project Mobile Theatre in the competition organized by the UIA in London. In 1972 the Auguste Perret Prize was granted by the UIA too. He could not accept it because he died before in a car accident when he was coming from Figueres, when he was designing two projects for Dali. With the title `Energetic Artifacts. From Fuller to Piñero (1961- 1972)´, this thesis relates the Work of Emilio Pérez Piñero with the neo avant-garde made by a young architects’ generation who is sharing the time with him. Several strategies have been used to formed relationships between them. They are described in the present work to set up a method that allows us to relate the work and ideas of the architects of the neo avant-garde with the ones from Piñero. This work is intended to analyze and explained the work of Pérez Piñero from the point of view of the international architects’ generation who is operating at the same time and finally to determinate if Piñero is not sharing the time with them but the concepts, ideas and architectural parameters.
Resumo:
Flat or worn wheels rolling on rough or corrugated tracks can provoke airborne noise and ground-borne vibration, which can be a serious concern for nearby neighbours of urban rail transit lines. Among the various treatments used to reduce vibration and noise, resilient wheels play an important role. In conventional resilient wheels, a slightly prestressed Vshaped rubber ring is mounted between the steel wheel centre and tyre. The elastic layer enhances rolling noise and vibration suppression, as well as impact reduction on the track. In this paper the effectiveness of resilient wheels in underground lines, in comparison to monobloc ones, is assessed. The analysed resilient wheel is able to carry greater loads than standard resilient wheels used for light vehicles. It also presents a greater radial resiliency and a higher axial stiffness than conventional Vwheels. The finite element method was used in this study. A quarter car model was defined, in which the wheelset was modelled as an elastic body. Several simulations were performed in order to assess the vibrational behaviour of elastic wheels, including modal, harmonic and random vibration analysis, the latter allowing the introduction of realistic vertical track irregularities, as well as the influence of the running speed. Due to numerical problems some simplifications were needed. Parametric variations were also performed, in which the sensitivity of the whole system to variations of rubber prestress and Poisson’s ratio of the elastic material was assessed.Results are presented in the frequency domain, showing a better performance of the resilient wheels for frequencies over 200 Hz. This result reveals the ability of the analyzed design to mitigate rolling noise, but not structural vibrations, which are primarily found in the lower frequency range.
Resumo:
In this paper, vehicle-track interaction for a new slab track design, conceived to reduce noise and vibration levels has been analyzed, assessing the derailment risk for trains running on curved track when encountering a broken rail. Two different types of rail fastening systems with different elasticities have been analysed and compared. Numerical methods were used in order to simulate the dynamic behaviour of the train-track interaction. Multibody system (MBS) modelling techniques were combined with techniques based on the finite element method (FEM). MBS modelling was used for modelling the vehicle and FEM for simulating the elastic track. The simulation model was validated by comparing simulated results to experimental data obtained in field testing. During the simulations various safety indices, characteristic of derailment risk, were analysed. The simulations realised at the maximum running velocity of 110 km/h showed a similar behaviour for several track types. When reducing the running speed, the safety indices worsened for both cases. Although the worst behaviour was observed for the track with a greater elasticity, in none of the simulations did a derailment occur when running over the broken rail.
Resumo:
This study focuses on the effectiveness of resilient wheels in reducing railway noise and vibrations, and compares the effectiveness of three types of wheels. The finite elements method has been used to characterise the vibratory behaviour of these wheels. The model has been excited with a realistic spectrum of vertical track irregularities, and a spectral analysis has been carried out. Results have been post-processed in order to estimate the sound power emitted. These calculations have been used to assess the effectiveness of the resilient wheel designs in reducing noise emitted to the environment and in propagating structural vibrations.
Resumo:
Vehicle–track interaction for a new resilient slab track designed to reduce noise and vibration levels was analysed, in order to assess the derailment risk on a curved track when encountering a broken rail. Sensitivity of the rail support spacing of the relative position of the rail breakage between two adjacent rail supports and of running speed were analysed for two different elasticities of the rail fastening system. In none of the cases analysed was observed an appreciable difference between either of the elastic systems. As was expected, the most unfavourable situations were those with greater rail support spacing and those with greater distance from the breakage to the nearest rail support, although in none of the simulations performed did a derailment occur when running over the broken rail. When varying the running speed, the most favourable condition was obtained for an intermediate speed, due to the superposition of two antagonistic effects.
Resumo:
One important task in the design of an antenna is to carry out an analysis to find out the characteristics of the antenna that best fulfills the specifications fixed by the application. After that, a prototype is manufactured and the next stage in design process is to check if the radiation pattern differs from the designed one. Besides the radiation pattern, other radiation parameters like directivity, gain, impedance, beamwidth, efficiency, polarization, etc. must be also evaluated. For this purpose, accurate antenna measurement techniques are needed in order to know exactly the actual electromagnetic behavior of the antenna under test. Due to this fact, most of the measurements are performed in anechoic chambers, which are closed areas, normally shielded, covered by electromagnetic absorbing material, that simulate free space propagation conditions, due to the absorption of the radiation absorbing material. Moreover, these facilities can be employed independently of the weather conditions and allow measurements free from interferences. Despite all the advantages of the anechoic chambers, the results obtained both from far-field measurements and near-field measurements are inevitably affected by errors. Thus, the main objective of this Thesis is to propose algorithms to improve the quality of the results obtained in antenna measurements by using post-processing techniques and without requiring additional measurements. First, a deep revision work of the state of the art has been made in order to give a general vision of the possibilities to characterize or to reduce the effects of errors in antenna measurements. Later, new methods to reduce the unwanted effects of four of the most commons errors in antenna measurements are described and theoretical and numerically validated. The basis of all them is the same, to perform a transformation from the measurement surface to another domain where there is enough information to easily remove the contribution of the errors. The four errors analyzed are noise, reflections, truncation errors and leakage and the tools used to suppress them are mainly source reconstruction techniques, spatial and modal filtering and iterative algorithms to extrapolate functions. Therefore, the main idea of all the methods is to modify the classical near-field-to-far-field transformations by including additional steps with which errors can be greatly suppressed. Moreover, the proposed methods are not computationally complex and, because they are applied in post-processing, additional measurements are not required. The noise is the most widely studied error in this Thesis, proposing a total of three alternatives to filter out an important noise contribution before obtaining the far-field pattern. The first one is based on a modal filtering. The second alternative uses a source reconstruction technique to obtain the extreme near-field where it is possible to apply a spatial filtering. The last one is to back-propagate the measured field to a surface with the same geometry than the measurement surface but closer to the AUT and then to apply also a spatial filtering. All the alternatives are analyzed in the three most common near-field systems, including comprehensive noise statistical analyses in order to deduce the signal-to-noise ratio improvement achieved in each case. The method to suppress reflections in antenna measurements is also based on a source reconstruction technique and the main idea is to reconstruct the field over a surface larger than the antenna aperture in order to be able to identify and later suppress the virtual sources related to the reflective waves. The truncation error presents in the results obtained from planar, cylindrical and partial spherical near-field measurements is the third error analyzed in this Thesis. The method to reduce this error is based on an iterative algorithm to extrapolate the reliable region of the far-field pattern from the knowledge of the field distribution on the AUT plane. The proper termination point of this iterative algorithm as well as other critical aspects of the method are also studied. The last part of this work is dedicated to the detection and suppression of the two most common leakage sources in antenna measurements. A first method tries to estimate the leakage bias constant added by the receiver’s quadrature detector to every near-field data and then suppress its effect on the far-field pattern. The second method can be divided into two parts; the first one to find the position of the faulty component that radiates or receives unwanted radiation, making easier its identification within the measurement environment and its later substitution; and the second part of this method is able to computationally remove the leakage effect without requiring the substitution of the faulty component. Resumen Una tarea importante en el diseño de una antena es llevar a cabo un análisis para averiguar las características de la antena que mejor cumple las especificaciones fijadas por la aplicación. Después de esto, se fabrica un prototipo de la antena y el siguiente paso en el proceso de diseño es comprobar si el patrón de radiación difiere del diseñado. Además del patrón de radiación, otros parámetros de radiación como la directividad, la ganancia, impedancia, ancho de haz, eficiencia, polarización, etc. deben ser también evaluados. Para lograr este propósito, se necesitan técnicas de medida de antenas muy precisas con el fin de saber exactamente el comportamiento electromagnético real de la antena bajo prueba. Debido a esto, la mayoría de las medidas se realizan en cámaras anecoicas, que son áreas cerradas, normalmente revestidas, cubiertas con material absorbente electromagnético. Además, estas instalaciones se pueden emplear independientemente de las condiciones climatológicas y permiten realizar medidas libres de interferencias. A pesar de todas las ventajas de las cámaras anecoicas, los resultados obtenidos tanto en medidas en campo lejano como en medidas en campo próximo están inevitablemente afectados por errores. Así, el principal objetivo de esta Tesis es proponer algoritmos para mejorar la calidad de los resultados obtenidos en medida de antenas mediante el uso de técnicas de post-procesado. Primeramente, se ha realizado un profundo trabajo de revisión del estado del arte con el fin de dar una visión general de las posibilidades para caracterizar o reducir los efectos de errores en medida de antenas. Después, se han descrito y validado tanto teórica como numéricamente nuevos métodos para reducir el efecto indeseado de cuatro de los errores más comunes en medida de antenas. La base de todos ellos es la misma, realizar una transformación de la superficie de medida a otro dominio donde hay suficiente información para eliminar fácilmente la contribución de los errores. Los cuatro errores analizados son ruido, reflexiones, errores de truncamiento y leakage y las herramientas usadas para suprimirlos son principalmente técnicas de reconstrucción de fuentes, filtrado espacial y modal y algoritmos iterativos para extrapolar funciones. Por lo tanto, la principal idea de todos los métodos es modificar las transformaciones clásicas de campo cercano a campo lejano incluyendo pasos adicionales con los que los errores pueden ser enormemente suprimidos. Además, los métodos propuestos no son computacionalmente complejos y dado que se aplican en post-procesado, no se necesitan medidas adicionales. El ruido es el error más ampliamente estudiado en esta Tesis, proponiéndose un total de tres alternativas para filtrar una importante contribución de ruido antes de obtener el patrón de campo lejano. La primera está basada en un filtrado modal. La segunda alternativa usa una técnica de reconstrucción de fuentes para obtener el campo sobre el plano de la antena donde es posible aplicar un filtrado espacial. La última es propagar el campo medido a una superficie con la misma geometría que la superficie de medida pero más próxima a la antena y luego aplicar también un filtrado espacial. Todas las alternativas han sido analizadas en los sistemas de campo próximos más comunes, incluyendo detallados análisis estadísticos del ruido con el fin de deducir la mejora de la relación señal a ruido lograda en cada caso. El método para suprimir reflexiones en medida de antenas está también basado en una técnica de reconstrucción de fuentes y la principal idea es reconstruir el campo sobre una superficie mayor que la apertura de la antena con el fin de ser capaces de identificar y después suprimir fuentes virtuales relacionadas con las ondas reflejadas. El error de truncamiento que aparece en los resultados obtenidos a partir de medidas en un plano, cilindro o en la porción de una esfera es el tercer error analizado en esta Tesis. El método para reducir este error está basado en un algoritmo iterativo para extrapolar la región fiable del patrón de campo lejano a partir de información de la distribución del campo sobre el plano de la antena. Además, se ha estudiado el punto apropiado de terminación de este algoritmo iterativo así como otros aspectos críticos del método. La última parte de este trabajo está dedicado a la detección y supresión de dos de las fuentes de leakage más comunes en medida de antenas. El primer método intenta realizar una estimación de la constante de fuga del leakage añadido por el detector en cuadratura del receptor a todos los datos en campo próximo y después suprimir su efecto en el patrón de campo lejano. El segundo método se puede dividir en dos partes; la primera de ellas para encontrar la posición de elementos defectuosos que radian o reciben radiación indeseada, haciendo más fácil su identificación dentro del entorno de medida y su posterior substitución. La segunda parte del método es capaz de eliminar computacionalmente el efector del leakage sin necesidad de la substitución del elemento defectuoso.
Resumo:
The elemental distribution for as-received (AR), H implanted (AI) and post-implanted annealed (A) Eurofer and ODS-Eurofer steels has been characterized by means of micro Particle Induced X-ray Emission (μ-PIXE), micro Elastic Recoil Detection (μ-ERD) and Secondary Ion Mass Spectrometry (SIMS). The temperature and time-induced H diffusion has been analyzed by Resonance Nuclear Reaction Analysis (RNRA), Thermal Desorption Spectroscopy (TDS), ERDA and SIMS techniques. μ-PIXE measurements point out the presence of inhomogeneities in the Y distribution for ODS-Eurofer samples. RNRA and SIMS experiments evidence that hydrogen easily outdiffuses in these steels even at room temperature. ERD data show that annealing at temperatures as low as 300 °C strongly accelerates the hydrogen diffusion process, driving out up to the 90% of the initial hydrogen.
Resumo:
One of the main obstacles to the widespread adoption of quantum cryptography has been the difficulty of integration into standard optical networks, largely due to the tremendous difference in power of classical signals compared with the single quantum used for quantum key distribution. This makes the technology expensive and hard to deploy. In this letter, we show an easy and straightforward integration method of quantum cryptography into optical access networks. In particular, we analyze how a quantum key distribution system can be seamlessly integrated in a standard access network based on the passive optical and time division multiplexing paradigms. The novelty of this proposal is based on the selective post-processing that allows for the distillation of secret keys avoiding the noise produced by other network users. Importantly, the proposal does not require the modification of the quantum or classical hardware specifications neither the use of any synchronization mechanism between the network and quantum cryptography devices.
Resumo:
Background Magnetoencephalography (MEG) provides a direct measure of brain activity with high combined spatiotemporal resolution. Preprocessing is necessary to reduce contributions from environmental interference and biological noise. New method The effect on the signal-to-noise ratio of different preprocessing techniques is evaluated. The signal-to-noise ratio (SNR) was defined as the ratio between the mean signal amplitude (evoked field) and the standard error of the mean over trials. Results Recordings from 26 subjects obtained during and event-related visual paradigm with an Elekta MEG scanner were employed. Two methods were considered as first-step noise reduction: Signal Space Separation and temporal Signal Space Separation, which decompose the signal into components with origin inside and outside the head. Both algorithm increased the SNR by approximately 100%. Epoch-based methods, aimed at identifying and rejecting epochs containing eye blinks, muscular artifacts and sensor jumps provided an SNR improvement of 5–10%. Decomposition methods evaluated were independent component analysis (ICA) and second-order blind identification (SOBI). The increase in SNR was of about 36% with ICA and 33% with SOBI. Comparison with existing methods No previous systematic evaluation of the effect of the typical preprocessing steps in the SNR of the MEG signal has been performed. Conclusions The application of either SSS or tSSS is mandatory in Elekta systems. No significant differences were found between the two. While epoch-based methods have been routinely applied the less often considered decomposition methods were clearly superior and therefore their use seems advisable.
Resumo:
We study dynamics of the bistable logistic map with delayed feedback, under the influence of white Gaussian noise and periodic modulation applied to the variable. This system may serve as a model to describe population dynamics under finite resources in noisy environment with seasonal fluctuations. While a very small amount of noise has no effect on the global structure of the coexisting attractors in phase space, an intermediate noise totally eliminates one of the attractors. Slow periodic modulation enhances the attractor annihilation.
Resumo:
Plants have extraordinary developmental plasticity as they continuously form organs during post-embryonic development. In addition they may regenerate organs upon in vitro hormonal induction. Advances in the field of plant regeneration show that the first steps of de novo organogenesis through in vitro culture in hormone containing media (via formation of a proliferating mass of cells or callus) require root post-embryonic developmental programs as well as regulators of auxin and cytokinin signaling pathways. We review how hormonal regulation is delivered during lateral root initiation and callus formation. Implications in reprograming, cell fate and pluripotency acquisition are discussed. Finally, we analyze the function of cell cycle regulators and connections with epigenetic regulation. Future work dissecting plant organogenesis driven by both endogenous and exogenous cues (upon hormonal induction) may reveal new paradigms of common regulation.