830 resultados para nonlinear system characterisation


Relevância:

30.00% 30.00%

Publicador:

Resumo:

El principal objetivo de la tesis es estudiar el acoplamiento entre los subsistemas de control de actitud y de control térmico de un pequeño satélite, con el fin de buscar la solución a los problemas relacionados con la determinación de los parámetros de diseño. Se considera la evolución de la actitud y de las temperaturas del satélite bajo la influencia de dos estrategias de orientación diferentes: 1) estabilización magnética pasiva de la orientación (PMAS, passive magnetic attitude stabilization), y 2) control de actitud magnético activo (AMAC, active magnetic attitude control). En primer lugar se presenta el modelo matemático del problema, que incluye la dinámica rotacional y el modelo térmico. En el problema térmico se considera un satélite cúbico modelizado por medio de siete nodos (seis externos y uno interno) aplicando la ecuación del balance térmico. Una vez establecido el modelo matemático del problema, se estudia la evolución que corresponde a las dos estrategias mencionadas. La estrategia PMAS se ha seleccionado por su simplicidad, fiabilidad, bajo coste, ahorrando consumo de potencia, masa coste y complejidad, comparado con otras estrategias. Se ha considerado otra estrategia de control que consigue que el satélite gire a una velocidad requerida alrededor de un eje deseado de giro, pudiendo controlar su dirección en un sistema inercial de referencia, ya que frecuentemente el subsistema térmico establece requisitos de giro alrededor de un eje del satélite orientado en una dirección perpendicular a la radiación solar incidente. En relación con el problema térmico, para estudiar la influencia de la velocidad de giro en la evolución de las temperaturas en diversos puntos del satélite, se ha empleado un modelo térmico linealizado, obtenido a partir de la formulación no lineal aplicando un método de perturbaciones. El resultado del estudio muestra que el tiempo de estabilización de la temperatura y la influencia de las cargas periódicas externas disminuye cuando aumenta la velocidad de giro. Los cambios de temperatura se reducen hasta ser muy pequeños para velocidades de rotación altas. En relación con la estrategia PMAC se ha observado que a pesar de su uso extendido entre los micro y nano satélites todavía presenta problemas que resolver. Estos problemas están relacionados con el dimensionamiento de los parámetros del sistema y la predicción del funcionamiento en órbita. Los problemas aparecen debido a la dificultad en la determinación de las características magnéticas de los cuerpos ferromagnéticos (varillas de histéresis) que se utilizan como amortiguadores de oscilaciones en los satélites. Para estudiar este problema se presenta un modelo analítico que permite estimar la eficiencia del amortiguamiento, y que se ha aplicado al estudio del comportamiento en vuelo de varios satélites, y que se ha empleado para comparar los resultados del modelo con los obtenidos en vuelo, observándose que el modelo permite explicar satisfactoriamente el comportamiento registrado. ABSTRACT The main objective of this thesis is to study the coupling between the attitude control and thermal control subsystems of a small satellite, and address the solution to some existing issues concerning the determination of their parameters. Through the thesis the attitude and temperature evolution of the satellite is studied under the influence of two independent attitude stabilization and control strategies: (1) passive magnetic attitude stabilization (PMAS), and (2) active magnetic attitude control (AMAC). In this regard the mathematical model of the problem is explained and presented. The mathematical model includes both the rotational dynamics and the thermal model. The thermal model is derived for a cubic satellite by solving the heat balance equation for 6 external and 1 internal nodes. Once established the mathematical model of the problem, the above mentioned attitude strategies were applied to the system and the temperature evolution of the 7 nodes of the satellite was studied. The PMAS technique has been selected to be studied due to its prevalent use, simplicity, reliability, and cost, as this strategy significantly saves the overall power, weight, cost, and reduces the complexity of the system compared to other attitude control strategies. In addition to that, another control law that provides the satellite with a desired spin rate along a desired axis of the satellite, whose direction can be controlled with respect to the inertial reference frame is considered, as the thermal subsystem of a satellite usually demands a spin requirement around an axis of the satellite which is positioned perpendicular to the direction of the coming solar radiation. Concerning the thermal problem, to study the influence of spin rate on temperature evolution of the satellite a linear approach of the thermal model is used, which is based on perturbation theory applied to the nonlinear differential equations of the thermal model of a spacecraft moving in a closed orbit. The results of this study showed that the temperature stabilization time and the periodic influence of the external thermal loads decreases by increasing the spin rate. However, the changes become insignificant for higher values of spin rate. Concerning the PMAS strategy, it was observed that in spite of its extended application to micro and nano satellites, still there are some issues to be solved regarding this strategy. These issues are related to the sizing of its system parameters and predicting the in-orbit performance. The problems were found to be rooted in the difficulties that exist in determining the magnetic characteristics of the ferromagnetic bodies (hysteresis rods) that are applied as damping devices on-board satellites. To address these issues an analytic model for estimating their damping efficiency is proposed and applied to several existing satellites in order to compare the results with their respective in-flight data. This model can explain the behavior showed by these satellites.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Con 1.300 millones de personas en el mundo sin acceso a la electricidad (la mayoría en entornos rurales de países empobrecidos), la energía solar fotovoltaica constituye una solución viable técnica y económicamente para electrificar las zonas más remotas del planeta donde las redes eléctricas convencionales no llegan. Casi todos los países en el mundo han desarrollado algún tipo de programa de electrificación fotovoltaica rural durante los últimos 40 años, principalmente los países más pobres, donde a través de diferentes modelos de financiación, se han instalado millones de sistemas solares domiciliarios (pequeños sistemas fotovoltaicos para uso doméstico). Durante este largo período, se han ido superando muchas barreras, como la mejora de la calidad de los sistemas fotovoltaicos, la reducción de costes, la optimización del diseño y del dimensionado de los sistemas, la disponibilidad financiera para implantar programas de electrificación rural, etc. Gracias a esto, la electrificación rural descentralizada ha experimentado recientemente un salto de escala caracterizada por la implantación de grandes programas con miles de sistemas solares domiciliarios e integrando largos períodos de mantenimiento. Muchos de estos grandes programas se están llevando a cabo con limitado éxito, ya que generalmente parten de supuestos e hipótesis poco contrastadas con la realidad, comprometiendo así un retorno económico que permita el desarrollo de esta actividad a largo plazo. En este escenario surge un nuevo reto: el de cómo garantizar la sostenibilidad de los grandes programas de electrificación rural fotovoltaica. Se argumenta que la principal causa de esta falta de rentabilidad es el imprevisto alto coste de la fase de operación y mantenimiento. Cuestiones clave tales como la estructura de costes de operación y mantenimiento o la fiabilidad de los componentes del sistema fotovoltaico no están bien caracterizados hoy en día. Esta situación limita la capacidad de diseñar estructuras de mantenimiento capaces de asegurar la sostenibilidad y la rentabilidad del servicio de operación y mantenimiento en estos programas. Esta tesis doctoral tiene como objetivo responder a estas cuestiones. Se ha realizado varios estudios sobre la base de un gran programa de electrificación rural fotovoltaica real llevado a cabo en Marruecos con más de 13.000 sistemas solares domiciliarios instalados. Sobre la base de este programa se ha hecho una evaluación en profundidad de la fiabilidad de los sistemas solares a partir de los datos de mantenimiento recogidos durante 5 años con más de 80.000 inputs. Los resultados han permitido establecer las funciones de fiabilidad de los equipos tal y como se comportan en condiciones reales de operación, las tasas de fallos y los tiempos medios hasta el fallo para los principales componentes del sistema, siendo este el primer caso de divulgación de resultados de este tipo en el campo de la electrificación rural fotovoltaica. Los dos principales componentes del sistema solar domiciliario, la batería y el módulo fotovoltaico, han sido analizados en campo a través de una muestra de 41 sistemas trabajando en condiciones reales pertenecientes al programa solar marroquí. Por un lado se ha estudiado la degradación de la capacidad de las baterías y por otro la degradación de potencia de los módulos fotovoltaicos. En el caso de las baterías, los resultados nos han permitido caracterizar la curva de degradación en capacidad llegando a obtener una propuesta de nueva definición del umbral de vida útil de las baterías en electrificación rural. También sobre la base del programa solar de Marruecos se ha llevado a cabo un estudio de caracterización de los costes reales de operación y mantenimiento a partir de la base de datos de contabilidad del programa registrados durante 5 años. Los resultados del estudio han permitido definir cuáles son costes que más incidencia tienen en el coste global. Se han obtenido los costes unitarios por sistema instalado y se han calculado los montantes de las cuotas de mantenimiento de los usuarios para garantizar la rentabilidad de la operación y mantenimiento. Finalmente, se propone un modelo de optimización matemática para diseñar estructuras de mantenimiento basado en los resultados de los estudios anteriores. La herramienta, elaborada mediante programación lineal entera mixta, se ha aplicado al programa marroquí con el fin de validar el modelo propuesto. ABSTRACT With 1,300 million people worldwide deprived of access to electricity (mostly in rural environments), photovoltaic solar energy has proven to be a cost‐effective solution and the only hope for electrifying the most remote inhabitants of the planet, where conventional electric grids do not reach because they are unaffordable. Almost all countries in the world have had some kind of rural photovoltaic electrification programme during the past 40 years, mainly the poorer countries, where through different organizational models, millions of solar home systems (small photovoltaic systems for domestic use) have been installed. During this long period, many barriers have been overcome, such as quality enhancement, cost reduction, the optimization of designing and sizing, financial availability, etc. Thanks to this, decentralized rural electrification has recently experienced a change of scale characterized by new programmes with thousands of solar home systems and long maintenance periods. Many of these large programmes are being developed with limited success, as they have generally been based on assumptions that do not correspond to reality, compromising the economic return that allows long term activity. In this scenario a new challenge emerges, which approaches the sustainability of large programmes. It is argued that the main cause of unprofitability is the unexpected high cost of the operation and maintenance of the solar systems. In fact, the lack of a paradigm in decentralized rural services has led to many private companies to carry out decentralized electrification programmes blindly. Issues such as the operation and maintenance cost structure or the reliability of the solar home system components have still not been characterized. This situation does not allow optimized maintenance structure to be designed to assure the sustainability and profitability of the operation and maintenance service. This PhD thesis aims to respond to these needs. Several studies have been carried out based on a real and large photovoltaic rural electrification programme carried out in Morocco with more than 13,000 solar home systems. An in‐depth reliability assessment has been made from a 5‐year maintenance database with more than 80,000 maintenance inputs. The results have allowed us to establish the real reliability functions, the failure rate and the main time to failure of the main components of the system, reporting these findings for the first time in the field of rural electrification. Both in‐field experiments on the capacity degradation of batteries and power degradation of photovoltaic modules have been carried out. During the experiments both samples of batteries and modules were operating under real conditions integrated into the solar home systems of the Moroccan programme. In the case of the batteries, the results have enabled us to obtain a proposal of definition of death of batteries in rural electrification. A cost assessment of the Moroccan experience based on a 5‐year accounting database has been carried out to characterize the cost structure of the programme. The results have allowed the major costs of the photovoltaic electrification to be defined. The overall cost ratio per installed system has been calculated together with the necessary fees that users would have to pay to make the operation and maintenance affordable. Finally, a mathematical optimization model has been proposed to design maintenance structures based on the previous study results. The tool has been applied to the Moroccan programme with the aim of validating the model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Arch bridge structural solution has been known for centuries, in fact the simple nature of arch that require low tension and shear strength was an advantage as the simple materials like stone and brick were the only option back in ancient centuries. By the pass of time especially after industrial revolution, the new materials were adopted in construction of arch bridges to reach longer spans. Nowadays one long span arch bridge is made of steel, concrete or combination of these two as "CFST", as the result of using these high strength materials, very long spans can be achieved. The current record for longest arch belongs to Chaotianmen bridge over Yangtze river in China with 552 meters span made of steel and the longest reinforced concrete type is Wanxian bridge which also cross the Yangtze river through a 420 meters span. Today the designer is no longer limited by span length as long as arch bridge is the most applicable solution among other approaches, i.e. cable stayed and suspended bridges are more reasonable if very long span is desired. Like any super structure, the economical and architectural aspects in construction of a bridge is extremely important, in other words, as a narrower bridge has better appearance, it also require smaller volume of material which make the design more economical. Design of such bridge, beside the high strength materials, requires precise structural analysis approaches capable of integrating the combination of material behaviour and complex geometry of structure and various types of loads which may be applied to bridge during its service life. Depend on the design strategy, analysis may only evaluates the linear elastic behaviour of structure or consider the nonlinear properties as well. Although most of structures in the past were designed to act in their elastic range, the rapid increase in computational capacity allow us to consider different sources of nonlinearities in order to achieve a more realistic evaluations where the dynamic behaviour of bridge is important especially in seismic zones where large movements may occur or structure experience P - _ effect during the earthquake. The above mentioned type of analysis is computationally expensive and very time consuming. In recent years, several methods were proposed in order to resolve this problem. Discussion of recent developments on these methods and their application on long span concrete arch bridges is the main goal of this research. Accordingly available long span concrete arch bridges have been studied to gather the critical information about their geometrical aspects and properties of their materials. Based on concluded information, several concrete arch bridges were designed for further studies. The main span of these bridges range from 100 to 400 meters. The Structural analysis methods implemented in in this study are as following: Elastic Analysis: Direct Response History Analysis (DRHA): This method solves the direct equation of motion over time history of applied acceleration or imposed load in linear elastic range. Modal Response History Analysis (MRHA): Similar to DRHA, this method is also based on time history, but the equation of motion is simplified to single degree of freedom system and calculates the response of each mode independently. Performing this analysis require less time than DRHA. Modal Response Spectrum Analysis (MRSA): As it is obvious from its name, this method calculates the peak response of structure for each mode and combine them using modal combination rules based on the introduced spectra of ground motion. This method is expected to be fastest among Elastic analysis. Inelastic Analysis: Nonlinear Response History Analysis (NL-RHA): The most accurate strategy to address significant nonlinearities in structural dynamics is undoubtedly the nonlinear response history analysis which is similar to DRHA but extended to inelastic range by updating the stiffness matrix for every iteration. This onerous task, clearly increase the computational cost especially for unsymmetrical buildings that requires to be analyzed in a full 3D model for taking the torsional effects in to consideration. Modal Pushover Analysis (MPA): The Modal Pushover Analysis is basically the MRHA but extended to inelastic stage. After all, the MRHA cannot solve the system of dynamics because the resisting force fs(u; u_ ) is unknown for inelastic stage. The solution of MPA for this obstacle is using the previously recorded fs to evaluate system of dynamics. Extended Modal Pushover Analysis (EMPA): Expanded Modal pushover is a one of very recent proposed methods which evaluates response of structure under multi-directional excitation using the modal pushover analysis strategy. In one specific mode,the original pushover neglect the contribution of the directions different than characteristic one, this is reasonable in regular symmetric building but a structure with complex shape like long span arch bridges may go through strong modal coupling. This method intend to consider modal coupling while it take same time of computation as MPA. Coupled Nonlinear Static Pushover Analysis (CNSP): The EMPA includes the contribution of non-characteristic direction to the formal MPA procedure. However the static pushovers in EMPA are performed individually for every mode, accordingly the resulted values from different modes can be combined but this is only valid in elastic phase; as soon as any element in structure starts yielding the neutral axis of that section is no longer fixed for both response during the earthquake, meaning the longitudinal deflection unavoidably affect the transverse one or vice versa. To overcome this drawback, the CNSP suggests executing pushover analysis for governing modes of each direction at the same time. This strategy is estimated to be more accurate than MPA and EMPA, moreover the calculation time is reduced because only one pushover analysis is required. Regardless of the strategy, the accuracy of structural analysis is highly dependent on modelling and numerical integration approaches used in evaluation of each method. Therefore the widely used Finite Element Method is implemented in process of all analysis performed in this research. In order to address the study, chapter 2, starts with gathered information about constructed long span arch bridges, this chapter continuous with geometrical and material definition of new models. Chapter 3 provides the detailed information about structural analysis strategies; furthermore the step by step description of procedure of all methods is available in Appendix A. The document ends with the description of results and conclusion of chapter 4.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We analyze a simple model of the heat transfer to and from a small satellite orbiting round a solar system planet. Our approach considers the satellite isothermal, with external heat input from the environment and from internal energy dissipation, and output to the environment as black-body radiation. The resulting nonlinear ordinary differential equation for the satellite’s temperature is analyzed by qualitative, perturbation and numerical methods, which prove that the temperature approaches a periodic pattern (attracting limit cycle). This approach can occur in two ways, according to the values of the parameters: (i) a slow decay towards the limit cycle over a time longer than the period, or (ii) a fast decay towards the limit cycle over a time shorter than the period. In the first case, an exactly soluble average equation is valid. We discuss the consequences of our model for the thermal stability of satellites.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The purpose of this Project is, first and foremost, to disclose the topic of nonlinear vibrations and oscillations in mechanical systems and, namely, nonlinear normal modes NNMs to a greater audience of researchers and technicians. To do so, first of all, the dynamical behavior and properties of nonlinear mechanical systems is outlined from the analysis of a pair of exemplary models with the harmonic balanced method. The conclusions drawn are contrasted with the Linear Vibration Theory. Then, it is argued how the nonlinear normal modes could, in spite of their limitations, predict the frequency response of a mechanical system. After discussing those introductory concepts, I present a Matlab package called 'NNMcont' developed by a group of researchers from the University of Liege. This package allows the analysis of nonlinear normal modes of vibration in a range of mechanical systems as extensions of the linear modes. This package relies on numerical methods and a 'continuation algorithm' for the computation of the nonlinear normal modes of a conservative mechanical system. In order to prove its functionality, a two degrees of freedom mechanical system with elastic nonlinearities is analized. This model comprises a mass suspended on a foundation by means of a spring-viscous damper mechanism -analogous to a very simplified model of most suspended structures and machines- that has attached a mass damper as a passive vibration control system. The results of the computation are displayed on frequency energy plots showing the NNMs branches along with modal curves and time-series plots for each normal mode. Finally, a critical analysis of the results obtained is carried out with an eye on devising what they can tell the researcher about the dynamical properties of the system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El uso de aritmética de punto fijo es una opción de diseño muy extendida en sistemas con fuertes restricciones de área, consumo o rendimiento. Para producir implementaciones donde los costes se minimicen sin impactar negativamente en la precisión de los resultados debemos llevar a cabo una asignación cuidadosa de anchuras de palabra. Encontrar la combinación óptima de anchuras de palabra en coma fija para un sistema dado es un problema combinatorio NP-hard al que los diseñadores dedican entre el 25 y el 50 % del ciclo de diseño. Las plataformas hardware reconfigurables, como son las FPGAs, también se benefician de las ventajas que ofrece la aritmética de coma fija, ya que éstas compensan las frecuencias de reloj más bajas y el uso más ineficiente del hardware que hacen estas plataformas respecto a los ASICs. A medida que las FPGAs se popularizan para su uso en computación científica los diseños aumentan de tamaño y complejidad hasta llegar al punto en que no pueden ser manejados eficientemente por las técnicas actuales de modelado de señal y ruido de cuantificación y de optimización de anchura de palabra. En esta Tesis Doctoral exploramos distintos aspectos del problema de la cuantificación y presentamos nuevas metodologías para cada uno de ellos: Las técnicas basadas en extensiones de intervalos han permitido obtener modelos de propagación de señal y ruido de cuantificación muy precisos en sistemas con operaciones no lineales. Nosotros llevamos esta aproximación un paso más allá introduciendo elementos de Multi-Element Generalized Polynomial Chaos (ME-gPC) y combinándolos con una técnica moderna basada en Modified Affine Arithmetic (MAA) estadístico para así modelar sistemas que contienen estructuras de control de flujo. Nuestra metodología genera los distintos caminos de ejecución automáticamente, determina las regiones del dominio de entrada que ejercitarán cada uno de ellos y extrae los momentos estadísticos del sistema a partir de dichas soluciones parciales. Utilizamos esta técnica para estimar tanto el rango dinámico como el ruido de redondeo en sistemas con las ya mencionadas estructuras de control de flujo y mostramos la precisión de nuestra aproximación, que en determinados casos de uso con operadores no lineales llega a tener tan solo una desviación del 0.04% con respecto a los valores de referencia obtenidos mediante simulación. Un inconveniente conocido de las técnicas basadas en extensiones de intervalos es la explosión combinacional de términos a medida que el tamaño de los sistemas a estudiar crece, lo cual conlleva problemas de escalabilidad. Para afrontar este problema presen tamos una técnica de inyección de ruidos agrupados que hace grupos con las señales del sistema, introduce las fuentes de ruido para cada uno de los grupos por separado y finalmente combina los resultados de cada uno de ellos. De esta forma, el número de fuentes de ruido queda controlado en cada momento y, debido a ello, la explosión combinatoria se minimiza. También presentamos un algoritmo de particionado multi-vía destinado a minimizar la desviación de los resultados a causa de la pérdida de correlación entre términos de ruido con el objetivo de mantener los resultados tan precisos como sea posible. La presente Tesis Doctoral también aborda el desarrollo de metodologías de optimización de anchura de palabra basadas en simulaciones de Monte-Cario que se ejecuten en tiempos razonables. Para ello presentamos dos nuevas técnicas que exploran la reducción del tiempo de ejecución desde distintos ángulos: En primer lugar, el método interpolativo aplica un interpolador sencillo pero preciso para estimar la sensibilidad de cada señal, y que es usado después durante la etapa de optimización. En segundo lugar, el método incremental gira en torno al hecho de que, aunque es estrictamente necesario mantener un intervalo de confianza dado para los resultados finales de nuestra búsqueda, podemos emplear niveles de confianza más relajados, lo cual deriva en un menor número de pruebas por simulación, en las etapas iniciales de la búsqueda, cuando todavía estamos lejos de las soluciones optimizadas. Mediante estas dos aproximaciones demostramos que podemos acelerar el tiempo de ejecución de los algoritmos clásicos de búsqueda voraz en factores de hasta x240 para problemas de tamaño pequeño/mediano. Finalmente, este libro presenta HOPLITE, una infraestructura de cuantificación automatizada, flexible y modular que incluye la implementación de las técnicas anteriores y se proporciona de forma pública. Su objetivo es ofrecer a desabolladores e investigadores un entorno común para prototipar y verificar nuevas metodologías de cuantificación de forma sencilla. Describimos el flujo de trabajo, justificamos las decisiones de diseño tomadas, explicamos su API pública y hacemos una demostración paso a paso de su funcionamiento. Además mostramos, a través de un ejemplo sencillo, la forma en que conectar nuevas extensiones a la herramienta con las interfaces ya existentes para poder así expandir y mejorar las capacidades de HOPLITE. ABSTRACT Using fixed-point arithmetic is one of the most common design choices for systems where area, power or throughput are heavily constrained. In order to produce implementations where the cost is minimized without negatively impacting the accuracy of the results, a careful assignment of word-lengths is required. The problem of finding the optimal combination of fixed-point word-lengths for a given system is a combinatorial NP-hard problem to which developers devote between 25 and 50% of the design-cycle time. Reconfigurable hardware platforms such as FPGAs also benefit of the advantages of fixed-point arithmetic, as it compensates for the slower clock frequencies and less efficient area utilization of the hardware platform with respect to ASICs. As FPGAs become commonly used for scientific computation, designs constantly grow larger and more complex, up to the point where they cannot be handled efficiently by current signal and quantization noise modelling and word-length optimization methodologies. In this Ph.D. Thesis we explore different aspects of the quantization problem and we present new methodologies for each of them: The techniques based on extensions of intervals have allowed to obtain accurate models of the signal and quantization noise propagation in systems with non-linear operations. We take this approach a step further by introducing elements of MultiElement Generalized Polynomial Chaos (ME-gPC) and combining them with an stateof- the-art Statistical Modified Affine Arithmetic (MAA) based methodology in order to model systems that contain control-flow structures. Our methodology produces the different execution paths automatically, determines the regions of the input domain that will exercise them, and extracts the system statistical moments from the partial results. We use this technique to estimate both the dynamic range and the round-off noise in systems with the aforementioned control-flow structures. We show the good accuracy of our approach, which in some case studies with non-linear operators shows a 0.04 % deviation respect to the simulation-based reference values. A known drawback of the techniques based on extensions of intervals is the combinatorial explosion of terms as the size of the targeted systems grows, which leads to scalability problems. To address this issue we present a clustered noise injection technique that groups the signals in the system, introduces the noise terms in each group independently and then combines the results at the end. In this way, the number of noise sources in the system at a given time is controlled and, because of this, the combinato rial explosion is minimized. We also present a multi-way partitioning algorithm aimed at minimizing the deviation of the results due to the loss of correlation between noise terms, in order to keep the results as accurate as possible. This Ph.D. Thesis also covers the development of methodologies for word-length optimization based on Monte-Carlo simulations in reasonable times. We do so by presenting two novel techniques that explore the reduction of the execution times approaching the problem in two different ways: First, the interpolative method applies a simple but precise interpolator to estimate the sensitivity of each signal, which is later used to guide the optimization effort. Second, the incremental method revolves on the fact that, although we strictly need to guarantee a certain confidence level in the simulations for the final results of the optimization process, we can do it with more relaxed levels, which in turn implies using a considerably smaller amount of samples, in the initial stages of the process, when we are still far from the optimized solution. Through these two approaches we demonstrate that the execution time of classical greedy techniques can be accelerated by factors of up to ×240 for small/medium sized problems. Finally, this book introduces HOPLITE, an automated, flexible and modular framework for quantization that includes the implementation of the previous techniques and is provided for public access. The aim is to offer a common ground for developers and researches for prototyping and verifying new techniques for system modelling and word-length optimization easily. We describe its work flow, justifying the taken design decisions, explain its public API and we do a step-by-step demonstration of its execution. We also show, through an example, the way new extensions to the flow should be connected to the existing interfaces in order to expand and improve the capabilities of HOPLITE.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

ACKNOWLEDGMENTS MW and RVD have been supported by the German Federal Ministry for Education and Research (BMBF) via the Young Investigators Group CoSy-CC2 (grant no. 01LN1306A). JFD thanks the Stordalen Foundation and BMBF (project GLUES) for financial support. JK acknowledges the IRTG 1740 funded by DFG and FAPESP. MT Gastner is acknowledged for providing his data on the airline, interstate, and Internet network. P Menck thankfully provided his data on the Scandinavian power grid. We thank S Willner on behalf of the entire zeean team for providing the data on the world trade network. All computations have been performed using the Python package pyunicorn [41] that is available at https://github.com/pik-copan/pyunicorn.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Acknowledgements The first author has been supported by a Georg Forster Research Fellowship granted by the Alexander von Humboldt Foundation, Germany

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Mathematical analysis of the subthreshold oscillatory properties of inferior olivary neurons in vitro indicates that the oscillation is nonlinear and supports low dimensional chaotic dynamics. This property leads to the generation of complex functional states that can be attained rapidly via phase coherence that conform to the category of “generalized synchronization.” Functionally, this translates into neuronal ensemble properties that can support maximum functional permissiveness and that rapidly can transform into robustly determined multicellular coherence.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Nonlinear analyses of infant heart rhythms reveal a marked rise in the complexity of the electrocardiogram with maturation. We find that normal mature infants (gestation greater than or equal to 35 weeks) have complex and distinctly nonlinear heart rhythms (consistent with recent reports for healthy adults) but that such nonlinearity is lacking in preterm infants (gestation > or = to 27 weeks) where parasympathetic-sympathetic interaction and function are presumed to be less well developed. Our study further shows that infants with clinical brain death and those treated with atropine exhibit a similar lack of nonlinear feedback control. These three lines of evidence support the hypothesis championed by Goldberger et al. [Goldberger, A.L., Rigney, D.R. & West, B.J. (1990) Sci. Am. 262, 43-49] that autonomic nervous system control underlies the nonlinearity and possible chaos of normal heart rhythms. This report demonstrates the acquisition of nonlinear heart rate dynamics and possible chaos in developing human infants and its loss in brain death and with the administration of atropine. It parallels earlier work documenting changes in the variability of heart rhythms in each of these cases and suggests that nonlinearity may provide additional power in characterizing physiological states.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the analysis of heart rate variability (HRV) are used temporal series that contains the distances between successive heartbeats in order to assess autonomic regulation of the cardiovascular system. These series are obtained from the electrocardiogram (ECG) signal analysis, which can be affected by different types of artifacts leading to incorrect interpretations in the analysis of the HRV signals. Classic approach to deal with these artifacts implies the use of correction methods, some of them based on interpolation, substitution or statistical techniques. However, there are few studies that shows the accuracy and performance of these correction methods on real HRV signals. This study aims to determine the performance of some linear and non-linear correction methods on HRV signals with induced artefacts by quantification of its linear and nonlinear HRV parameters. As part of the methodology, ECG signals of rats measured using the technique of telemetry were used to generate real heart rate variability signals without any error. In these series were simulated missing points (beats) in different quantities in order to emulate a real experimental situation as accurately as possible. In order to compare recovering efficiency, deletion (DEL), linear interpolation (LI), cubic spline interpolation (CI), moving average window (MAW) and nonlinear predictive interpolation (NPI) were used as correction methods for the series with induced artifacts. The accuracy of each correction method was known through the results obtained after the measurement of the mean value of the series (AVNN), standard deviation (SDNN), root mean square error of the differences between successive heartbeats (RMSSD), Lomb\'s periodogram (LSP), Detrended Fluctuation Analysis (DFA), multiscale entropy (MSE) and symbolic dynamics (SD) on each HRV signal with and without artifacts. The results show that, at low levels of missing points the performance of all correction techniques are very similar with very close values for each HRV parameter. However, at higher levels of losses only the NPI method allows to obtain HRV parameters with low error values and low quantity of significant differences in comparison to the values calculated for the same signals without the presence of missing points.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A dedicated mission to investigate exoplanetary atmospheres represents a major milestone in our quest to understand our place in the universe by placing our Solar System in context and by addressing the suitability of planets for the presence of life. EChO—the Exoplanet Characterisation Observatory—is a mission concept specifically geared for this purpose. EChO will provide simultaneous, multi-wavelength spectroscopic observations on a stable platform that will allow very long exposures. The use of passive cooling, few moving parts and well established technology gives a low-risk and potentially long-lived mission. EChO will build on observations by Hubble, Spitzer and ground-based telescopes, which discovered the first molecules and atoms in exoplanetary atmospheres. However, EChO’s configuration and specifications are designed to study a number of systems in a consistent manner that will eliminate the ambiguities affecting prior observations. EChO will simultaneously observe a broad enough spectral region—from the visible to the mid-infrared—to constrain from one single spectrum the temperature structure of the atmosphere, the abundances of the major carbon and oxygen bearing species, the expected photochemically-produced species and magnetospheric signatures. The spectral range and resolution are tailored to separate bands belonging to up to 30 molecules and retrieve the composition and temperature structure of planetary atmospheres. The target list for EChO includes planets ranging from Jupiter-sized with equilibrium temperatures T_ eq up to 2,000 K, to those of a few Earth masses, with T _eq \u223c 300 K. The list will include planets with no Solar System analog, such as the recently discovered planets GJ1214b, whose density lies between that of terrestrial and gaseous planets, or the rocky-iron planet 55 Cnc e, with day-side temperature close to 3,000 K. As the number of detected exoplanets is growing rapidly each year, and the mass and radius of those detected steadily decreases, the target list will be constantly adjusted to include the most interesting systems. We have baselined a dispersive spectrograph design covering continuously the 0.4–16 μm spectral range in 6 channels (1 in the visible, 5 in the InfraRed), which allows the spectral resolution to be adapted from several tens to several hundreds, depending on the target brightness. The instrument will be mounted behind a 1.5 m class telescope, passively cooled to 50 K, with the instrument structure and optics passively cooled to \u223c45 K. EChO will be placed in a grand halo orbit around L2. This orbit, in combination with an optimised thermal shield design, provides a highly stable thermal environment and a high degree of visibility of the sky to observe repeatedly several tens of targets over the year. Both the baseline and alternative designs have been evaluated and no critical items with Technology Readiness Level (TRL) less than 4–5 have been identified. We have also undertaken a first-order cost and development plan analysis and find that EChO is easily compatible with the ESA M-class mission framework.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cold atoms in optical potentials provide an ideal test bed to explore quantum nonlinear dynamics. Atoms are prepared in a magneto-optic trap or as a dilute Bose-Einstein condensate and subjected to a far detuned optical standing wave that is modulated. They exhibit a wide range of dynamics, some of which can be explained by classical theory while other aspects show the underlying quantum nature of the system. The atoms have a mixed phase space containing regions of regular motion which appear as distinct peaks in the atomic momentum distribution embedded in a sea of chaos. The action of the atoms is of the order of Planck's constant, making quantum effects significant. This tutorial presents a detailed description of experiments measuring the evolution of atoms in time-dependent optical potentials. Experimental methods are developed providing means for the observation and selective loading of regions of regular motion. The dependence of the atomic dynamics on the system parameters is explored and distinct changes in the atomic momentum distribution are observed which are explained by the applicable quantum and classical theory. The observation of a bifurcation sequence is reported and explained using classical perturbation theory. Experimental methods for the accurate control of the momentum of an ensemble of atoms are developed. They use phase space resonances and chaotic transients providing novel ensemble atomic beamsplitters. The divergence between quantum and classical nonlinear dynamics is manifest in the experimental observation of dynamical tunnelling. It involves no potential barrier. However a constant of motion other than energy still forbids classically this quantum allowed motion. Atoms coherently tunnel back and forth between their initial state of oscillatory motion and the state 180 out of phase with the initial state.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The multibody dynamics of a satellite in circular orbit, modeled as a central body with two hinge-connected deployable solar panel arrays, is investigated. Typically, the solar panel arrays are deployed in orbit using preloaded torsional springs at the hinges in a near symmetrical accordion manner, to minimize the shock loads at the hinges. There are five degrees of freedom of the interconnected rigid bodies, composed of coupled attitude motions (pitch, yaw and roll) of the central body plus relative rotations of the solar panel arrays. The dynamical equations of motion of the satellite system are derived using Kane's equations. These are then used to investigate the dynamic behavior of the system during solar panel deployment via the 7-8th-order Runge-Kutta integration algorithms and results are compared with approximate analytical solutions. Chaotic attitude motions of the completely deployed satellite in circular orbit under the influence of the gravity-gradient torques are subsequently investigated analytically using Melnikov's method and confirmed via numerical integration. The Hamiltonian equations in terms of Deprit's variables are used to facilitate the analysis. (C) 2003 Published by Elsevier Ltd.