32 resultados para Hexarotor. Dynamic modeling. Robust backstepping control. EKF Attitude Estimation

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In recent decades, full electric and hybrid electric vehicles have emerged as an alternative to conventional cars due to a range of factors, including environmental and economic aspects. These vehicles are the result of considerable efforts to seek ways of reducing the use of fossil fuel for vehicle propulsion. Sophisticated technologies such as hybrid and electric powertrains require careful study and optimization. Mathematical models play a key role at this point. Currently, many advanced mathematical analysis tools, as well as computer applications have been built for vehicle simulation purposes. Given the great interest of hybrid and electric powertrains, along with the increasing importance of reliable computer-based models, the author decided to integrate both aspects in the research purpose of this work. Furthermore, this is one of the first final degree projects held at the ETSII (Higher Technical School of Industrial Engineers) that covers the study of hybrid and electric propulsion systems. The present project is based on MBS3D 2.0, a specialized software for the dynamic simulation of multibody systems developed at the UPM Institute of Automobile Research (INSIA). Automobiles are a clear example of complex multibody systems, which are present in nearly every field of engineering. The work presented here benefits from the availability of MBS3D software. This program has proven to be a very efficient tool, with a highly developed underlying mathematical formulation. On this basis, the focus of this project is the extension of MBS3D features in order to be able to perform dynamic simulations of hybrid and electric vehicle models. This requires the joint simulation of the mechanical model of the vehicle, together with the model of the hybrid or electric powertrain. These sub-models belong to completely different physical domains. In fact the powertrain consists of energy storage systems, electrical machines and power electronics, connected to purely mechanical components (wheels, suspension, transmission, clutch…). The challenge today is to create a global vehicle model that is valid for computer simulation. Therefore, the main goal of this project is to apply co-simulation methodologies to a comprehensive model of an electric vehicle, where sub-models from different areas of engineering are coupled. The created electric vehicle (EV) model consists of a separately excited DC electric motor, a Li-ion battery pack, a DC/DC chopper converter and a multibody vehicle model. Co-simulation techniques allow car designers to simulate complex vehicle architectures and behaviors, which are usually difficult to implement in a real environment due to safety and/or economic reasons. In addition, multi-domain computational models help to detect the effects of different driving patterns and parameters and improve the models in a fast and effective way. Automotive designers can greatly benefit from a multidisciplinary approach of new hybrid and electric vehicles. In this case, the global electric vehicle model includes an electrical subsystem and a mechanical subsystem. The electrical subsystem consists of three basic components: electric motor, battery pack and power converter. A modular representation is used for building the dynamic model of the vehicle drivetrain. This means that every component of the drivetrain (submodule) is modeled separately and has its own general dynamic model, with clearly defined inputs and outputs. Then, all the particular submodules are assembled according to the drivetrain configuration and, in this way, the power flow across the components is completely determined. Dynamic models of electrical components are often based on equivalent circuits, where Kirchhoff’s voltage and current laws are applied to draw the algebraic and differential equations. Here, Randles circuit is used for dynamic modeling of the battery and the electric motor is modeled through the analysis of the equivalent circuit of a separately excited DC motor, where the power converter is included. The mechanical subsystem is defined by MBS3D equations. These equations consider the position, velocity and acceleration of all the bodies comprising the vehicle multibody system. MBS3D 2.0 is entirely written in MATLAB and the structure of the program has been thoroughly studied and understood by the author. MBS3D software is adapted according to the requirements of the applied co-simulation method. Some of the core functions are modified, such as integrator and graphics, and several auxiliary functions are added in order to compute the mathematical model of the electrical components. By coupling and co-simulating both subsystems, it is possible to evaluate the dynamic interaction among all the components of the drivetrain. ‘Tight-coupling’ method is used to cosimulate the sub-models. This approach integrates all subsystems simultaneously and the results of the integration are exchanged by function-call. This means that the integration is done jointly for the mechanical and the electrical subsystem, under a single integrator and then, the speed of integration is determined by the slower subsystem. Simulations are then used to show the performance of the developed EV model. However, this project focuses more on the validation of the computational and mathematical tool for electric and hybrid vehicle simulation. For this purpose, a detailed study and comparison of different integrators within the MATLAB environment is done. Consequently, the main efforts are directed towards the implementation of co-simulation techniques in MBS3D software. In this regard, it is not intended to create an extremely precise EV model in terms of real vehicle performance, although an acceptable level of accuracy is achieved. The gap between the EV model and the real system is filled, in a way, by introducing the gas and brake pedals input, which reflects the actual driver behavior. This input is included directly in the differential equations of the model, and determines the amount of current provided to the electric motor. For a separately excited DC motor, the rotor current is proportional to the traction torque delivered to the car wheels. Therefore, as it occurs in the case of real vehicle models, the propulsion torque in the mathematical model is controlled through acceleration and brake pedal commands. The designed transmission system also includes a reduction gear that adapts the torque coming for the motor drive and transfers it. The main contribution of this project is, therefore, the implementation of a new calculation path for the wheel torques, based on performance characteristics and outputs of the electric powertrain model. Originally, the wheel traction and braking torques were input to MBS3D through a vector directly computed by the user in a MATLAB script. Now, they are calculated as a function of the motor current which, in turn, depends on the current provided by the battery pack across the DC/DC chopper converter. The motor and battery currents and voltages are the solutions of the electrical ODE (Ordinary Differential Equation) system coupled to the multibody system. Simultaneously, the outputs of MBS3D model are the position, velocity and acceleration of the vehicle at all times. The motor shaft speed is computed from the output vehicle speed considering the wheel radius, the gear reduction ratio and the transmission efficiency. This motor shaft speed, somehow available from MBS3D model, is then introduced in the differential equations corresponding to the electrical subsystem. In this way, MBS3D and the electrical powertrain model are interconnected and both subsystems exchange values resulting as expected with tight-coupling approach.When programming mathematical models of complex systems, code optimization is a key step in the process. A way to improve the overall performance of the integration, making use of C/C++ as an alternative programming language, is described and implemented. Although this entails a higher computational burden, it leads to important advantages regarding cosimulation speed and stability. In order to do this, it is necessary to integrate MATLAB with another integrated development environment (IDE), where C/C++ code can be generated and executed. In this project, C/C++ files are programmed in Microsoft Visual Studio and the interface between both IDEs is created by building C/C++ MEX file functions. These programs contain functions or subroutines that can be dynamically linked and executed from MATLAB. This process achieves reductions in simulation time up to two orders of magnitude. The tests performed with different integrators, also reveal the stiff character of the differential equations corresponding to the electrical subsystem, and allow the improvement of the cosimulation process. When varying the parameters of the integration and/or the initial conditions of the problem, the solutions of the system of equations show better dynamic response and stability, depending on the integrator used. Several integrators, with variable and non-variable step-size, and for stiff and non-stiff problems are applied to the coupled ODE system. Then, the results are analyzed, compared and discussed. From all the above, the project can be divided into four main parts: 1. Creation of the equation-based electric vehicle model; 2. Programming, simulation and adjustment of the electric vehicle model; 3. Application of co-simulation methodologies to MBS3D and the electric powertrain subsystem; and 4. Code optimization and study of different integrators. Additionally, in order to deeply understand the context of the project, the first chapters include an introduction to basic vehicle dynamics, current classification of hybrid and electric vehicles and an explanation of the involved technologies such as brake energy regeneration, electric and non-electric propulsion systems for EVs and HEVs (hybrid electric vehicles) and their control strategies. Later, the problem of dynamic modeling of hybrid and electric vehicles is discussed. The integrated development environment and the simulation tool are also briefly described. The core chapters include an explanation of the major co-simulation methodologies and how they have been programmed and applied to the electric powertrain model together with the multibody system dynamic model. Finally, the last chapters summarize the main results and conclusions of the project and propose further research topics. In conclusion, co-simulation methodologies are applicable within the integrated development environments MATLAB and Visual Studio, and the simulation tool MBS3D 2.0, where equation-based models of multidisciplinary subsystems, consisting of mechanical and electrical components, are coupled and integrated in a very efficient way.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper, a system that allows applying precision agriculture techniques is described. The application is based on the deployment of a team of unmanned aerial vehicles that are able to take georeferenced pictures in order to create a full map by applying mosaicking procedures for postprocessing. The main contribution of this work is practical experimentation with an integrated tool. Contributions in different fields are also reported. Among them is a new one-phase automatic task partitioning manager, which is based on negotiation among the aerial vehicles, considering their state and capabilities. Once the individual tasks are assigned, an optimal path planning algorithm is in charge of determining the best path for each vehicle to follow. Also, a robust flight control based on the use of a control law that improves the maneuverability of the quadrotors has been designed. A set of field tests was performed in order to analyze all the capabilities of the system, from task negotiations to final performance. These experiments also allowed testing control robustness under different weather conditions.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Esta tesis realiza una contribución metodológica al problema de la gestión óptima de embalses hidroeléctricos durante eventos de avenidas, considerando un enfoque estocástico y multiobjetivo. Para ello se propone una metodología de evaluación de estrategias de laminación en un contexto probabilístico y multiobjetivo. Además se desarrolla un entorno dinámico de laminación en tiempo real con pronósticos que combina un modelo de optimización y algoritmos de simulación. Estas herramientas asisten a los gestores de las presas en la toma de decisión respecto de cuál es la operación más adecuada del embalse. Luego de una detallada revisión de la bibliografía, se observó que los trabajos en el ámbito de la gestión óptima de embalses en avenidas utilizan, en general, un número reducido de series de caudales o hidrogramas para caracterizar los posibles escenarios. Limitando el funcionamiento satisfactorio de un modelo determinado a situaciones hidrológicas similares. Por otra parte, la mayoría de estudios disponibles en este ámbito abordan el problema de la laminación en embalses multipropósito durante la temporada de avenidas, con varios meses de duración. Estas características difieren de la realidad de la gestión de embalses en España. Con los avances computacionales en materia de gestión de información en tiempo real, se observó una tendencia a la implementación de herramientas de operación en tiempo real con pronósticos para determinar la operación a corto plazo (involucrando el control de avenidas). La metodología de evaluación de estrategias propuesta en esta tesis se basa en determinar el comportamiento de éstas frente a un espectro de avenidas características de la solicitación hidrológica. Con ese fin, se combina un sistema de evaluación mediante indicadores y un entorno de generación estocástica de avenidas, obteniéndose un sistema implícitamente estocástico. El sistema de evaluación consta de tres etapas: caracterización, síntesis y comparación, a fin de poder manejar la compleja estructura de datos resultante y realizar la evaluación. En la primera etapa se definen variables de caracterización, vinculadas a los aspectos que se quieren evaluar (seguridad de la presa, control de inundaciones, generación de energía, etc.). Estas variables caracterizan el comportamiento del modelo para un aspecto y evento determinado. En la segunda etapa, la información de estas variables se sintetiza en un conjunto de indicadores, lo más reducido posible. Finalmente, la comparación se lleva a cabo a partir de la comparación de esos indicadores, bien sea mediante la agregación de dichos objetivos en un indicador único, o bien mediante la aplicación del criterio de dominancia de Pareto obteniéndose un conjunto de soluciones aptas. Esta metodología se aplicó para calibrar los parámetros de un modelo de optimización de embalse en laminación y su comparación con otra regla de operación, mediante el enfoque por agregación. Luego se amplió la metodología para evaluar y comparar reglas de operación existentes para el control de avenidas en embalses hidroeléctricos, utilizando el criterio de dominancia. La versatilidad de la metodología permite otras aplicaciones, tales como la determinación de niveles o volúmenes de seguridad, o la selección de las dimensiones del aliviadero entre varias alternativas. Por su parte, el entorno dinámico de laminación al presentar un enfoque combinado de optimización-simulación, permite aprovechar las ventajas de ambos tipos de modelos, facilitando la interacción con los operadores de las presas. Se mejoran los resultados respecto de los obtenidos con una regla de operación reactiva, aun cuando los pronósticos se desvían considerablemente del hidrograma real. Esto contribuye a reducir la tan mencionada brecha entre el desarrollo teórico y la aplicación práctica asociada a los modelos de gestión óptima de embalses. This thesis presents a methodological contribution to address the problem about how to operate a hydropower reservoir during floods in order to achieve an optimal management considering a multiobjective and stochastic approach. A methodology is proposed to assess the flood control strategies in a multiobjective and probabilistic framework. Additionally, a dynamic flood control environ was developed for real-time operation, including forecasts. This dynamic platform combines simulation and optimization models. These tools may assist to dam managers in the decision making process, regarding the most appropriate reservoir operation to be implemented. After a detailed review of the bibliography, it was observed that most of the existing studies in the sphere of flood control reservoir operation consider a reduce number of hydrographs to characterize the reservoir inflows. Consequently, the adequate functioning of a certain strategy may be limited to similar hydrologic scenarios. In the other hand, most of the works in this context tackle the problem of multipurpose flood control operation considering the entire flood season, lasting some months. These considerations differ from the real necessity in the Spanish context. The implementation of real-time reservoir operation is gaining popularity due to computational advances and improvements in real-time data management. The methodology proposed in this thesis for assessing the strategies is based on determining their behavior for a wide range of floods, which are representative of the hydrological forcing of the dam. An evaluation algorithm is combined with a stochastic flood generation system to obtain an implicit stochastic analysis framework. The evaluation system consists in three stages: characterizing, synthesizing and comparing, in order to handle the complex structure of results and, finally, conduct the evaluation process. In the first stage some characterization variables are defined. These variables should be related to the different aspects to be evaluated (such as dam safety, flood protection, hydropower, etc.). Each of these variables characterizes the behavior of a certain operating strategy for a given aspect and event. In the second stage this information is synthesized obtaining a reduced group of indicators or objective functions. Finally, the indicators are compared by means of an aggregated approach or by a dominance criterion approach. In the first case, a single optimum solution may be achieved. However in the second case, a set of good solutions is obtained. This methodology was applied for calibrating the parameters of a flood control model and to compare it with other operating policy, using an aggregated method. After that, the methodology was extent to assess and compared some existing hydropower reservoir flood control operation, considering the Pareto approach. The versatility of the method allows many other applications, such as determining the safety levels, defining the spillways characteristics, among others. The dynamic framework for flood control combines optimization and simulation models, exploiting the advantages of both techniques. This facilitates the interaction between dam operators and the model. Improvements are obtained applying this system when compared with a reactive operating policy, even if the forecasts deviate significantly from the observed hydrograph. This approach contributes to reduce the gap between the theoretical development in the field of reservoir management and its practical applications.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Ripple-based controls can strongly reduce the required output capacitance in PowerSoC converter thanks to a very fast dynamic response. Unfortunately, these controls are prone to sub-harmonic oscillations and several parameters affect the stability of these systems. This paper derives and validates a simulation-based modeling and stability analysis of a closed-loop V 2Ic control applied to a 5 MHz Buck converter using discrete modeling and Floquet theory to predict stability. This allows the derivation of sensitivity analysis to design robust systems. The work is extended to different V 2 architectures using the same methodology.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

La electrónica digital moderna presenta un desafío a los diseñadores de sistemas de potencia. El creciente alto rendimiento de microprocesadores, FPGAs y ASICs necesitan sistemas de alimentación que cumplan con requirimientos dinámicos y estáticos muy estrictos. Específicamente, estas alimentaciones son convertidores DC-DC de baja tensión y alta corriente que necesitan ser diseñados para tener un pequeño rizado de tensión y una pequeña desviación de tensión de salida bajo transitorios de carga de una alta pendiente. Además, dependiendo de la aplicación, se necesita cumplir con otros requerimientos tal y como proveer a la carga con ”Escalado dinámico de tensión”, donde el convertidor necesitar cambiar su tensión de salida tan rápidamente posible sin sobreoscilaciones, o ”Posicionado Adaptativo de la Tensión” donde la tensión de salida se reduce ligeramente cuanto más grande sea la potencia de salida. Por supuesto, desde el punto de vista de la industria, las figuras de mérito de estos convertidores son el coste, la eficiencia y el tamaño/peso. Idealmente, la industria necesita un convertidor que es más barato, más eficiente, más pequeño y que aún así cumpla con los requerimienos dinámicos de la aplicación. En este contexto, varios enfoques para mejorar la figuras de mérito de estos convertidores se han seguido por la industria y la academia tales como mejorar la topología del convertidor, mejorar la tecnología de semiconducores y mejorar el control. En efecto, el control es una parte fundamental en estas aplicaciones ya que un control muy rápido hace que sea más fácil que una determinada topología cumpla con los estrictos requerimientos dinámicos y, consecuentemente, le da al diseñador un margen de libertar más amplio para mejorar el coste, la eficiencia y/o el tamaño del sistema de potencia. En esta tesis, se investiga cómo diseñar e implementar controles muy rápidos para el convertidor tipo Buck. En esta tesis se demuestra que medir la tensión de salida es todo lo que se necesita para lograr una respuesta casi óptima y se propone una guía de diseño unificada para controles que sólo miden la tensión de salida Luego, para asegurar robustez en controles muy rápidos, se proponen un modelado y un análisis de estabilidad muy precisos de convertidores DC-DC que tienen en cuenta circuitería para sensado y elementos parásitos críticos. También, usando este modelado, se propone una algoritmo de optimización que tiene en cuenta las tolerancias de los componentes y sensados distorsionados. Us ando este algoritmo, se comparan controles muy rápidos del estado del arte y su capacidad para lograr una rápida respuesta dinámica se posiciona según el condensador de salida utilizado. Además, se propone una técnica para mejorar la respuesta dinámica de los controladores. Todas las propuestas se han corroborado por extensas simulaciones y prototipos experimentales. Con todo, esta tesis sirve como una metodología para ingenieros para diseñar e implementar controles rápidos y robustos de convertidores tipo Buck. ABSTRACT Modern digital electronics present a challenge to designers of power systems. The increasingly high-performance of microprocessors, FPGAs (Field Programmable Gate Array) and ASICs (Application-Specific Integrated Circuit) require power supplies to comply with very demanding static and dynamic requirements. Specifically, these power supplies are low-voltage/high-current DC-DC converters that need to be designed to exhibit low voltage ripple and low voltage deviation under high slew-rate load transients. Additionally, depending on the application, other requirements need to be met such as to provide to the load ”Dynamic Voltage Scaling” (DVS), where the converter needs to change the output voltage as fast as possible without underdamping, or ”Adaptive Voltage Positioning” (AVP) where the output voltage is slightly reduced the greater the output power. Of course, from the point of view of the industry, the figures of merit of these converters are the cost, efficiency and size/weight. Ideally, the industry needs a converter that is cheaper, more efficient, smaller and that can still meet the dynamic requirements of the application. In this context, several approaches to improve the figures of merit of these power supplies are followed in the industry and academia such as improving the topology of the converter, improving the semiconductor technology and improving the control. Indeed, the control is a fundamental part in these applications as a very fast control makes it easier for the topology to comply with the strict dynamic requirements and, consequently, gives the designer a larger margin of freedom to improve the cost, efficiency and/or size of the power supply. In this thesis, how to design and implement very fast controls for the Buck converter is investigated. This thesis proves that sensing the output voltage is all that is needed to achieve an almost time-optimal response and a unified design guideline for controls that only sense the output voltage is proposed. Then, in order to assure robustness in very fast controls, a very accurate modeling and stability analysis of DC-DC converters is proposed that takes into account sensing networks and critical parasitic elements. Also, using this modeling approach, an optimization algorithm that takes into account tolerances of components and distorted measurements is proposed. With the use of the algorithm, very fast analog controls of the state-of-the-art are compared and their capabilities to achieve a fast dynamic response are positioned de pending on the output capacitor. Additionally, a technique to improve the dynamic response of controllers is also proposed. All the proposals are corroborated by extensive simulations and experimental prototypes. Overall, this thesis serves as a methodology for engineers to design and implement fast and robust controls for Buck-type converters.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Las futuras misiones para misiles aire-aire operando dentro de la atmósfera requieren la interceptación de blancos a mayores velocidades y más maniobrables, incluyendo los esperados vehículos aéreos de combate no tripulados. La intercepción tiene que lograrse desde cualquier ángulo de lanzamiento. Una de las principales discusiones en la tecnología de misiles en la actualidad es cómo satisfacer estos nuevos requisitos incrementando la capacidad de maniobra del misil y en paralelo, a través de mejoras en los métodos de guiado y control modernos. Esta Tesis aborda estos dos objetivos simultáneamente, al proponer un diseño integrando el guiado y el control de vuelo (autopiloto) y aplicarlo a misiles con control aerodinámico simultáneo en canard y cola. Un primer avance de los resultados obtenidos ha sido publicado recientemente en el Journal of Aerospace Engineering, en Abril de 2015, [Ibarrondo y Sanz-Aranguez, 2015]. El valor del diseño integrado obtenido es que permite al misil cumplir con los requisitos operacionales mencionados empleando únicamente control aerodinámico. El diseño propuesto se compara favorablemente con esquemas más tradicionales, consiguiendo menores distancias de paso al blanco y necesitando de menores esfuerzos de control incluso en presencia de ruidos. En esta Tesis se demostrará cómo la introducción del doble mando, donde tanto el canard como las aletas de cola son móviles, puede mejorar las actuaciones de un misil existente. Comparado con un misil con control en cola, el doble control requiere sólo introducir dos servos adicionales para accionar los canards también en guiñada y cabeceo. La sección de cola será responsable de controlar el misil en balanceo mediante deflexiones diferenciales de los controles. En el caso del doble mando, la complicación añadida es que los vórtices desprendidos de los canards se propagan corriente abajo y pueden incidir sobre las superficies de cola, alterando sus características de control. Como un primer aporte, se ha desarrollado un modelo analítico completo para la aerodinámica no lineal de un misil con doble control, incluyendo la caracterización de este efecto de acoplamiento aerodinámico. Hay dos modos de funcionamiento en picado y guiñada para un misil de doble mando: ”desviación” y ”opuesto”. En modo ”desviación”, los controles actúan en la misma dirección, generando un cambio inmediato en la sustentación y produciendo un movimiento de translación en el misil. La respuesta es rápida, pero en el modo ”desviación” los misiles con doble control pueden tener dificultades para alcanzar grandes ángulos de ataque y altas aceleraciones laterales. Cuando los controles actúan en direcciones opuestas, el misil rota y el ángulo de ataque del fuselaje se incrementa para generar mayores aceleraciones en estado estacionario, aunque el tiempo de respuesta es mayor. Con el modelo aerodinámico completo, es posible obtener una parametrización dependiente de los estados de la dinámica de corto periodo del misil. Debido al efecto de acoplamiento entre los controles, la respuesta en bucle abierto no depende linealmente de los controles. El autopiloto se optimiza para obtener la maniobra requerida por la ley de guiado sin exceder ninguno de los límites aerodinámicos o mecánicos del misil. Una segunda contribución de la tesis es el desarrollo de un autopiloto con múltiples entradas de control y que integra la aerodinámica no lineal, controlando los tres canales de picado, guiñada y cabeceo de forma simultánea. Las ganancias del autopiloto dependen de los estados del misil y se calculan a cada paso de integración mediante la resolución de una ecuación de Riccati de orden 21x21. Las ganancias obtenidas son sub-óptimas, debido a que una solución completa de la ecuación de Hamilton-Jacobi-Bellman no puede obtenerse de manera práctica, y se asumen ciertas simplificaciones. Se incorpora asimismo un mecanismo que permite acelerar la respuesta en caso necesario. Como parte del autopiloto, se define una estrategia para repartir el esfuerzo de control entre el canard y la cola. Esto se consigue mediante un controlador aumentado situado antes del bucle de optimización, que minimiza el esfuerzo total de control para maniobrar. Esta ley de alimentación directa mantiene al misil cerca de sus condiciones de equilibrio, garantizando una respuesta transitoria adecuada. El controlador no lineal elimina la respuesta de fase no-mínima característica de la cola. En esta Tesis se consideran dos diseños para el guiado y control, el control en Doble-Lazo y el control Integrado. En la aproximación de Doble-Lazo, el autopiloto se sitúa dentro de un bucle interior y se diseña independientemente del guiado, que conforma el bucle más exterior del control. Esta estructura asume que existe separación espectral entre los dos, esto es, que los tiempos de respuesta del autopiloto son mucho mayores que los tiempos característicos del guiado. En el estudio se combina el autopiloto desarrollado con una ley de guiado óptimo. Los resultados obtenidos demuestran que se consiguen aumentos muy importantes en las actuaciones frente a misiles con control canard o control en cola, y que la interceptación, cuando se lanza cerca del curso de colisión, se consigue desde cualquier ángulo alrededor del blanco. Para el misil de doble mando, la estrategia óptima resulta en utilizar el modo de control opuesto en la aproximación al blanco y utilizar el modo de desviación justo antes del impacto. Sin embargo la lógica de doble bucle no consigue el impacto cuando hay desviaciones importantes con respecto al curso de colisión. Una de las razones es que parte de la demanda de guiado se pierde, ya que el misil solo es capaz de modificar su aceleración lateral, y no tiene control sobre su aceleración axial, a no ser que incorpore un motor de empuje regulable. La hipótesis de separación mencionada, y que constituye la base del Doble-Bucle, puede no ser aplicable cuando la dinámica del misil es muy alta en las proximidades del blanco. Si se combinan el guiado y el autopiloto en un único bucle, la información de los estados del misil está disponible para el cálculo de la ley de guiado, y puede calcularse la estrategia optima de guiado considerando las capacidades y la actitud del misil. Una tercera contribución de la Tesis es la resolución de este segundo diseño, la integración no lineal del guiado y del autopiloto (IGA) para el misil de doble control. Aproximaciones anteriores en la literatura han planteado este sistema en ejes cuerpo, resultando en un sistema muy inestable debido al bajo amortiguamiento del misil en cabeceo y guiñada. Las simplificaciones que se tomaron también causan que el misil se deslice alrededor del blanco y no consiga la intercepción. En nuestra aproximación el problema se plantea en ejes inerciales y se recurre a la dinámica de los cuaterniones, eliminado estos inconvenientes. No se limita a la dinámica de corto periodo del misil, porque se construye incluyendo de modo explícito la velocidad dentro del bucle de optimización. La formulación resultante en el IGA es independiente de la maniobra del blanco, que sin embargo se ha de incluir en el cálculo del modelo en Doble-bucle. Un típico inconveniente de los sistemas integrados con controlador proporcional, es el problema de las escalas. Los errores de guiado dominan sobre los errores de posición del misil y saturan el controlador, provocando la pérdida del misil. Este problema se ha tratado aquí con un controlador aumentado previo al bucle de optimización, que define un estado de equilibrio local para el sistema integrado, que pasa a actuar como un regulador. Los criterios de actuaciones para el IGA son los mismos que para el sistema de Doble-Bucle. Sin embargo el problema matemático resultante es muy complejo. El problema óptimo para tiempo finito resulta en una ecuación diferencial de Riccati con condiciones terminales, que no puede resolverse. Mediante un cambio de variable y la introducción de una matriz de transición, este problema se transforma en una ecuación diferencial de Lyapunov que puede resolverse mediante métodos numéricos. La solución resultante solo es aplicable en un entorno cercano del blanco. Cuando la distancia entre misil y blanco es mayor, se desarrolla una solución aproximada basada en la solución de una ecuación algebraica de Riccati para cada paso de integración. Los resultados que se han obtenido demuestran, a través de análisis numéricos en distintos escenarios, que la solución integrada es mejor que el sistema de Doble-Bucle. Las trayectorias resultantes son muy distintas. El IGA preserva el guiado del misil y consigue maximizar el uso de la propulsión, consiguiendo la interceptación del blanco en menores tiempos de vuelo. El sistema es capaz de lograr el impacto donde el Doble-Bucle falla, y además requiere un orden menos de magnitud en la cantidad de cálculos necesarios. El efecto de los ruidos radar, datos discretos y errores del radomo se investigan. El IGA es más robusto, resultando menos afectado por perturbaciones que el Doble- Bucle, especialmente porque el núcleo de optimización en el IGA es independiente de la maniobra del blanco. La estimación de la maniobra del blanco es siempre imprecisa y contaminada por ruido, y degrada la precisión de la solución de Doble-Bucle. Finalmente, como una cuarta contribución, se demuestra que el misil con guiado IGA es capaz de realizar una maniobra de defensa contra un blanco que ataque por su cola, sólo con control aerodinámico. Las trayectorias estudiadas consideran una fase pre-programada de alta velocidad de giro, manteniendo siempre el misil dentro de su envuelta de vuelo. Este procedimiento no necesita recurrir a soluciones técnicamente más complejas como el control vectorial del empuje o control por chorro para ejecutar esta maniobra. En todas las demostraciones matemáticas se utiliza el producto de Kronecker como una herramienta practica para manejar las parametrizaciones dependientes de variables, que resultan en matrices de grandes dimensiones. ABSTRACT Future missions for air to air endo-atmospheric missiles require the interception of targets with higher speeds and more maneuverable, including forthcoming unmanned supersonic combat vehicles. The interception will need to be achieved from any angle and off-boresight launch conditions. One of the most significant discussions in missile technology today is how to satisfy these new operational requirements by increasing missile maneuvering capabilities and in parallel, through the development of more advanced guidance and control methods. This Thesis addresses these two objectives by proposing a novel optimal integrated guidance and autopilot design scheme, applicable to more maneuverable missiles with forward and rearward aerodynamic controls. A first insight of these results have been recently published in the Journal of Aerospace Engineering in April 2015, [Ibarrondo and Sanz-Aránguez, 2015]. The value of this integrated solution is that it allows the missile to comply with the aforementioned requirements only by applying aerodynamic control. The proposed design is compared against more traditional guidance and control approaches with positive results, achieving reduced control efforts and lower miss distances with the integrated logic even in the presence of noises. In this Thesis it will be demonstrated how the dual control missile, where canard and tail fins are both movable, can enhance the capabilities of an existing missile airframe. Compared to a tail missile, dual control only requires two additional servos to actuate the canards in pitch and yaw. The tail section will be responsible to maintain the missile stabilized in roll, like in a classic tail missile. The additional complexity is that the vortices shed from the canard propagate downstream where they interact with the tail surfaces, altering the tail expected control characteristics. These aerodynamic phenomena must be properly described, as a preliminary step, with high enough precision for advanced guidance and control studies. As a first contribution we have developed a full analytical model of the nonlinear aerodynamics of a missile with dual control, including the characterization of this cross-control coupling effect. This development has been produced from a theoretical model validated with reliable practical data obtained from wind tunnel experiments available in the scientific literature, complement with computer fluid dynamics and semi-experimental methods. There are two modes of operating a missile with forward and rear controls, ”divert” and ”opposite” modes. In divert mode, controls are deflected in the same direction, generating an increment in direct lift and missile translation. Response is fast, but in this mode, dual control missiles may have difficulties in achieving large angles of attack and high level of lateral accelerations. When controls are deflected in opposite directions (opposite mode) the missile airframe rotates and the body angle of attack is increased to generate greater accelerations in steady-state, although the response time is larger. With the aero-model, a state dependent parametrization of the dual control missile short term dynamics can be obtained. Due to the cross-coupling effect, the open loop dynamics for the dual control missile is not linearly dependent of the fin positions. The short term missile dynamics are blended with the servo system to obtain an extended autopilot model, where the response is linear with the control fins turning rates, that will be the control variables. The flight control loop is optimized to achieve the maneuver required by the guidance law without exceeding any of the missile aerodynamic or mechanical limitations. The specific aero-limitations and relevant performance indicators for the dual control are set as part of the analysis. A second contribution of this Thesis is the development of a step-tracking multi-input autopilot that integrates non-linear aerodynamics. The designed dual control missile autopilot is a full three dimensional autopilot, where roll, pitch and yaw are integrated, calculating command inputs simultaneously. The autopilot control gains are state dependent, and calculated at each integration step solving a matrix Riccati equation of order 21x21. The resulting gains are sub-optimal as a full solution for the Hamilton-Jacobi-Bellman equation cannot be resolved in practical terms and some simplifications are taken. Acceleration mechanisms with an λ-shift is incorporated in the design. As part of the autopilot, a strategy is defined for proper allocation of control effort between canard and tail channels. This is achieved with an augmented feed forward controller that minimizes the total control effort of the missile to maneuver. The feedforward law also maintains the missile near trim conditions, obtaining a well manner response of the missile. The nonlinear controller proves to eliminate the non-minimum phase effect of the tail. Two guidance and control designs have been considered in this Thesis: the Two- Loop and the Integrated approaches. In the Two-Loop approach, the autopilot is placed in an inner loop and designed separately from an outer guidance loop. This structure assumes that spectral separation holds, meaning that the autopilot response times are much higher than the guidance command updates. The developed nonlinear autopilot is linked in the study to an optimal guidance law. Simulations are carried on launching close to collision course against supersonic and highly maneuver targets. Results demonstrate a large boost in performance provided by the dual control versus more traditional canard and tail missiles, where interception with the dual control close to collision course is achieved form 365deg all around the target. It is shown that for the dual control missile the optimal flight strategy results in using opposite control in its approach to target and quick corrections with divert just before impact. However the Two-Loop logic fails to achieve target interception when there are large deviations initially from collision course. One of the reasons is that part of the guidance command is not followed, because the missile is not able to control its axial acceleration without a throttleable engine. Also the separation hypothesis may not be applicable for a high dynamic vehicle like a dual control missile approaching a maneuvering target. If the guidance and autopilot are combined into a single loop, the guidance law will have information of the missile states and could calculate the most optimal approach to the target considering the actual capabilities and attitude of the missile. A third contribution of this Thesis is the resolution of the mentioned second design, the non-linear integrated guidance and autopilot (IGA) problem for the dual control missile. Previous approaches in the literature have posed the problem in body axes, resulting in high unstable behavior due to the low damping of the missile, and have also caused the missile to slide around the target and not actually hitting it. The IGA system is posed here in inertial axes and quaternion dynamics, eliminating these inconveniences. It is not restricted to the missile short term dynamic, and we have explicitly included the missile speed as a state variable. The IGA formulation is also independent of the target maneuver model that is explicitly included in the Two-loop optimal guidance law model. A typical problem of the integrated systems with a proportional control law is the problem of scales. The guidance errors are larger than missile state errors during most of the flight and result in high gains, control saturation and loss of control. It has been addressed here with an integrated feedforward controller that defines a local equilibrium state at each flight point and the controller acts as a regulator to minimize the IGA states excursions versus the defined feedforward state. The performance criteria for the IGA are the same as in the Two-Loop case. However the resulting optimization problem is mathematically very complex. The optimal problem in a finite-time horizon results in an irresoluble state dependent differential Riccati equation with terminal conditions. With a change of variable and the introduction of a transition matrix, the equation is transformed into a time differential Lyapunov equation that can be solved with known numerical methods in real time. This solution results range limited, and applicable when the missile is in a close neighborhood of the target. For larger ranges, an approximate solution is used, obtained from solution of an algebraic matrix Riccati equation at each integration step. The results obtained show, by mean of several comparative numerical tests in diverse homing scenarios, than the integrated approach is a better solution that the Two- Loop scheme. Trajectories obtained are very different in the two cases. The IGA fully preserves the guidance command and it is able to maximize the utilization of the missile propulsion system, achieving interception with lower miss distances and in lower flight times. The IGA can achieve interception against off-boresight targets where the Two- Loop was not able to success. As an additional advantage, the IGA also requires one order of magnitude less calculations than the Two-Loop solution. The effects of radar noises, discrete radar data and radome errors are investigated. IGA solution is robust, and less affected by radar than the Two-Loop, especially because the target maneuvers are not part of the IGA core optimization loop. Estimation of target acceleration is always imprecise and noisy and degrade the performance of the two-Loop solution. The IGA trajectories are such that minimize the impact of radome errors in the guidance loop. Finally, as a fourth contribution, it is demonstrated that the missile with IGA guidance is capable of performing a defense against attacks from its rear hemisphere, as a tail attack, only with aerodynamic control. The studied trajectories have a preprogrammed high rate turn maneuver, maintaining the missile within its controllable envelope. This solution does not recur to more complex features in service today, like vector control of the missile thrust or side thrusters. In all the mathematical treatments and demonstrations, the Kronecker product has been introduced as a practical tool to handle the state dependent parametrizations that have resulted in very high order matrix equations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Recently, vision-based advanced driver-assistance systems (ADAS) have received a new increased interest to enhance driving safety. In particular, due to its high performance–cost ratio, mono-camera systems are arising as the main focus of this field of work. In this paper we present a novel on-board road modeling and vehicle detection system, which is a part of the result of the European I-WAY project. The system relies on a robust estimation of the perspective of the scene, which adapts to the dynamics of the vehicle and generates a stabilized rectified image of the road plane. This rectified plane is used by a recursive Bayesian classi- fier, which classifies pixels as belonging to different classes corresponding to the elements of interest of the scenario. This stage works as an intermediate layer that isolates subsequent modules since it absorbs the inherent variability of the scene. The system has been tested on-road, in different scenarios, including varied illumination and adverse weather conditions, and the results have been proved to be remarkable even for such complex scenarios.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Modeling phase is fundamental both in the analysis process of a dynamic system and the design of a control system. If this phase is in-line is even more critical and the only information of the system comes from input/output data. Some adaptation algorithms for fuzzy system based on extended Kalman filter are presented in this paper, which allows obtaining accurate models without renounce the computational efficiency that characterizes the Kalman filter, and allows its implementation in-line with the process

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In the field of detection and monitoring of dynamic objects in quasi-static scenes, background subtraction techniques where background is modeled at pixel-level, although showing very significant limitations, are extensively used. In this work we propose a novel approach to background modeling that operates at region-level in a wavelet based multi-resolution framework. Based on a segmentation of the background, characterization is made for each region independently as a mixture of K Gaussian modes, considering the model of the approximation and detail coefficients at the different wavelet decomposition levels. Background region characterization is updated along time, and the detection of elements of interest is carried out computing the distance between background region models and those of each incoming image in the sequence. The inclusion of the context in the modeling scheme through each region characterization makes the model robust, being able to support not only gradual illumination and long-term changes, but also sudden illumination changes and the presence of strong shadows in the scene

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In arid countries worldwide, social conflicts between irrigation-based human development and the conservation of aquatic ecosystems are widespread and attract many public debates. This research focuses on the analysis of water and agricultural policies aimed at conserving groundwater resources and maintaining rurallivelihoods in a basin in Spain's central arid region. Intensive groundwater mining for irrigation has caused overexploitation of the basin's large aquifer, the degradation of reputed wetlands and has given rise to notable social conflicts over the years. With the aim of tackling the multifaceted socio-ecological interactions of complex water systems, the methodology used in this study consists in a novel integration into a common platform of an economic optimization model and a hydrology model WEAP (Water Evaluation And Planning system). This robust tool is used to analyze the spatial and temporal effects of different water and agricultural policies under different climate scenarios. It permits the prediction of different climate and policy outcomes across farm types (water stress impacts and adaptation), at basin's level (aquifer recovery), and along the policies’ implementation horizon (short and long run). Results show that the region's current quota-based water policies may contribute to reduce water consumption in the farms but will not be able to recover the aquifer and will inflict income losses to the rural communities. This situation would worsen in case of drought. Economies of scale and technology are evidenced as larger farms with cropping diversification and those equipped with modern irrigation will better adapt to water stress conditions. However, the long-term sustainability of the aquifer and the maintenance of rurallivelihoods will be attained only if additional policy measures are put in place such as the control of illegal abstractions and the establishing of a water bank. Within the policy domain, the research contributes to the new sustainable development strategy of the EU by concluding that, in water-scarce regions, effective integration of water and agricultural policies is essential for achieving the water protection objectives of the EU policies. Therefore, the design and enforcement of well-balanced region-specific polices is a major task faced by policy makers for achieving successful water management that will ensure nature protection and human development at tolerable social costs. From a methodological perspective, this research initiative contributes to better address hydrological questions as well as economic and social issues in complex water and human systems. Its integrated vision provides a valuable illustration to inform water policy and management decisions within contexts of water-related conflicts worldwide.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

La temperatura es una preocupación que juega un papel protagonista en el diseño de circuitos integrados modernos. El importante aumento de las densidades de potencia que conllevan las últimas generaciones tecnológicas ha producido la aparición de gradientes térmicos y puntos calientes durante el funcionamiento normal de los chips. La temperatura tiene un impacto negativo en varios parámetros del circuito integrado como el retardo de las puertas, los gastos de disipación de calor, la fiabilidad, el consumo de energía, etc. Con el fin de luchar contra estos efectos nocivos, la técnicas de gestión dinámica de la temperatura (DTM) adaptan el comportamiento del chip en función en la información que proporciona un sistema de monitorización que mide en tiempo de ejecución la información térmica de la superficie del dado. El campo de la monitorización de la temperatura en el chip ha llamado la atención de la comunidad científica en los últimos años y es el objeto de estudio de esta tesis. Esta tesis aborda la temática de control de la temperatura en el chip desde diferentes perspectivas y niveles, ofreciendo soluciones a algunos de los temas más importantes. Los niveles físico y circuital se cubren con el diseño y la caracterización de dos nuevos sensores de temperatura especialmente diseñados para los propósitos de las técnicas DTM. El primer sensor está basado en un mecanismo que obtiene un pulso de anchura variable dependiente de la relación de las corrientes de fuga con la temperatura. De manera resumida, se carga un nodo del circuito y posteriormente se deja flotando de tal manera que se descarga a través de las corrientes de fugas de un transistor; el tiempo de descarga del nodo es la anchura del pulso. Dado que la anchura del pulso muestra una dependencia exponencial con la temperatura, la conversión a una palabra digital se realiza por medio de un contador logarítmico que realiza tanto la conversión tiempo a digital como la linealización de la salida. La estructura resultante de esta combinación de elementos se implementa en una tecnología de 0,35 _m. El sensor ocupa un área muy reducida, 10.250 nm2, y consume muy poca energía, 1.05-65.5nW a 5 muestras/s, estas cifras superaron todos los trabajos previos en el momento en que se publicó por primera vez y en el momento de la publicación de esta tesis, superan a todas las implementaciones anteriores fabricadas en el mismo nodo tecnológico. En cuanto a la precisión, el sensor ofrece una buena linealidad, incluso sin calibrar; se obtiene un error 3_ de 1,97oC, adecuado para tratar con las aplicaciones de DTM. Como se ha explicado, el sensor es completamente compatible con los procesos de fabricación CMOS, este hecho, junto con sus valores reducidos de área y consumo, lo hacen especialmente adecuado para la integración en un sistema de monitorización de DTM con un conjunto de monitores empotrados distribuidos a través del chip. Las crecientes incertidumbres de proceso asociadas a los últimos nodos tecnológicos comprometen las características de linealidad de nuestra primera propuesta de sensor. Con el objetivo de superar estos problemas, proponemos una nueva técnica para obtener la temperatura. La nueva técnica también está basada en las dependencias térmicas de las corrientes de fuga que se utilizan para descargar un nodo flotante. La novedad es que ahora la medida viene dada por el cociente de dos medidas diferentes, en una de las cuales se altera una característica del transistor de descarga |la tensión de puerta. Este cociente resulta ser muy robusto frente a variaciones de proceso y, además, la linealidad obtenida cumple ampliamente los requisitos impuestos por las políticas DTM |error 3_ de 1,17oC considerando variaciones del proceso y calibrando en dos puntos. La implementación de la parte sensora de esta nueva técnica implica varias consideraciones de diseño, tales como la generación de una referencia de tensión independiente de variaciones de proceso, que se analizan en profundidad en la tesis. Para la conversión tiempo-a-digital, se emplea la misma estructura de digitalización que en el primer sensor. Para la implementación física de la parte de digitalización, se ha construido una biblioteca de células estándar completamente nueva orientada a la reducción de área y consumo. El sensor resultante de la unión de todos los bloques se caracteriza por una energía por muestra ultra baja (48-640 pJ) y un área diminuta de 0,0016 mm2, esta cifra mejora todos los trabajos previos. Para probar esta afirmación, se realiza una comparación exhaustiva con más de 40 propuestas de sensores en la literatura científica. Subiendo el nivel de abstracción al sistema, la tercera contribución se centra en el modelado de un sistema de monitorización que consiste de un conjunto de sensores distribuidos por la superficie del chip. Todos los trabajos anteriores de la literatura tienen como objetivo maximizar la precisión del sistema con el mínimo número de monitores. Como novedad, en nuestra propuesta se introducen nuevos parámetros de calidad aparte del número de sensores, también se considera el consumo de energía, la frecuencia de muestreo, los costes de interconexión y la posibilidad de elegir diferentes tipos de monitores. El modelo se introduce en un algoritmo de recocido simulado que recibe la información térmica de un sistema, sus propiedades físicas, limitaciones de área, potencia e interconexión y una colección de tipos de monitor; el algoritmo proporciona el tipo seleccionado de monitor, el número de monitores, su posición y la velocidad de muestreo _optima. Para probar la validez del algoritmo, se presentan varios casos de estudio para el procesador Alpha 21364 considerando distintas restricciones. En comparación con otros trabajos previos en la literatura, el modelo que aquí se presenta es el más completo. Finalmente, la última contribución se dirige al nivel de red, partiendo de un conjunto de monitores de temperatura de posiciones conocidas, nos concentramos en resolver el problema de la conexión de los sensores de una forma eficiente en área y consumo. Nuestra primera propuesta en este campo es la introducción de un nuevo nivel en la jerarquía de interconexión, el nivel de trillado (o threshing en inglés), entre los monitores y los buses tradicionales de periféricos. En este nuevo nivel se aplica selectividad de datos para reducir la cantidad de información que se envía al controlador central. La idea detrás de este nuevo nivel es que en este tipo de redes la mayoría de los datos es inútil, porque desde el punto de vista del controlador sólo una pequeña cantidad de datos |normalmente sólo los valores extremos| es de interés. Para cubrir el nuevo nivel, proponemos una red de monitorización mono-conexión que se basa en un esquema de señalización en el dominio de tiempo. Este esquema reduce significativamente tanto la actividad de conmutación sobre la conexión como el consumo de energía de la red. Otra ventaja de este esquema es que los datos de los monitores llegan directamente ordenados al controlador. Si este tipo de señalización se aplica a sensores que realizan conversión tiempo-a-digital, se puede obtener compartición de recursos de digitalización tanto en tiempo como en espacio, lo que supone un importante ahorro de área y consumo. Finalmente, se presentan dos prototipos de sistemas de monitorización completos que de manera significativa superan la características de trabajos anteriores en términos de área y, especialmente, consumo de energía. Abstract Temperature is a first class design concern in modern integrated circuits. The important increase in power densities associated to recent technology evolutions has lead to the apparition of thermal gradients and hot spots during run time operation. Temperature impacts several circuit parameters such as speed, cooling budgets, reliability, power consumption, etc. In order to fight against these negative effects, dynamic thermal management (DTM) techniques adapt the behavior of the chip relying on the information of a monitoring system that provides run-time thermal information of the die surface. The field of on-chip temperature monitoring has drawn the attention of the scientific community in the recent years and is the object of study of this thesis. This thesis approaches the matter of on-chip temperature monitoring from different perspectives and levels, providing solutions to some of the most important issues. The physical and circuital levels are covered with the design and characterization of two novel temperature sensors specially tailored for DTM purposes. The first sensor is based upon a mechanism that obtains a pulse with a varying width based on the variations of the leakage currents on the temperature. In a nutshell, a circuit node is charged and subsequently left floating so that it discharges away through the subthreshold currents of a transistor; the time the node takes to discharge is the width of the pulse. Since the width of the pulse displays an exponential dependence on the temperature, the conversion into a digital word is realized by means of a logarithmic counter that performs both the timeto- digital conversion and the linearization of the output. The structure resulting from this combination of elements is implemented in a 0.35_m technology and is characterized by very reduced area, 10250 nm2, and power consumption, 1.05-65.5 nW at 5 samples/s, these figures outperformed all previous works by the time it was first published and still, by the time of the publication of this thesis, they outnumber all previous implementations in the same technology node. Concerning the accuracy, the sensor exhibits good linearity, even without calibration it displays a 3_ error of 1.97oC, appropriate to deal with DTM applications. As explained, the sensor is completely compatible with standard CMOS processes, this fact, along with its tiny area and power overhead, makes it specially suitable for the integration in a DTM monitoring system with a collection of on-chip monitors distributed across the chip. The exacerbated process fluctuations carried along with recent technology nodes jeop-ardize the linearity characteristics of the first sensor. In order to overcome these problems, a new temperature inferring technique is proposed. In this case, we also rely on the thermal dependencies of leakage currents that are used to discharge a floating node, but now, the result comes from the ratio of two different measures, in one of which we alter a characteristic of the discharging transistor |the gate voltage. This ratio proves to be very robust against process variations and displays a more than suficient linearity on the temperature |1.17oC 3_ error considering process variations and performing two-point calibration. The implementation of the sensing part based on this new technique implies several issues, such as the generation of process variations independent voltage reference, that are analyzed in depth in the thesis. In order to perform the time-to-digital conversion, we employ the same digitization structure the former sensor used. A completely new standard cell library targeting low area and power overhead is built from scratch to implement the digitization part. Putting all the pieces together, we achieve a complete sensor system that is characterized by ultra low energy per conversion of 48-640pJ and area of 0.0016mm2, this figure outperforms all previous works. To prove this statement, we perform a thorough comparison with over 40 works from the scientific literature. Moving up to the system level, the third contribution is centered on the modeling of a monitoring system consisting of set of thermal sensors distributed across the chip. All previous works from the literature target maximizing the accuracy of the system with the minimum number of monitors. In contrast, we introduce new metrics of quality apart form just the number of sensors; we consider the power consumption, the sampling frequency, the possibility to consider different types of monitors and the interconnection costs. The model is introduced in a simulated annealing algorithm that receives the thermal information of a system, its physical properties, area, power and interconnection constraints and a collection of monitor types; the algorithm yields the selected type of monitor, the number of monitors, their position and the optimum sampling rate. We test the algorithm with the Alpha 21364 processor under several constraint configurations to prove its validity. When compared to other previous works in the literature, the modeling presented here is the most complete. Finally, the last contribution targets the networking level, given an allocated set of temperature monitors, we focused on solving the problem of connecting them in an efficient way from the area and power perspectives. Our first proposal in this area is the introduction of a new interconnection hierarchy level, the threshing level, in between the monitors and the traditional peripheral buses that applies data selectivity to reduce the amount of information that is sent to the central controller. The idea behind this new level is that in this kind of networks most data are useless because from the controller viewpoint just a small amount of data |normally extreme values| is of interest. To cover the new interconnection level, we propose a single-wire monitoring network based on a time-domain signaling scheme that significantly reduces both the switching activity over the wire and the power consumption of the network. This scheme codes the information in the time domain and allows a straightforward obtention of an ordered list of values from the maximum to the minimum. If the scheme is applied to monitors that employ TDC, digitization resource sharing is achieved, producing an important saving in area and power consumption. Two prototypes of complete monitoring systems are presented, they significantly overcome previous works in terms of area and, specially, power consumption.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In this paper, we use ARIMA modelling to estimate a set of characteristics of a short-term indicator (for example, the index of industrial production), as trends, seasonal variations, cyclical oscillations, unpredictability, deterministic effects (as a strike), etc. Thus for each sector and product (more than 1000), we construct a vector of values corresponding to the above-mentioned characteristics, that can be used for data editing.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Numerical explorations show how the known periodic solutions of the Hill problem are modified in the case of the attitude-orbit coupling that may occur for large satellite structures. We focus on the case in which the elongation is the dominant satellite’s characteristic and find that a rotating structure may remain with its largest dimension in a plane parallel to the plane of the primaries. In this case, the effect produced by the non-negligible physical length is dynamically equivalent to the perturbation produced by an oblate central body on a mass-point satellite. Based on this, it is demonstrated that the attitude-orbital coupling of a long enough body may change the dynamical characteristics of a periodic orbit about the collinear Lagrangian points.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

For the past 20 years, dynamic analysis of shells has been one of the most fascinating fields for research. Using the new light materials the building engineer soon discovered that the subsequent reduction of gravity forces produced not only the desired shape freedom but the appearance of ecologic loads as the first factor of design; loads which present strong random properties and marked dynamic influence. On the other hand, the technological advance in the aeronautical and astronautical field placed the engineers in front of shell structures of nonconventional shape and able to sustain substantialy dynamic loads. The response to the increasingly challenger problems of the last two decades has been very bright; new forms, new materials and new methods of analysis have arosen in the design of off-shore platforms, nuclear vessels, space crafts, etc. Thanks to the intensity of the lived years we have at our disposition a coherent and homogeneous amount of knowledge which enable us to face problems of inconceivable complexity when IASS was founded. The open minded approach to classical problems and the impact of the computer are, probably, important factors in the Renaissance we have enjoyed these years, and a good proof of this are the papers presented to the previous IASS meetings as well as that we are going to consider in this one. Particularly striking is the great number of papers based on a mathematical modeling in front of the meagerness of those treating laboratory experiments on physical models. The universal entering of the computer into almost every phase of our lifes, and the cost of physical models, are –may be- reasons for this lack of experimental methods. Nevertheless they continue offering useful results as are those obtained with the shaking-table in which the computer plays an essential role in the application of loads as well as in the instantaneous treatment of control data. Plates 1 and 2 record the papers presented under dynamic heading, 40% of them are from Japan in good correlation with the relevance that Japanese research has traditionally showed in this area. Also interesting is to find old friends as profesors Tanaka, Nishimura and Kostem who presented valuable papers in previous IASS conferences. As we see there are papers representative of all tendencies, even purely analytical! Better than discuss them in detail, which can be done after the authors presentation, I think we can comment in the general pattern of the dynamical approach are summarized in plate 3.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Numerical explorations show how the known periodic solutions of the Hill problem are modified in the case of the attitude-orbit coupling that may occur for large satellite structures. We focus on the case in which the elongation is the dominant satellite?s characteristic and find that a rotating structure may remain with its largest dimension in a plane parallel to the plane of the primaries. In this case, the effect produced by the non-negligible physical dimension is dynamically equivalent to the perturbation produced by an oblate central body on a masspoint satellite. Based on this, it is demonstrated that the attitude-orbital coupling of a long enough body may change the dynamical characteristics of a periodic orbit about the collinear Lagrangian points.