852 resultados para Physical Design
Resumo:
Los arrays de ranuras son sistemas de antennas conocidos desde los años 40, principalmente destinados a formar parte de sistemas rádar de navíos de combate y grandes estaciones terrenas donde el tamaño y el peso no eran altamente restrictivos. Con el paso de los años y debido sobre todo a importantes avances en materiales y métodos de fabricación, el rango de aplicaciones de este tipo de sistemas radiantes creció en gran medida. Desde nuevas tecnologías biomédicas, sistemas anticolisión en automóviles y navegación en aviones, enlaces de comunicaciones de alta tasa binaria y corta distancia e incluso sistemas embarcados en satélites para la transmisión de señal de televisión. Dentro de esta familia de antennas, existen dos grupos que destacan por ser los más utilizados: las antennas de placas paralelas con las ranuras distribuidas de forma circular o espiral y las agrupaciones de arrays lineales construidos sobre guia de onda. Continuando con las tareas de investigación desarrolladas durante los últimos años en el Instituto de Tecnología de Tokyo y en el Grupo de Radiación de la Universidad Politécnica de Madrid, la totalidad de esta tesis se centra en este último grupo, aunque como se verá se separa en gran medida de las técnicas de diseño y metodologías convencionales. Los arrays de ranuras rectas y paralelas al eje de la guía rectangular que las alimenta son, sin ninguna duda, los modelos más empleados debido a la fiabilidad que presentan a altas frecuencias, su capacidad para gestionar grandes cantidades de potencia y la sencillez de su diseño y fabricación. Sin embargo, también presentan desventajas como estrecho ancho de banda en pérdidas de retorno y rápida degradación del diagrama de radiación con la frecuencia. Éstas son debidas a la naturaleza resonante de sus elementos radiantes: al perder la resonancia, el sistema global se desajusta y sus prestaciones degeneran. En arrays bidimensionales de slots rectos, el campo eléctrico queda polarizado sobre el plano transversal a las ranuras, correspondiéndose con el plano de altos lóbulos secundarios. Esta tesis tiene como objetivo el desarrollo de un método sistemático de diseño de arrays de ranuras inclinadas y desplazadas del centro (en lo sucesivo “ranuras compuestas”), definido en 1971 como uno de los desafíos a superar dentro del mundo del diseño de antennas. La técnica empleada se basa en el Método de los Momentos, la Teoría de Circuitos y la Teoría de Conexión Aleatoria de Matrices de Dispersión. Al tratarse de un método circuital, la primera parte de la tesis se corresponde con el estudio de la aplicabilidad de las redes equivalentes fundamentales, su capacidad para recrear fenómenos físicos de la ranura, las limitaciones y ventajas que presentan para caracterizar las diferentes configuraciones de slot compuesto. Se profundiza en las diferencias entre las redes en T y en ! y se condiciona la selección de una u otra dependiendo del tipo de elemento radiante. Una vez seleccionado el tipo de red a emplear en el diseño del sistema, se ha desarrollado un algoritmo de cascadeo progresivo desde el puerto alimentador hacia el cortocircuito que termina el modelo. Este algoritmo es independiente del número de elementos, la frecuencia central de funcionamiento, del ángulo de inclinación de las ranuras y de la red equivalente seleccionada (en T o en !). Se basa en definir el diseño del array como un Problema de Satisfacción de Condiciones (en inglés, Constraint Satisfaction Problem) que se resuelve por un método de Búsqueda en Retroceso (Backtracking algorithm). Como resultado devuelve un circuito equivalente del array completo adaptado a su entrada y cuyos elementos consumen una potencia acorde a una distribución de amplitud dada para el array. En toda agrupación de antennas, el acoplo mutuo entre elementos a través del campo radiado representa uno de los principales problemas para el ingeniero y sus efectos perjudican a las prestaciones globales del sistema, tanto en adaptación como en capacidad de radiación. El empleo de circuito equivalente se descartó por la dificultad que suponía la caracterización de estos efectos y su inclusión en la etapa de diseño. En esta tesis doctoral el acoplo también se ha modelado como una red equivalente cuyos elementos son transformadores ideales y admitancias, conectada al conjunto de redes equivalentes que representa el array. Al comparar los resultados estimados en términos de pérdidas de retorno y radiación con aquellos obtenidos a partir de programas comerciales populares como CST Microwave Studio se confirma la validez del método aquí propuesto, el primer método de diseño sistemático de arrays de ranuras compuestos alimentados por guía de onda rectangular. Al tratarse de ranuras no resonantes, el ancho de banda en pérdidas de retorno es mucho mas amplio que el que presentan arrays de slots rectos. Para arrays bidimensionales, el ángulo de inclinación puede ajustarse de manera que el campo quede polarizado en los planos de bajos lóbulos secundarios. Además de simulaciones se han diseñado, construido y medido dos prototipos centrados en la frecuencia de 12GHz, de seis y diez elementos. Las medidas de pérdidas de retorno y diagrama de radiación revelan excelentes resultados, certificando la bondad del método genuino Method of Moments - Forward Matching Procedure desarrollado a lo largo de esta tésis. Abstract The slot antenna arrays are well known systems from the decade of 40s, mainly intended to be part of radar systems of large warships and terrestrial stations where size and weight were not highly restrictive. Over the years, mainly due to significant advances in materials and manufacturing methods, the range of applications of this type of radiating systems grew significantly. From new biomedical technologies, collision avoidance systems in cars and aircraft navigation, short communication links with high bit transfer rate and even embedded systems in satellites for television broadcast. Within this family of antennas, two groups stand out as being the most frequent in the literature: parallel plate antennas with slots placed in a circular or spiral distribution and clusters of waveguide linear arrays. To continue the vast research work carried out during the last decades in the Tokyo Institute of Technology and in the Radiation Group at the Universidad Politécnica de Madrid, this thesis focuses on the latter group, although it represents a technique that drastically breaks with traditional design methodologies. The arrays of slots straight and parallel to the axis of the feeding rectangular waveguide are without a doubt the most used models because of the reliability that they present at high frequencies, its ability to handle large amounts of power and their simplicity of design and manufacturing. However, there also exist disadvantages as narrow bandwidth in return loss and rapid degradation of the radiation pattern with frequency. These are due to the resonant nature of radiating elements: away from the resonance status, the overall system performance and radiation pattern diminish. For two-dimensional arrays of straight slots, the electric field is polarized transverse to the radiators, corresponding to the plane of high side-lobe level. This thesis aims to develop a systematic method of designing arrays of angled and displaced slots (hereinafter "compound slots"), defined in 1971 as one of the challenges to overcome in the world of antenna design. The used technique is based on the Method of Moments, Circuit Theory and the Theory of Scattering Matrices Connection. Being a circuitry-based method, the first part of this dissertation corresponds to the study of the applicability of the basic equivalent networks, their ability to recreate the slot physical phenomena, their limitations and advantages presented to characterize different compound slot configurations. It delves into the differences of T and ! and determines the selection of the most suitable one depending on the type of radiating element. Once the type of network to be used in the system design is selected, a progressive algorithm called Forward Matching Procedure has been developed to connect the proper equivalent networks from the feeder port to shorted ending. This algorithm is independent of the number of elements, the central operating frequency, the angle of inclination of the slots and selected equivalent network (T or ! networks). It is based on the definition of the array design as a Constraint Satisfaction Problem, solved by means of a Backtracking Algorithm. As a result, the method returns an equivalent circuit of the whole array which is matched at its input port and whose elements consume a power according to a given amplitude distribution for the array. In any group of antennas, the mutual coupling between elements through the radiated field represents one of the biggest problems that the engineer faces and its effects are detrimental to the overall performance of the system, both in radiation capabilities and return loss. The employment of an equivalent circuit for the array design was discarded by some authors because of the difficulty involved in the characterization of the coupling effects and their inclusion in the design stage. In this thesis the coupling has also been modeled as an equivalent network whose elements are ideal transformers and admittances connected to the set of equivalent networks that represent the antennas of the array. By comparing the estimated results in terms of return loss and radiation with those obtained from popular commercial software as CST Microwave Studio, the validity of the proposed method is fully confirmed, representing the first method of systematic design of compound-slot arrays fed by rectangular waveguide. Since these slots do not work under the resonant status, the bandwidth in return loss is much wider than the longitudinal-slot arrays. For the case of two-dimensional arrays, the angle of inclination can be adjusted so that the field is polarized at the low side-lobe level plane. Besides the performed full-wave simulations two prototypes of six and ten elements for the X-band have been designed, built and measured, revealing excellent results and agreement with the expected results. These facts certify that the genuine technique Method of Moments - Matching Forward Procedure developed along this thesis is valid and trustable.
Resumo:
The application of liquid metal technology in fusion devices requires R&D related to many phenomena: interaction between liquid metals and structural material as corrosion, erosion and passivation techniques; magneto-hydrodynamics; free surface fluid-dynamics and any other physical aspect that will be needed for their safe reliable operation. In particular, there is a significant shortage of experimental facilities dedicated to the development of the lithium technology. In the framework of the TECHNOFUSION project, an experimental laboratory devoted to the lithium technology development is proposed, in order to shed some light in the path to IFMIF and the design of chamber's first wall and divertors. The conceptual design foresee a development in two stages, the first one consisting on a material testing loop. The second stage proposes the construction of a mock-up of the IFMIF target that will allow to assess the behaviour of a free-surface lithium target under vacuum conditions. In this paper, such conceptual design is addressed.
Resumo:
A broadband primary standard for thermal noise measurements is presented and its thermal and electromagnetic behaviour is analysed by means of a novel hybrid analytical?numerical simulation methodology. The standard consists of a broadband termination connected to a 3.5mm coaxial airline partially immersed in liquid nitrogen and is designed in order to obtain a low reflectivity and a low uncertainty in the noise temperature. A detailed sensitivity analysis is made in order to highlight the critical characteristics that mostly affect the uncertainty in the noise temperature, and also to determine the manufacturing and operation tolerances for a proper performance in the range 10MHz to 26.5 GHz. Aspects such as the thermal bead design, the level of liquid nitrogen or the uncertainties associated with the temperatures, the physical properties of the materials in the standard and the simulation techniques are discussed.
Resumo:
Embedded context management in resource-constrained devices (e.g. mobile phones, autonomous sensors or smart objects) imposes special requirements in terms of lightness for data modelling and reasoning. In this paper, we explore the state-of-the-art on data representation and reasoning tools for embedded mobile reasoning and propose a light inference system (LIS) aiming at simplifying embedded inference processes offering a set of functionalities to avoid redundancy in context management operations. The system is part of a service-oriented mobile software framework, conceived to facilitate the creation of context-aware applications?it decouples sensor data acquisition and context processing from the application logic. LIS, composed of several modules, encapsulates existing lightweight tools for ontology data management and rule-based reasoning, and it is ready to run on Java-enabled handheld devices. Data management and reasoning processes are designed to handle a general ontology that enables communication among framework components. Both the applications running on top of the framework and the framework components themselves can configure the rule and query sets in order to retrieve the information they need from LIS. In order to test LIS features in a real application scenario, an ?Activity Monitor? has been designed and implemented: a personal health-persuasive application that provides feedback on the user?s lifestyle, combining data from physical and virtual sensors. In this case of use, LIS is used to timely evaluate the user?s activity level, to decide on the convenience of triggering notifications and to determine the best interface or channel to deliver these context-aware alerts.
Resumo:
Este trabajo aborda la metodología seguida para llevar a cabo el proyecto de investigación PRONAF (Clinical Trials Gov.: number NCT01116856.) Background: At present, scientific consensus exists on the multifactorial etiopatogenia of obesity. Both professionals and researchers agree that treatment must also have a multifactorial approach, including diet, physical activity, pharmacology and/or surgical treatment. These two last ones should be reserved for those cases of morbid obesities or in case of failure of the previous ones. The aim of the PRONAF study is to determine what type of exercise combined with caloric restriction is the most appropriate to be included in overweigth and obesity intervention programs, and the aim of this paper is to describe the design and the evaluation methods used to carry out the PRONAF study. Methods/design: One-hundred nineteen overweight (46 males) and 120 obese (61 males) subjects aged 18–50 years were randomly assigned to a strength training group, an endurance training group, a combined strength + endurance training group or a diet and physical activity recommendations group. The intervention period was 22 weeks (in all cases 3 times/wk of training for 22 weeks and 2 weeks for pre and post evaluation). All subjects followed a hypocaloric diet (25-30% less energy intake than the daily energy expenditure estimated by accelerometry). 29–34% of the total energy intake came from fat, 14–20% from protein, and 50–55% from carbohydrates. The mayor outcome variables assesed were, biochemical and inflamatory markers, body composition, energy balance, physical fitness, nutritional habits, genetic profile and quality of life. 180 (75.3%) subjects finished the study, with a dropout rate of 24.7%. Dropout reasons included: personal reasons 17 (28.8%), low adherence to exercise 3 (5.1%), low adherence to diet 6 (10.2%), job change 6 (10.2%), and lost interest 27 (45.8%). Discussion: Feasibility of the study has been proven, with a low dropout rate which corresponds to the estimated sample size. Transfer of knowledge is foreseen as a spin-off, in order that overweight and obese subjects can benefit from the results. The aim is to transfer it to sports centres. Effectiveness on individual health-related parameter in order to determine the most effective training programme will be analysed in forthcoming publications.
Resumo:
En la interacción con el entorno que nos rodea durante nuestra vida diaria (utilizar un cepillo de dientes, abrir puertas, utilizar el teléfono móvil, etc.) y en situaciones profesionales (intervenciones médicas, procesos de producción, etc.), típicamente realizamos manipulaciones avanzadas que incluyen la utilización de los dedos de ambas manos. De esta forma el desarrollo de métodos de interacción háptica multi-dedo dan lugar a interfaces hombre-máquina más naturales y realistas. No obstante, la mayoría de interfaces hápticas disponibles en el mercado están basadas en interacciones con un solo punto de contacto; esto puede ser suficiente para la exploración o palpación del entorno pero no permite la realización de tareas más avanzadas como agarres. En esta tesis, se investiga el diseño mecánico, control y aplicaciones de dispositivos hápticos modulares con capacidad de reflexión de fuerzas en los dedos índice, corazón y pulgar del usuario. El diseño mecánico de la interfaz diseñada, ha sido optimizado con funciones multi-objetivo para conseguir una baja inercia, un amplio espacio de trabajo, alta manipulabilidad y reflexión de fuerzas superiores a 3 N en el espacio de trabajo. El ancho de banda y la rigidez del dispositivo se han evaluado mediante simulación y experimentación real. Una de las áreas más importantes en el diseño de estos dispositivos es el efector final, ya que es la parte que está en contacto con el usuario. Durante este trabajo se ha diseñado un dedal de bajo peso, adaptable a diferentes usuarios que, mediante la incorporación de sensores de contacto, permite estimar fuerzas normales y tangenciales durante la interacción con entornos reales y virtuales. Para el diseño de la arquitectura de control, se estudiaron los principales requisitos para estos dispositivos. Entre estos, cabe destacar la adquisición, procesado e intercambio a través de internet de numerosas señales de control e instrumentación; la computación de equaciones matemáticas incluyendo la cinemática directa e inversa, jacobiana, algoritmos de detección de agarres, etc. Todos estos componentes deben calcularse en tiempo real garantizando una frecuencia mínima de 1 KHz. Además, se describen sistemas para manipulación de precisión virtual y remota; así como el diseño de un método denominado "desacoplo cinemático iterativo" para computar la cinemática inversa de robots y la comparación con otros métodos actuales. Para entender la importancia de la interacción multimodal, se ha llevado a cabo un estudio para comprobar qué estímulos sensoriales se correlacionan con tiempos de respuesta más rápidos y de mayor precisión. Estos experimentos se desarrollaron en colaboración con neurocientíficos del instituto Technion Israel Institute of Technology. Comparando los tiempos de respuesta en la interacción unimodal (auditiva, visual y háptica) con combinaciones bimodales y trimodales de los mismos, se demuestra que el movimiento sincronizado de los dedos para generar respuestas de agarre se basa principalmente en la percepción háptica. La ventaja en el tiempo de procesamiento de los estímulos hápticos, sugiere que los entornos virtuales que incluyen esta componente sensorial generan mejores contingencias motoras y mejoran la credibilidad de los eventos. Se concluye que, los sistemas que incluyen percepción háptica dotan a los usuarios de más tiempo en las etapas cognitivas para rellenar información de forma creativa y formar una experiencia más rica. Una aplicación interesante de los dispositivos hápticos es el diseño de nuevos simuladores que permitan entrenar habilidades manuales en el sector médico. En colaboración con fisioterapeutas de Griffith University en Australia, se desarrolló un simulador que permite realizar ejercicios de rehabilitación de la mano. Las propiedades de rigidez no lineales de la articulación metacarpofalange del dedo índice se estimaron mediante la utilización del efector final diseñado. Estos parámetros, se han implementado en un escenario que simula el comportamiento de la mano humana y que permite la interacción háptica a través de esta interfaz. Las aplicaciones potenciales de este simulador están relacionadas con entrenamiento y educación de estudiantes de fisioterapia. En esta tesis, se han desarrollado nuevos métodos que permiten el control simultáneo de robots y manos robóticas en la interacción con entornos reales. El espacio de trabajo alcanzable por el dispositivo háptico, se extiende mediante el cambio de modo de control automático entre posición y velocidad. Además, estos métodos permiten reconocer el gesto del usuario durante las primeras etapas de aproximación al objeto para su agarre. Mediante experimentos de manipulación avanzada de objetos con un manipulador y diferentes manos robóticas, se muestra que el tiempo en realizar una tarea se reduce y que el sistema permite la realización de la tarea con precisión. Este trabajo, es el resultado de una colaboración con investigadores de Harvard BioRobotics Laboratory. ABSTRACT When we interact with the environment in our daily life (using a toothbrush, opening doors, using cell-phones, etc.), or in professional situations (medical interventions, manufacturing processes, etc.) we typically perform dexterous manipulations that involve multiple fingers and palm for both hands. Therefore, multi-Finger haptic methods can provide a realistic and natural human-machine interface to enhance immersion when interacting with simulated or remote environments. Most commercial devices allow haptic interaction with only one contact point, which may be sufficient for some exploration or palpation tasks but are not enough to perform advanced object manipulations such as grasping. In this thesis, I investigate the mechanical design, control and applications of a modular haptic device that can provide force feedback to the index, thumb and middle fingers of the user. The designed mechanical device is optimized with a multi-objective design function to achieve a low inertia, a large workspace, manipulability, and force-feedback of up to 3 N within the workspace; the bandwidth and rigidity for the device is assessed through simulation and real experimentation. One of the most important areas when designing haptic devices is the end-effector, since it is in contact with the user. In this thesis the design and evaluation of a thimble-like, lightweight, user-adaptable, and cost-effective device that incorporates four contact force sensors is described. This design allows estimation of the forces applied by a user during manipulation of virtual and real objects. The design of a real-time, modular control architecture for multi-finger haptic interaction is described. Requirements for control of multi-finger haptic devices are explored. Moreover, a large number of signals have to be acquired, processed, sent over the network and mathematical computations such as device direct and inverse kinematics, jacobian, grasp detection algorithms, etc. have to be calculated in Real Time to assure the required high fidelity for the haptic interaction. The Hardware control architecture has different modules and consists of an FPGA for the low-level controller and a RT controller for managing all the complex calculations (jacobian, kinematics, etc.); this provides a compact and scalable solution for the required high computation capabilities assuring a correct frequency rate for the control loop of 1 kHz. A set-up for dexterous virtual and real manipulation is described. Moreover, a new algorithm named the iterative kinematic decoupling method was implemented to solve the inverse kinematics of a robotic manipulator. In order to understand the importance of multi-modal interaction including haptics, a subject study was carried out to look for sensory stimuli that correlate with fast response time and enhanced accuracy. This experiment was carried out in collaboration with neuro-scientists from Technion Israel Institute of Technology. By comparing the grasping response times in unimodal (auditory, visual, and haptic) events with the response times in events with bimodal and trimodal combinations. It is concluded that in grasping tasks the synchronized motion of the fingers to generate the grasping response relies on haptic cues. This processing-speed advantage of haptic cues suggests that multimodalhaptic virtual environments are superior in generating motor contingencies, enhancing the plausibility of events. Applications that include haptics provide users with more time at the cognitive stages to fill in missing information creatively and form a richer experience. A major application of haptic devices is the design of new simulators to train manual skills for the medical sector. In collaboration with physical therapists from Griffith University in Australia, we developed a simulator to allow hand rehabilitation manipulations. First, the non-linear stiffness properties of the metacarpophalangeal joint of the index finger were estimated by using the designed end-effector; these parameters are implemented in a scenario that simulates the behavior of the human hand and that allows haptic interaction through the designed haptic device. The potential application of this work is related to educational and medical training purposes. In this thesis, new methods to simultaneously control the position and orientation of a robotic manipulator and the grasp of a robotic hand when interacting with large real environments are studied. The reachable workspace is extended by automatically switching between rate and position control modes. Moreover, the human hand gesture is recognized by reading the relative movements of the index, thumb and middle fingers of the user during the early stages of the approximation-to-the-object phase and then mapped to the robotic hand actuators. These methods are validated to perform dexterous manipulation of objects with a robotic manipulator, and different robotic hands. This work is the result of a research collaboration with researchers from the Harvard BioRobotics Laboratory. The developed experiments show that the overall task time is reduced and that the developed methods allow for full dexterity and correct completion of dexterous manipulations.
Resumo:
After being designed, a product has to be manufactured, which means converting concepts and information into a real, physical object. This requires a big amount of resources and a careful planning. The product manufacturing must be designed too, and that is called Industrialization Design. An accepted methodology for this activity is starting defining simple structures and then progressively increasing the detail degree of the manufacturing solution. The impact of decisions taken at first stages of Industrialization Design is remarkable, and software tools to assist designers are required. In this paper a Knowledge Based Application prototype for the Industrialization Design is presented. The application is implemented within the environment CATIA V5/DELMIA. A case study with a simple Product from aerospace sector illustrates the prototype development.
Resumo:
The number of breast cancer survivors increases every year, thanks to the development of new treatments and screening techniques. However, patients present with numerous side effects that may affect their quality of life. Exercise has been demostrated to reduce some of these side effects, but in spite of this, few breast cancer patientes know and follow the exercise recommendations needed to remain healthy. In this review, we describe the differente breast cancer treatments and the related side effects and implications of exercise in relation to these. We propose that exercise could be and integrative complementary intervention to improve physiological, physical and psychological factors that affect survival and quality of life of these patients. For that reason, the main objective of this review is to provide a general overview of exercise benefits in breast cancer patients and recommendations of how to design exercise interventions in patients with different side effects.
Resumo:
Objective: To examine the association of breakfast consumption with objectively measured and self-reported physical activity, sedentary time and physical fitness. Design: The HELENA (Healthy Lifestyle in Europe by Nutrition in Adolescence) Cross-Sectional Study. Breakfast consumption was assessed by two non-consecutive 24 h recalls and by a ‘Food Choices and Preferences’ questionnaire. Physical activity, sedentary time and physical fitness components (cardiorespiratory fitness, muscular fitness and speed/agility) were measured and self-reported. Socio-economic status was assessed by questionnaire. Setting: Ten European cities. Subjects: Adolescents (n 2148; aged 12?5–17?5 years). Results: Breakfast consumption was not associated with measured or self-reported physical activity. However, 24 h recall breakfast consumption was related to measured sedentary time in males and females; although results were not confirmed when using other methods to assess breakfast patterns or sedentary time. Breakfast consumption was not related to muscular fitness and speed/agility in males and females. However, male breakfast consumers had higher cardiorespiratory fitness compared with occasional breakfast consumers and breakfast skippers, while no differences were observed in females. Overall, results were consistent using different methods to assess breakfast consumption or cardiorespiratory fitness (all P#0?005). In addition, both male and female breakfast skippers (assessed by 24 h recall) were less likely to have high measured cardiorespiratory fitness compared with breakfast consumers (OR50?33; 95% CI 0?18, 0?59 and OR50?56; 95 %CI 0?32, 0?98,respectively). Results persisted across methods. Conclusions: Skipping breakfast does not seem to be related to physical activity,sedentary time or muscular fitness and speed/agility as physical fitness components in European adolescents; yet it is associated with both measured and self-reported cardiorespiratory fitness, which extends previous findings.
Resumo:
High-Performance Computing, Cloud computing and next-generation applications such e-Health or Smart Cities have dramatically increased the computational demand of Data Centers. The huge energy consumption, increasing levels of CO2 and the economic costs of these facilities represent a challenge for industry and researchers alike. Recent research trends propose the usage of holistic optimization techniques to jointly minimize Data Center computational and cooling costs from a multilevel perspective. This paper presents an analysis on the parameters needed to integrate the Data Center in a holistic optimization framework and leverages the usage of Cyber-Physical systems to gather workload, server and environmental data via software techniques and by deploying a non-intrusive Wireless Sensor Net- work (WSN). This solution tackles data sampling, retrieval and storage from a reconfigurable perspective, reducing the amount of data generated for optimization by a 68% without information loss, doubling the lifetime of the WSN nodes and allowing runtime energy minimization techniques in a real scenario.
Resumo:
After being designed, a product has to be manufactured, which means converting concepts and information into a real, physical object. This requires a big amount of resources and a careful planning. The product manufacturing must be designed too, and that is called Industrialization Design. An accepted methodology for this activity is starting defining simple structures and then progressively increasing the detail degree of the manufacturing solution. The impact of decisions taken at first stages of Industrialization Design is remarkable, and software tools to assist designers are required. In this paper a Knowledge Based Application prototype for the Industrialization Design is presented. The application is implemented within the environment CATIA V5/DELMIA. A case study with a simple Product from aerospace sector illustrates the prototype development.
Resumo:
La región del espectro electromagnético comprendida entre 100 GHz y 10 THz alberga una gran variedad de aplicaciones en campos tan dispares como la radioastronomía, espectroscopíamolecular, medicina, seguridad, radar, etc. Los principales inconvenientes en el desarrollo de estas aplicaciones son los altos costes de producción de los sistemas trabajando a estas frecuencias, su costoso mantenimiento, gran volumen y baja fiabilidad. Entre las diferentes tecnologías a frecuencias de THz, la tecnología de los diodos Schottky juega un importante papel debido a su madurez y a la sencillez de estos dispositivos. Además, los diodos Schottky pueden operar tanto a temperatura ambiente como a temperaturas criogénicas, con altas eficiencias cuando se usan como multiplicadores y con moderadas temperaturas de ruido en mezcladores. El principal objetivo de esta tesis doctoral es analizar los fenómenos físicos responsables de las características eléctricas y del ruido en los diodos Schottky, así como analizar y diseñar circuitos multiplicadores y mezcladores en bandas milimétricas y submilimétricas. La primera parte de la tesis presenta un análisis de los fenómenos físicos que limitan el comportamiento de los diodos Schottky de GaAs y GaN y de las características del espectro de ruido de estos dispositivos. Para llevar a cabo este análisis, un modelo del diodo basado en la técnica de Monte Carlo se ha considerado como referencia debido a la elevada precisión y fiabilidad de este modelo. Además, el modelo de Monte Carlo permite calcular directamente el espectro de ruido de los diodos sin necesidad de utilizar ningún modelo analítico o empírico. Se han analizado fenómenos físicos como saturación de la velocidad, inercia de los portadores, dependencia de la movilidad electrónica con la longitud de la epicapa, resonancias del plasma y efectos no locales y no estacionarios. También se ha presentado un completo análisis del espectro de ruido para diodos Schottky de GaAs y GaN operando tanto en condiciones estáticas como variables con el tiempo. Los resultados obtenidos en esta parte de la tesis contribuyen a mejorar la comprensión de la respuesta eléctrica y del ruido de los diodos Schottky en condiciones de altas frecuencias y/o altos campos eléctricos. También, estos resultados han ayudado a determinar las limitaciones de modelos numéricos y analíticos usados en el análisis de la respuesta eléctrica y del ruido electrónico en los diodos Schottky. La segunda parte de la tesis está dedicada al análisis de multiplicadores y mezcladores mediante una herramienta de simulación de circuitos basada en la técnica de balance armónico. Diferentes modelos basados en circuitos equivalentes del dispositivo, en las ecuaciones de arrastre-difusión y en la técnica de Monte Carlo se han considerado en este análisis. El modelo de Monte Carlo acoplado a la técnica de balance armónico se ha usado como referencia para evaluar las limitaciones y el rango de validez de modelos basados en circuitos equivalentes y en las ecuaciones de arrastredifusión para el diseño de circuitos multiplicadores y mezcladores. Una notable característica de esta herramienta de simulación es que permite diseñar circuitos Schottky teniendo en cuenta tanto la respuesta eléctrica como el ruido generado en los dispositivos. Los resultados de las simulaciones presentados en esta parte de la tesis, tanto paramultiplicadores comomezcladores, se han comparado con resultados experimentales publicados en la literatura. El simulador que integra el modelo de Monte Carlo con la técnica de balance armónico permite analizar y diseñar circuitos a frecuencias superiores a 1 THz. ABSTRACT The terahertz region of the electromagnetic spectrum(100 GHz-10 THz) presents a wide range of applications such as radio-astronomy, molecular spectroscopy, medicine, security and radar, among others. The main obstacles for the development of these applications are the high production cost of the systems working at these frequencies, highmaintenance, high volume and low reliability. Among the different THz technologies, Schottky technology plays an important rule due to its maturity and the inherent simplicity of these devices. Besides, Schottky diodes can operate at both room and cryogenic temperatures, with high efficiency in multipliers and moderate noise temperature in mixers. This PhD. thesis is mainly concerned with the analysis of the physical processes responsible for the characteristics of the electrical response and noise of Schottky diodes, as well as the analysis and design of frequency multipliers and mixers at millimeter and submillimeter wavelengths. The first part of the thesis deals with the analysis of the physical phenomena limiting the electrical performance of GaAs and GaN Schottky diodes and their noise performance. To carry out this analysis, a Monte Carlo model of the diode has been used as a reference due to the high accuracy and reliability of this diode model at millimeter and submillimter wavelengths. Besides, the Monte Carlo model provides a direct description of the noise spectra of the devices without the necessity of any additional analytical or empirical model. Physical phenomena like velocity saturation, carrier inertia, dependence of the electron mobility on the epilayer length, plasma resonance and nonlocal effects in time and space have been analysed. Also, a complete analysis of the current noise spectra of GaAs and GaN Schottky diodes operating under static and time varying conditions is presented in this part of the thesis. The obtained results provide a better understanding of the electrical and the noise responses of Schottky diodes under high frequency and/or high electric field conditions. Also these results have helped to determine the limitations of numerical and analytical models used in the analysis of the electrical and the noise responses of these devices. The second part of the thesis is devoted to the analysis of frequency multipliers and mixers by means of an in-house circuit simulation tool based on the harmonic balance technique. Different lumped equivalent circuits, drift-diffusion and Monte Carlo models have been considered in this analysis. The Monte Carlo model coupled to the harmonic balance technique has been used as a reference to evaluate the limitations and range of validity of lumped equivalent circuit and driftdiffusion models for the design of frequency multipliers and mixers. A remarkable feature of this reference simulation tool is that it enables the design of Schottky circuits from both electrical and noise considerations. The simulation results presented in this part of the thesis for both multipliers and mixers have been compared with measured results available in the literature. In addition, the Monte Carlo simulation tool allows the analysis and design of circuits above 1 THz.
Resumo:
Linear Fresnel collectors still present a large margin to improve efficiency. Solar fields of this kind installed until current time, both prototypes and commercial plants, are designed with widths and shifts of mirrors that are constant across the solar field. However, the physical processes that limit the width of the mirrors depend on their relative locations to the receiver; the same applies to shading and blocking effects, that oblige to have a minimum shift between mirrors. In this work such phenomena are studied analytically in order to obtain a coherent design, able to improve the efficiency with no increase in cost. A ray tracing simulation along one year has been carried out for a given design, obtaining a moderate increase in radiation collecting efficiency in comparison to conventional designs. Moreover, this analytic theory can guide future designs aiming at fully optimizing linear Fresnel collectors' performance.
Resumo:
The new reactor concepts proposed in the Generation IV International Forum (GIF) are conceived to improve the use of natural resources, reduce the amount of high-level radioactive waste and excel in their reliability and safe operation. Among these novel designs sodium fast reactors (SFRs) stand out due to their technological feasibility as demonstrated in several countries during the last decades. As part of the contribution of EURATOM to GIF the CP-ESFR is a collaborative project with the objective, among others, to perform extensive analysis on safety issues involving renewed SFR demonstrator designs. The verification of computational tools able to simulate the plant behaviour under postulated accidental conditions by code-to-code comparison was identified as a key point to ensure reactor safety. In this line, several organizations employed coupled neutronic and thermal-hydraulic system codes able to simulate complex and specific phenomena involving multi-physics studies adapted to this particular fast reactor technology. In the “Introduction” of this paper the framework of this study is discussed, the second section describes the envisaged plant design and the commonly agreed upon modelling guidelines. The third section presents a comparative analysis of the calculations performed by each organisation applying their models and codes to a common agreed transient with the objective to harmonize the models as well as validating the implementation of all relevant physical phenomena in the different system codes.
Resumo:
INTRODUCCIÓN: El riesgo de padecer enfermedades cardiovasculares y los índices de obesidad infantil han ido en aumento durante los últimos años empobreciendo la salud de la población. La Teoría de Barker relaciona el estado de salud de la madre con el desarrollo fetal, asociando a un deficiente estado físico y hábitos de vida negativos de la mujer embarazada con el aumento del riesgo de padecer cardiopatías en la infancia y adolescencia, así como predisponer al recién nacido a padecer sobrepeso y/u obesidad en su vida posterior. Por otro lado los estudios efectuados sobre ejercicio físico durante el embarazo reportan beneficios para salud materna y fetal. Uno de los parámetros más utilizados para comprobar la salud fetal es su frecuencia cardiaca, mediante la que se comprueba el buen desarrollo del sistema nervioso autónomo. Si se observa este parámetro en presencia de ejercicio materno podría encontrarse una respuesta crónica del corazón fetal al ejercicio materno como consecuencia de una adaptación y mejora en el funcionamiento del sistema nervioso autónomo del feto. De esta forma podría mejorar su salud cardiovascular intrauterina, lo que podría mantenerse en su vida posterior descendiendo el riesgo de padecer enfermedades cardiovasculares en la edad adulta. OBJETIVOS: Conocer la influencia de un programa de ejercicio físico supervisado en la frecuencia cardiaca fetal (FCF) en reposo y después del ejercicio materno en relación con gestantes sedentarias mediante la realización de un protocolo específico. Conocer la influencia de un programa de ejercicio físico en el desarrollo del sistema nervioso autónomo fetal, relacionado con el tiempo de recuperación de la FCF. MATERIAL Y MÉTODO: Se diseñó un ensayo clínico aleatorizado multicéntrico en el que participaron 81 gestantes (GC=38, GE=43). El estudio fue aprobado por el comité ético de los hospitales que participaron en el estudio. Todas las gestantes fueron informadas y firmaron un consentimiento para su participación en el estudio. Las participantes del GE recibieron una intervención basada en un programa de ejercicio físico desarrollado durante la gestación (12-36 semanas de gestación) con una frecuencia de tres veces por semana. Todas las gestantes realizaron un protocolo de medida de la FCF entre las semanas 34-36 de gestación. Dicho protocolo consistía en dos test llevados a cabo caminando a diferentes intensidades (40% y 60% de la frecuencia cardiaca de reserva). De este protocolo se obtuvieron las principales variables de estudio: FCF en reposo, FCF posejercicio al 40 y al 60% de intensidad, tiempo de recuperación de la frecuencia cardiaca fetal en ambos esfuerzos. El material utilizado para la realización del protocolo fue un monitor de frecuencia cardiaca para controlar la frecuencia cardiaca de la gestante y un monitor fetal inalámbrico (telemetría fetal) para registrar el latido fetal durante todo el protocolo. RESULTADOS: No se encontraron diferencias estadísticamente significativas en la FCF en reposo entre grupos (GE=140,88 lat/min vs GC= 141,95 lat/min; p>,05). Se encontraron diferencias estadísticamente significativas en el tiempo de recuperación de la FCF entre los fetos de ambos grupos (GE=135,65 s vs GC=426,11 s esfuerzo al 40%; p<,001); (GE=180,26 s vs GC=565,61 s esfuerzo al 60%; p<,001). Se encontraron diferencias estadísticamente significativas en la FCF posejercicio al 40% (GE=139,93 lat/min vs GC=147,87 lat/min; p<,01). No se encontraron diferencias estadísticamente significativas en la FCF posejercicio al 60% (GE=143,74 lat/min vs GC=148,08 lat/min; p>,05). CONLUSIÓN: El programa de ejercicio físico desarrollado durante la gestación influyó sobre el corazón fetal de los fetos de las gestantes del GE en relación con el tiempo de recuperación de la FCF. Los resultados muestran un posible mejor funcionamiento del sistema nervioso autónomo en fetos de gestantes activas durante el embarazo. ABSTRACT INTRODUCTION: The risk to suffer cardiovascular diseases and childhood obesity index has grown in the last years worsening the health around the population. Barker´s Theory related maternal health with fetal development establishing an association between a poorly physical state and an unhealthy lifestyle in the pregnant woman with the risk to suffer heart disease during childhood and adolescence, childhood overweight and/or obese is related to maternal lifestyle. By the other way researches carried out about physical exercise and pregnancy show benefits in maternal and fetal health. One of the most studied parameters to check fetal health is its heart rate, correct fetal autonomic nervous system development and work is also corroborated by fetal heart rate. Looking at this parameter during maternal exercise a chronic response of fetal heart could be found due to an adaptation and improvement in the working of the autonomic nervous system. Therefore its cardiovascular health could be enhanced during its intrauterine life and maybe it could be maintained in its posterior life descending the risk to suffer cardiovascular diseases in adult life. OBJECTIVES: To know the influence of a supervised physical activity program in the fetal heart rate (FHR) at rest, FHR after maternal exercise related to sedentary pregnant women by a FHR assessment protocol. To know the influence of a physical activity program in the development of the autonomic nervous system related to FHR recovery time. MATERIAL AND METHOD: A multicentric randomized clinical trial was design in which 81 pregnant women participated (CG=38, EG=43). The study was approved by the ethics committee of all of the hospitals participating in the study. All of the participants signed an informed consent for their participation in the study. EG participants received an intervention based on a physical activity program carried out during gestation (12-36 gestation weeks) with a three days a week frequency. All of the participants were tested between 34-36 weeks of gestation by a specific FHR assessment protocol. The mentioned protocol consisted in two test performed walking and at a two different intensities (40% and 60% of the reserve heart rate). From this protocol we obtained the main research variables: FHR at rest, FHR post-exercise at 40% and 60% intensity, and FHR recovery time at both walking test. The material used to perform the protocol were a FH monitor to check maternal HR and a wireless fetal monitor (Telemetry) to register fetal beats during the whole protocol. RESULTS: There were no statistical differences in FHR at rest between groups (EG=140,88 beats/min vs CG= 141,95 beats/min; p>,05). There were statistical differences in FHR recovery time in both walking tests between groups (EG=135,65 s vs CG=426,11 s test at 40% intensity; p<,001); (EG=180,26 s vs CG=565,61 s test at 60% intensity; p<,001). Statistical differences were found in FHR post-exercise at 40% intensity between groups (EG=139,93 beats/min vs CG=147,87 beats/min; p<,01). No statistical differences were found in FHR at rest post-exercise at 60% intensity between groups (EG=143,74 beats/min vs CG=148,08 beats/min; p>,05). CONCLUSIONS: The physical activity program performed during gestation had an influence in fetal heart of the fetus from mother in the EG related to FHR recovery time. These results show a possible enhancement on autonomic nervous system working in fetus from active mothers during gestation.