885 resultados para long-distance propagation simulation
Resumo:
The delay caused by the reflected ray in broadband communication has a great influence on the communications in subway tunnel. This paper presents measurements taken in subway tunnels at 2.4 GHz, with 5 MHz bandwidth. According to propagation characteristics of tunnel, the measurements were carried out with a frequency domain channel sounding technique, in three typical scenarios: line of sight (LOS), Non-line-of-sight (NLOS) and far line of sight (FLOS), which lead to different delay distributions. Firstly IFFT was chosen to get channel impulse response (CIR) h(t) from measured three-dimensional transfer functions. Power delay profile (PDP) was investigated to give an overview of broadband channel model. Thereafter, a long delay caused by the obturation of tunnel is observed and investigated in all the scenarios. The measurements show that the reflection can be greatly remained by the tunnel, which leads to long delay cluster where the reflection, but direct ray, makes the main contribution for radio wave propagation. Four important parameters: distribution of whole PDP power, first peak arriving time, reflection cluster duration and PDP power distribution of reflection cluster were studied to give a detailed description of long delay characteristic in tunnel. This can be used to ensure high capacity communication in tunnels
Resumo:
This paper investigates the propagation of airblast from quarry blasting. Peak overpressure is calculated as a function of blasting parameters (explosive mass per delay and velocity at which the detonation sequence proceeds along the bench) and polar coordinates of the point of interest (distance to the blast and azimuth with respect to the free face of the blast). The model is in the form of the product of a classical scaled distance attenuation law times a directional correction factor. The latter considers the influence of the bench face, and attenuates overpressure at the top level and amplifies it at the bottom. Such factor also accounts for the effect of the delay by amplifying the pressure in the direction of the initiation sequence if the velocity of initiation exceeds half the speed of sound and up to an initiation velocity in the range of the speed of sound. The model has been fitted to an empirical data set composed by 134 airblast records monitored in 47 blasts at two quarries. The measurements were made at distances to the blast less than 450 m. The model is statistically significant and has a determination coefficient of 0.869
Resumo:
Esta tesis analiza los elementos que afectan a la evaluación del rendimiento dentro de la técnica de radiodiagnóstico mediante tomografía por emisión de positrones (PET), centrándose en escáneres preclínicos. Se exploran las posibilidades de los protocolos estándar de evaluación sobre los siguientes aspectos: su uso como herramienta para validar programas de simulación Montecarlo, como método para la comparación de escáneres y su validez en el estudio del efecto sobre la calidad de imagen al utilizar radioisótopos alternativos. Inicialmente se estudian los métodos de evaluación orientados a la validación de simulaciones PET, para ello se presenta el programa GAMOS como entorno de simulación y se muestran los resultados de su validación basada en el estándar NEMA NU 4-2008 para escáneres preclínicos. Esta validación se ha realizado mediante la comparación de los resultados simulados frente a adquisiciones reales en el equipo ClearPET, describiendo la metodología de evaluación y selección de los parámetros NEMA. En este apartado también se mencionan las aportaciones desarrolladas en GAMOS para aplicaciones PET, como la inclusión de herramientas para la reconstrucción de imágenes. Por otro lado, la evaluación NEMA del ClearPET es utilizada para comparar su rendimiento frente a otro escáner preclínico: el sistema rPET-1. Esto supone la primera caracterización NEMA NU 4 completa de ambos equipos; al mismo tiempo que se analiza cómo afectan las importantes diferencias de diseño entre ellos, especialmente el tamaño axial del campo de visión y la configuración de los detectores. El 68Ga es uno de los radioisótopos no convencionales en imagen PET que está experimentando un mayor desarrollo, sin embargo, presenta la desventaja del amplio rango o distancia recorrida por el positrón emitido. Además del rango del positrón, otra propiedad física característica de los radioisótopos PET que puede afectar a la imagen es la emisión de fotones gamma adicionales, tal como le ocurre al isótopo 48V. En esta tesis se evalúan dichos efectos mediante estudios de resolución espacial y calidad de imagen NEMA. Finalmente, se analiza el alcance del protocolo NEMA NU 4-2008 cuando se utiliza para este propósito, adaptándolo a tal fin y proponiendo posibles modificaciones. Abstract This thesis analyzes the factors affecting the performance evaluation in positron emission tomography (PET) imaging, focusing on preclinical scanners. It explores the possibilities of standard protocols of assessment on the following aspects: their use as tools to validate Monte Carlo simulation programs, their usefulness as a method for comparing scanners and their validity in the study of the effect of alternative radioisotopes on image quality. Initially we study the methods of performance evaluation oriented to validate PET simulations. For this we present the GAMOS program as a simulation framework and show the results of its validation based on the standard NEMA NU 4-2008 for preclinical PET scanners. This has been accomplished by comparing simulated results against experimental acquisitions in the ClearPET scanner, describing the methodology for the evaluation and selection of NEMA parameters. This section also mentions the contributions developed in GAMOS for PET applications, such as the inclusion of tools for image reconstruction. Furthermore, the evaluation of the ClearPET scanner is used to compare its performance against another preclinical scanner, specifically the rPET-1 system. This is the first complete NEMA NU 4 based characterization study of both systems. At the same time we analyze how do the significant design differences of these two systems, especially the size of the axial field of view and the detectors configuration affect their performance characteristics. 68Ga is one of the unconventional radioisotopes in PET imaging the use of which is currently significantly increasing; however, it presents the disadvantage of the long positron range (distance traveled by the emitted positron before annihilating with an electron). Besides the positron range, additional gamma photon emission is another physical property characteristic of PET radioisotopes that can affect the reconstructed image quality, as it happens to the isotope 48V. In this thesis we assess these effects through studies of spatial resolution and image quality. Finally, we analyze the scope of the NEMA NU 4-2008 to carry out such studies, adapting it and proposing possible modifications.
Resumo:
SIMLIDAR is an application developed in Cþþ that generates an artificial orchard using a Lindenmayer system. The application simulates the lateral interaction between the artificial orchard and a laser scanner or LIDAR (Light Detection and Ranging). To best highlight the unique qualities of the LIDAR simulation, this work focuses on apple trees without leaves, i.e. the woody structure. The objective is to simulate a terrestrial laser sensor (LIDAR) when applied to different artificially created orchards and compare the simulated characteristics of trees with the parameters obtained with the LIDAR. The scanner is mounted on a virtual tractor and measures the distance between the origin of the laser beam and the nearby plant object. This measurement is taken with an angular scan in a plane which is perpendicular to the route of the virtual tractor. SIMLIDAR determines the distance measured in a bi-dimensional matrix N M, where N is the number of angular scans and M is the number of steps in the tractor route. In order to test the data and performance of SIMLIDAR, the simulation has been applied to 42 different artificial orchards. After previously defining and calculating two vegetative parameters (wood area and wood projected area) of the simulated trees, a good correlation (R2 ¼ 0.70e0.80) was found between these characteristics and the wood area detected (impacted) by the laser beam. The designed software can be valuable in horticulture for estimating biomass and optimising the pesticide treatments that are performed in winter.
Resumo:
This paper is concerned with the study of non-Markovian queuing systems in container terminals. The methodology presented has been applied to analyze the ship traffic in the port of Valencia located in the Western Mediterranean. Two container terminals have been studied: the public container terminal of NOATUM and the dedicated container terminal of MSC. This paper contains the results of a simulation model based on queuing theory. The methodology presented is found to be effective in replicating realistic ship traffic operations in port as well as in conducting capacity evaluations. Thus the methodology can be used for capacity planning (long term), tactical planning (medium term) and even for the container terminal design (port enlargement purposes).
Resumo:
Corrosion of reinforcing steel in concrete due to chloride ingress is one of the main causes of the deterioration of reinforced concrete structures. Structures most affected by such a corrosion are marine zone buildings and structures exposed to de-icing salts like highways and bridges. Such process is accompanied by an increase in volume of the corrosión products on the rebarsconcrete interface. Depending on the level of oxidation, iron can expand as much as six times its original volume. This increase in volume exerts tensile stresses in the surrounding concrete which result in cracking and spalling of the concrete cover if the concrete tensile strength is exceeded. The mechanism by which steel embedded in concrete corrodes in presence of chloride is the local breakdown of the passive layer formed in the highly alkaline condition of the concrete. It is assumed that corrosion initiates when a critical chloride content reaches the rebar surface. The mathematical formulation idealized the corrosion sequence as a two-stage process: an initiation stage, during which chloride ions penetrate to the reinforcing steel surface and depassivate it, and a propagation stage, in which active corrosion takes place until cracking of the concrete cover has occurred. The aim of this research is to develop computer tools to evaluate the duration of the service life of reinforced concrete structures, considering both the initiation and propagation periods. Such tools must offer a friendly interface to facilitate its use by the researchers even though their background is not in numerical simulation. For the evaluation of the initiation period different tools have been developed: Program TavProbabilidade: provides means to carry out a probability analysis of a chloride ingress model. Such a tool is necessary due to the lack of data and general uncertainties associated with the phenomenon of the chloride diffusion. It differs from the deterministic approach because it computes not just a chloride profile at a certain age, but a range of chloride profiles for each probability or occurrence. Program TavProbabilidade_Fiabilidade: carries out reliability analyses of the initiation period. It takes into account the critical value of the chloride concentration on the steel that causes breakdown of the passive layer and the beginning of the propagation stage. It differs from the deterministic analysis in that it does not predict if the corrosion is going to begin or not, but to quantifies the probability of corrosion initiation. Program TavDif_1D: was created to do a one dimension deterministic analysis of the chloride diffusion process by the finite element method (FEM) which numerically solves Fick’second Law. Despite of the different FEM solver already developed in one dimension, the decision to create a new code (TavDif_1D) was taken because of the need to have a solver with friendly interface for pre- and post-process according to the need of IETCC. An innovative tool was also developed with a systematic method devised to compare the ability of the different 1D models to predict the actual evolution of chloride ingress based on experimental measurements, and also to quantify the degree of agreement of the models with each others. For the evaluation of the entire service life of the structure: a computer program has been developed using finite elements method to do the coupling of both service life periods: initiation and propagation. The program for 2D (TavDif_2D) allows the complementary use of two external programs in a unique friendly interface: • GMSH - an finite element mesh generator and post-processing viewer • OOFEM – a finite element solver. This program (TavDif_2D) is responsible to decide in each time step when and where to start applying the boundary conditions of fracture mechanics module in function of the amount of chloride concentration and corrosion parameters (Icorr, etc). This program is also responsible to verify the presence and the degree of fracture in each element to send the Information of diffusion coefficient variation with the crack width. • GMSH - an finite element mesh generator and post-processing viewer • OOFEM – a finite element solver. The advantages of the FEM with the interface provided by the tool are: • the flexibility to input the data such as material property and boundary conditions as time dependent function. • the flexibility to predict the chloride concentration profile for different geometries. • the possibility to couple chloride diffusion (initiation stage) with chemical and mechanical behavior (propagation stage). The OOFEM code had to be modified to accept temperature, humidity and the time dependent values for the material properties, which is necessary to adequately describe the environmental variations. A 3-D simulation has been performed to simulate the behavior of the beam on both, action of the external load and the internal load caused by the corrosion products, using elements of imbedded fracture in order to plot the curve of the deflection of the central region of the beam versus the external load to compare with the experimental data.
Resumo:
The new Spanish Regulation in Building Acoustic establishes values and limits for the different acoustic magnitudes whose fulfillment can be verify by means field measurements. In this sense, an essential aspect of a field measurement is to give the measured magnitude and the uncertainty associated to such a magnitude. In the calculus of the uncertainty it is very usual to follow the uncertainty propagation method as described in the Guide to the expression of Uncertainty in Measurements (GUM). Other option is the numerical calculus based on the distribution propagation method by means of Monte Carlo simulation. In fact, at this stage, it is possible to find several publications developing this last method by using different software programs. In the present work, we used Excel for the Monte Carlo simulation for the calculus of the uncertainty associated to the different magnitudes derived from the field measurements following ISO 140-4, 140-5 and 140-7. We compare the results with the ones obtained by the uncertainty propagation method. Although both methods give similar values, some small differences have been observed. Some arguments to explain such differences are the asymmetry of the probability distributions associated to the entry magnitudes,the overestimation of the uncertainty following the GUM
Resumo:
Distributed target tracking in wireless sensor networks (WSN) is an important problem, in which agreement on the target state can be achieved using conventional consensus methods, which take long to converge. We propose distributed particle filtering based on belief propagation (DPF-BP) consensus, a fast method for target tracking. According to our simulations, DPF-BP provides better performance than DPF based on standard belief consensus (DPF-SBC) in terms of disagreement in the network. However, in terms of root-mean square error, it can outperform DPF-SBC only for a specific number of consensus iterations.
Resumo:
Se describen las variaciones de temperaturas y de tensiones durante la construcción de presas de hormigón compactado. The curing of concrete is an exothermic process. The heat of hydration generated induces temperature increases in the concrete, which will disappear in the long term by heat conduction in the concrete mass and thermal exchanges with the environment. The problem is of particularly interest for large concrete masses, as is the case of dams, because the time involved in the heat diffusion process grows with the square of the dimensions and a hotter dam interior implies the possibility of cracking the exposed surfaces of the dam. The Cuira dam, currently being built in Venezuela using roller compacted concrete, is a 134 m high, arch-gravity dam. In support of the design, different strategies were analysed, including various combinations of cooling of the water and the aggregate in order to achieve acceptable results. The calculations were conducted with Abaqus, taking into account all the necessary mechanical and thermal characteristics, as well as the relevant non-linearities. The analyses led to the conclusion that no cooling was required, even taking into account the stress state imposed by an early and rapid filling of the reservoir.
Resumo:
Use of computational fluid dynamic (CFD) methods to predict the power production from wind entire wind farms in flat and complex terrain is presented in this paper. Two full 3D Navier–Stokes solvers for incompressible flow are employed that incorporate the k–ε and k–ω turbulence models respectively. The wind turbines (W/Ts) are modelled as momentum absorbers by means of their thrust coefficient using the actuator disk approach. The WT thrust is estimated using the wind speed one diameter upstream of the rotor at hub height. An alternative method that employs an induction-factor based concept is also tested. This method features the advantage of not utilizing the wind speed at a specific distance from the rotor disk, which is a doubtful approximation when a W/T is located in the wake of another and/or the terrain is complex. To account for the underestimation of the near wake deficit, a correction is introduced to the turbulence model. The turbulence time scale is bounded using the general “realizability” constraint for the turbulent velocities. Application is made on two wind farms, a five-machine one located in flat terrain and another 43-machine one located in complex terrain. In the flat terrain case, the combination of the induction factor method along with the turbulence correction provides satisfactory results. In the complex terrain case, there are some significant discrepancies with the measurements, which are discussed. In this case, the induction factor method does not provide satisfactory results.
Resumo:
En la última década ha aumentado en gran medida el interés por las redes móviles Ad Hoc. La naturaleza dinámica y sin infraestructura de estas redes, exige un nuevo conjunto de algoritmos y estrategias para proporcionar un servicio de comunicación fiable extremo a extremo. En el contexto de las redes móviles Ad Hoc, el encaminamiento surge como una de las áreas más interesantes para transmitir información desde una fuente hasta un destino, con la calidad de servicio de extremo a extremo. Debido a las restricciones inherentes a las redes móviles, los modelos de encaminamiento tradicionales sobre los que se fundamentan las redes fijas, no son aplicables a las redes móviles Ad Hoc. Como resultado, el encaminamiento en redes móviles Ad Hoc ha gozado de una gran atención durante los últimos años. Esto ha llevado al acrecentamiento de numerosos protocolos de encaminamiento, tratando de cubrir con cada uno de ellos las necesidades de los diferentes tipos de escenarios. En consecuencia, se hace imprescindible estudiar el comportamiento de estos protocolos bajo configuraciones de red variadas, con el fin de ofrecer un mejor encaminamiento respecto a los existentes. El presente trabajo de investigación muestra precisamente una solución de encaminamiento en las redes móviles Ad Hoc. Dicha solución se basa en el mejoramiento de un algoritmo de agrupamiento y la creación de un modelo de encaminamiento; es decir, un modelo que involucra la optimización de un protocolo de enrutamiento apoyado de un mecanismo de agrupación. El algoritmo mejorado, denominado GMWCA (Group Management Weighted Clustering Algorithm) y basado en el WCA (Weighted Clustering Algorithm), permite calcular el mejor número y tamaño de grupos en la red. Con esta mejora se evitan constantes reagrupaciones y que los jefes de clústeres tengan más tiempo de vida intra-clúster y por ende una estabilidad en la comunicación inter-clúster. En la tesis se detallan las ventajas de nuestro algoritmo en relación a otras propuestas bajo WCA. El protocolo de enrutamiento Ad Hoc propuesto, denominado QoS Group Cluster Based Routing Protocol (QoSG-CBRP), utiliza como estrategia el empleo de clúster y jerarquías apoyada en el algoritmo de agrupamiento. Cada clúster tiene un jefe de clúster (JC), quien administra la información de enrutamiento y la envía al destino cuando esta fuera de su área de cobertura. Para evitar que haya constantes reagrupamientos y llamados al algoritmo de agrupamiento se consideró agregarle un jefe de cluster de soporte (JCS), el que asume las funciones del JC, siempre y cuando este haya roto el enlace con los otros nodos comunes del clúster por razones de alejamiento o por desgaste de batería. Matemáticamente y a nivel de algoritmo se han demostrado las mejoras del modelo propuesto, el cual ha involucrado el mejoramiento a nivel de algoritmo de clustering y del protocolo de enrutamiento. El protocolo QoSG-CBRP, se ha implementado en la herramienta de simulación Network Simulator 2 (NS2), con la finalidad de ser comparado con el protocolo de enrutamiento jerárquico Cluster Based Routing Protocol (CBRP) y con un protocolo de enrutamiento Ad Hoc reactivo denominado Ad Hoc On Demand Distance Vector Routing (AODV). Estos protocolos fueron elegidos por ser los que mejor comportamiento presentaron dentro de sus categorías. Además de ofrecer un panorama general de los actuales protocolos de encaminamiento en redes Ad Hoc, este proyecto presenta un procedimiento integral para el análisis de capacidades de la propuesta del nuevo protocolo con respecto a otros, sobre redes que tienen un alto número de nodos. Estas prestaciones se miden en base al concepto de eficiencia de encaminamiento bajo parámetros de calidad de servicio (QoS), permitiendo establecer el camino más corto posible entre un nodo origen y un nodo destino. Con ese fin se han realizado simulaciones con diversos escenarios para responder a los objetivos de la tesis. La conclusiones derivadas del análisis de los resultados permiten evaluar cualitativamente las capacidades que presenta el protocolo dentro del modelo propuesto, al mismo tiempo que avizora un atractivo panorama en líneas futuras de investigación. ABSTRACT In the past decade, the interest in mobile Ad Hoc networks has greatly increased. The dynamic nature of these networks without infrastructure requires a new set of algorithms and strategies to provide a reliable end-to-end communication service. In the context of mobile Ad Hoc networks, routing emerges as one of the most interesting areas for transmitting information from a source to a destination, with the quality of service from end-to-end. Due to the constraints of mobile networks, traditional routing models that are based on fixed networks are not applicable to Ad Hoc mobile networks. As a result, the routing in mobile Ad Hoc networks has experienced great attention in recent years. This has led to the enhancement of many routing protocols, trying to cover with each one of them, the needs of different types of scenarios. Consequently, it is essential to study the behavior of these protocols under various network configurations, in order to provide a better routing scheme. Precisely, the present research shows a routing solution in mobile Ad Hoc networks. This solution is based on the improvement of a clustering algorithm, and the creation of a routing model, ie a model that involves optimizing a routing protocol with the support of a grouping mechanism. The improved algorithm called GMWCA (Group Management Weighted Clustering Algorithm) and based on the WCA (Weighted Clustering Algorithm), allows to calculate the best number and size of groups in the network. With this enhancement, constant regroupings are prevented and cluster heads are living longer intra-cluster lives and therefore stability in inter-cluster communication. The thesis details the advantages of our algorithm in relation to other proposals under WCA. The Ad Hoc routing protocol proposed, called QoS Group Cluster Based Routing Protocol (QoSG-CBRP), uses a cluster-employment strategy and hierarchies supported by the clustering algorithm. Each cluster has a cluster head (JC), who manages the routing information and sends it to the destination when is out of your coverage area. To avoid constant rearrangements and clustering algorithm calls, adding a support cluster head (JCS) was considered. The JCS assumes the role of the JC as long as JC has broken the link with the other nodes in the cluster for common restraining reasons or battery wear. Mathematically and at an algorithm level, the improvements of the proposed model have been showed, this has involved the improvement level clustering algorithm and the routing protocol. QoSG-CBRP protocol has been implemented in the simulation tool Network Simulator 2 (NS2), in order to be compared with the hierarchical routing protocol Cluster Based Routing Protocol (CBRP) and with the reactive routing protocol Ad Hoc On Demand Distance Vector Routing (AODV). These protocols were chosen because they showed the best individual performance in their categories. In addition to providing an overview of existing routing protocols in Ad Hoc networks, this project presents a comprehensive procedure for capacity analysis of the proposed new protocol with respect to others on networks that have a high number of nodes. These benefits are measured based on the concept of routing efficiency under the quality of service (QoS) parameters, thus allowing for the shortest possible path between a source node and a destination node. To meet the objectives of the thesis, simulations have been performed with different scenarios. The conclusions derived from the analysis of the results to assess qualitatively the protocol capabilities presented in the proposed model, while an attractive scenario for future research appears.
Resumo:
Tropospheric phenomena such as clouds and mainly rain cause higher attenuation at Ka-band than at lower frequencies. In this collaborative paper, the main results of four long-term Ka-band propagation campaigns are presented. The experiments are carried out in Ottawa, Canada (satellite Anik F2); Aveiro, Portugal; Madrid, Spain; and Toulouse, France (satellite HotBird 6 in the last three cases) and have been running since 2004 in Aveiro, 2006 in Ottawa and Madrid, and 2008 in Toulouse. After a brief introduction of the experiments, rain rate and excess attenuation results are discussed, first for a common two-year measurement period and later for the whole database available. Seasonal attenuation statistics for Madrid, Ottawa and Aveiro are compared. Finally, fade duration and fade slope statistics derived at three locations are presented and discussed.
Resumo:
Ultrasonic sound velocity measurements with hand-held equipment remain due to their simplicity among the most used methods for non-destructive grading of sawn woods, yet a dedicated normalization effort with respect to strength classes for Spanish species is still required. As part of an ongoing project with the aim of definition of standard testing methods, the effect of the dimensions of commonly tested Scots pine (Pinus sylvestris L.) timbers and equipment testing frequency on ultrasonic velocity were investigated. A dedicated full-wave finite-difference time-domain software allowed simulation of pulse propagation through timbers of representative length and section combinations. Sound velocity measurements vL were performed along the grain with the indirect method at 22 kHz and 45 kHz for grids of measurement points at specific distances. For sample sections larger than the cross-sectional wavelength ?RT, the simulated sound velocity vL converges to vL = (CL/?)0.5. For smaller square sections the sound velocity drops down to vL = (EL/?)0.5, where CL, EL and ? are the stiffness, E-modul and density, respectively. The experiments confirm a linear regression between time of flight and measurement distance even at less than two wavelength menor que2?L distance, the fitted sound speed values increased by 15% between the two tested frequencies.
Resumo:
Desde la aparición del turborreactor, el motor aeróbico con turbomaquinaria ha demostrado unas prestaciones excepcionales en los regímenes subsónico y supersónico bajo. No obstante, la operación a velocidades superiores requiere sistemas más complejos y pesados, lo cual ha imposibilitado la ejecución de estos conceptos. Los recientes avances tecnológicos, especialmente en materiales ligeros, han restablecido el interés por los motores de ciclo combinado. La simulación numérica de estos nuevos conceptos es esencial para estimar las prestaciones de la planta propulsiva, así como para abordar las dificultades de integración entre célula y motor durante las primeras etapas de diseño. Al mismo tiempo, la evaluación de estos extraordinarios motores requiere una metodología de análisis distinta. La tesis doctoral versa sobre el diseño y el análisis de los mencionados conceptos propulsivos mediante el modelado numérico y la simulación dinámica con herramientas de vanguardia. Las distintas arquitecturas presentadas por los ciclos combinados basados en sendos turborreactor y motor cohete, así como los diversos sistemas comprendidos en cada uno de ellos, hacen necesario establecer una referencia común para su evaluación. Es más, la tendencia actual hacia aeronaves "más eléctricas" requiere una nueva métrica para juzgar la aptitud de un proceso de generación de empuje en el que coexisten diversas formas de energía. A este respecto, la combinación del Primer y Segundo Principios define, en un marco de referencia absoluto, la calidad de la trasferencia de energía entre los diferentes sistemas. Esta idea, que se ha estado empleando desde hace mucho tiempo en el análisis de plantas de potencia terrestres, ha sido extendida para relacionar la misión de la aeronave con la ineficiencia de cada proceso involucrado en la generación de empuje. La metodología se ilustra mediante el estudio del motor de ciclo combinado variable de una aeronave para el crucero a Mach 5. El diseño de un acelerador de ciclo combinado basado en el turborreactor sirve para subrayar la importancia de la integración del motor y la célula. El diseño está limitado por la trayectoria ascensional y el espacio disponible en la aeronave de crucero supersónico. Posteriormente se calculan las prestaciones instaladas de la planta propulsiva en función de la velocidad y la altitud de vuelo y los parámetros de control del motor: relación de compresión, relación aire/combustible y área de garganta. ABSTRACT Since the advent of the turbojet, the air-breathing engine with rotating machinery has demonstrated exceptional performance in the subsonic and low supersonic regimes. However, the operation at higher speeds requires further system complexity and weight, which so far has impeded the realization of these concepts. Recent technology developments, especially in lightweight materials, have restored the interest towards combined-cycle engines. The numerical simulation of these new concepts is essential at the early design stages to compute a first estimate of the engine performance in addition to addressing airframe-engine integration issues. In parallel, a different analysis methodology is required to evaluate these unconventional engines. The doctoral thesis concerns the design and analysis of the aforementioned engine concepts by means of numerical modeling and dynamic simulation with state-of-the-art tools. A common reference is needed to evaluate the different architectures of the turbine and the rocket-based combined-cycle engines as well as the various systems within each one of them. Furthermore, the actual trend towards more electric aircraft necessitates a common metric to judge the suitability of a thrust generation process where different forms of energy coexist. In line with this, the combination of the First and the Second Laws yields the quality of the energy being transferred between the systems on an absolute reference frame. This idea, which has been since long applied to the analysis of on-ground power plants, was extended here to relate the aircraft mission with the inefficiency of every process related to the thrust generation. The methodology is illustrated with the study of a variable- combined-cycle engine for a Mach 5 cruise aircraft. The design of a turbine-based combined-cycle booster serves to highlight the importance of the engine-airframe integration. The design is constrained by the ascent trajectory and the allocated space in the supersonic cruise aircraft. The installed performance of the propulsive plant is then computed as a function of the flight speed and altitude and the engine control parameters: pressure ratio, air-to-fuel ratio and throat area.
Resumo:
Sight distance is of major importance for road safety either when designing new roads or analysing the alignment of existing roads. It is essential that available sight distance in roads is long enough for emergency stops or overtaking manoeuvres. Also, it is vital for engineers/researchers that the tools used for that analysis are both powerful and intuitive. Based on ArcGIS, the application to be presented not only performs an exhaustive sight distance calculation, but allows an accurate analysis of 3D alignment, using all new tools, from a Digital Elevation Model and vehicle trajectory. The software has been successfully utilised to analyse several two-lane rural roads in Spain. In addition, the software produces thematic maps representing sight distance in which supplementary information about crashes, traffic flow, speed or design consistency could be included, allowing traffic safety studies.