14 resultados para Synthetic cartography

em Universidad Politécnica de Madrid


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nowadays, many researches focus their efforts in studies and applications on the Learning area. However, there is a lack of a reference system that permits to know the positioning and the existing links between Learning and Information Technologies. This paper proposes a Cartography where explains the relationships between the elements that compose the Learning Theories and Information Technologies, considering the own features of the learner and the Information Technologies Properties. This intersection will allow us to know what Information Technologies Properties promote Learning Futures.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hojas Kilométricas (Kilometric Sheets). Specifically, the study focuses on those sheets referring to the city centre and surrounding area of the Royal Site of Aranjuez, a town in the south of the Province of Madrid. The aim of this study is to restore the actual size and measurements of scanned images of the Hojas Kilométricas. This would allow us, among other things, to reestablish both the format and scale of the original plans. To achieve this goal it is necessary to rectify and then georeference these images, i.e. assign them a geographic reference system. This procedure is essential in the overlaying and comparison of the Hojas Kilométricas of the Royal Site with other historical cartography as well as other sources related to the same area from different time periods. Subsequent research would allow us, for example, to reconstruct the time-evolution of the urban area, to spot new construction and to pinpoint the locations of any altered or missing buildings or architectural features. In addition, this would allow us to develop and integrate databases for GIS models applicable to the management of our cultural heritage.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Zinc chelates have been widely used to correct deficiencies in this micronutrient in different soil types and under different moisture conditions. The aging of the metal in soil could cause a change in its availability. Over time the most labile forms of Zn could decrease in activity and extractability and change to more stable forms. Various soil parameters, such as redox conditions, time, soil type and moisture conditions, affect the aging process and modify the solubility of the metal. In general, redox conditions influence pH and also the chemical forms dissolved in the soil solution. Soil pH also affects Zn solubility; at high pH values, most of the Zn is present in forms that are not bioavailable to plants. The objective of this study was to determine the changes in Zn over time in a soil solution in a waterlogged acidic soil to which synthetic and natural chelates were applied

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Adding Zn improves crop growth, increases seed yield and also positively affects nutritional quality. After Zn fertilization, there is normally a period of several years in which residual effects provide an adequate supply of Zn to successive crops. Immediately after the application of Zn sources water-soluble Zn slowly but continually decreases. Various factors, including time and moisture conditions, affect the aging process and modify the solubility of the metal in soil and therefore its availability. In previous experiments, we studied the residual effect of synthetic chelates, obtained that the amounts of potentially available Zn decreased in the second cropping year due to aging processes. The present study was undertaken to verify variations in the residual effects of applying four different synthetic Zn sources

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This article presents a cartographic system to facilitate cooperative manoeuvres among autonomous vehicles in a well-known environment. The main objective is to design an extended cartographic system to help in the navigation of autonomous vehicles. This system has to allow the vehicles not only to access the reference points needed for navigation, but also noticeable information such as the location and type of traffic signals, the proximity to a crossing, the streets en route, etc. To do this, a hierarchical representation of the information has been chosen, where the information has been stored in two levels. The lower level contains the archives with the Universal Traverse Mercator (UTM) coordinates of the points that define the reference segments to follow. The upper level contains a directed graph with the relational database in which streets, crossings, roundabouts and other points of interest are represented. Using this new system it is possible to know when the vehicle approaches a crossing, what other paths arrive at that crossing, and, should there be other vehicles circulating on those paths and arriving at the crossing, which one has the highest priority. The data obtained from the cartographic system is used by the autonomous vehicles for cooperative manoeuvres.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Synthetic Aperture Radar (SAR) images a target region reflectivity function in the multi-dimensional spatial domain of range and cross-range. SAR synthesizes a large aperture radar in order to achieve a finer azimuth resolution than the one provided by any on-board real antenna. Conventional SAR techniques assume a single reflection of transmitted waveforms from targets. Nevertheless, today¿s new scenes force SAR systems to work in urban environments. Consequently, multiple-bounce returns are added to directscatter echoes. We refer to these as ghost images, since they obscure true target image and lead to poor resolution. By analyzing the quadratic phase error (QPE), this paper demonstrates that Earth¿s curvature influences the defocusing degree of multipath returns. In addition to the QPE, other parameters such as integrated sidelobe ratio (ISLR), peak sidelobe ratio (PSLR), contrast (C) and entropy (E) provide us with the tools to identify direct-scatter echoes in images containing undesired returns coming from multipath.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The accuracy of Tomás López´s historical cartography of the Canary Islands included in the “Atlas Particular” of the Kingdoms of Spain, Portugal and Adjacent Islands” is analyzed. For this purpose, we propose a methodology based on Geographic Information Systems (GIS), a comparison of digitized historical cartography population centres with current ones. This study shows that the lineal error value is small for the smaller islands: Lanzarote, El Hierro, La Palma and La Gomera. In the large islands of Tenerife, Fuerteventura and Gran Canaria, the error is smaller in central zones but increases towards the coast. This indicates that Tomás López began his cartography starting from central island zones, accumulating errors due to lack of geodetic references as he moved toward the coast.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

cartografía de incertidumbres

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A simple and scalable chemical approach has been proposed for the generation of 1-dimensional nanostructures of two most important inorganic materials such as zinc oxide and cadmium sulfide. By controlling the growth habit of the nanostructures with manipulated reaction conditions, the diameter and uniformity of the nanowires/nanorods were tailored. We studied extensively optical behavior and structural growth of CdS NWs and ZnO NRs doped ferroelectric liquid crystal Felix-017/100. Due to doping band gap has been changed and several blue shifts occurred in photoluminescence spectra because of nanoconfinement effect and mobility of charges.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Foliage Penetration (FOPEN) radar systems were introduced in 1960, and have been constantly improved by several organizations since that time. The use of Synthetic Aperture Radar (SAR) approaches for this application has important advantages, due to the need for high resolution in two dimensions. The design of this type of systems, however, includes some complications that are not present in standard SAR systems. FOPEN SAR systems need to operate with a low central frequency (VHF or UHF bands) in order to be able to penetrate the foliage. High bandwidth is also required to obtain high resolution. Due to the low central frequency, large integration angles are required during SAR image formation, and therefore the Range Migration Algorithm (RMA) is used. This project thesis identifies the three main complications that arise due to these requirements. First, a high fractional bandwidth makes narrowband propagation models no longer valid. Second, the VHF and UHF bands are used by many communications systems. The transmitted signal spectrum needs to be notched to avoid interfering them. Third, those communications systems cause Radio Frequency Interference (RFI) on the received signal. The thesis carries out a thorough analysis of the three problems, their degrading effects and possible solutions to compensate them. The UWB model is applied to the SAR signal, and the degradation induced by it is derived. The result is tested through simulation of both a single pulse stretch processor and the complete RMA image formation. Both methods show that the degradation is negligible, and therefore the UWB propagation effect does not need compensation. A technique is derived to design a notched transmitted signal. Then, its effect on the SAR image formation is evaluated analytically. It is shown that the stretch processor introduces a processing gain that reduces the degrading effects of the notches. The remaining degrading effect after processing gain is assessed through simulation, and an experimental graph of degradation as a function of percentage of nulled frequencies is obtained. The RFI is characterized and its effect on the SAR processor is derived. Once again, a processing gain is found to be introduced by the receiver. As the RFI power can be much higher than that of the desired signal, an algorithm is proposed to remove the RFI from the received signal before RMA processing. This algorithm is a modification of the Chirp Least Squares Algorithm (CLSA) explained in [4], which adapts it to deramped signals. The algorithm is derived analytically and then its performance is evaluated through simulation, showing that it is effective in removing the RFI and reducing the degradation caused by both RFI and notching. Finally, conclusions are drawn as to the importance of each one of the problems in SAR system design.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

En los últimos años, y a la luz de los retos a los que se enfrenta la sociedad, algunas voces están urgiendo a dejar atrás los paradigmas modernos —eficiencia y rendimiento— que sustentan a las llamadas prácticas sostenibles, y están alentando a repensar, en el contexto de los cambios científicos y culturales, una agenda termodinámica y ecológica para la arquitectura. La cartografía que presenta esta tesis doctoral se debe de entender en este contexto. Alineándose con esta necesidad, se esfuerza por dar a este empeño la profundidad histórica de la que carece. De este modo, el esfuerzo por dotar a la arquitectura de una agenda de base científica, se refuerza con una discusión cultural sobre el progresivo empoderamiento de las ideas termodinámicas en la arquitectura. Esta cartografía explora la historia de las ideas termodinámicas en la arquitectura desde el principio del siglo XX hasta la actualidad. Estudia, con el paso de los sistemas en equilibrio a los alejados del equilibrio como trasfondo, como las ideas termodinámicas han ido infiltrándose gradualmente en la arquitectura. Este esfuerzo se ha planteado desde un doble objetivo. Primero, adquirir una distancia crítica respecto de las prácticas modernas, de modo que se refuerce y recalibre el armazón intelectual y las herramientas sobre las que se está apoyando esta proyecto termodinámico. Y segundo, desarrollar una aproximación proyectual sobre la que se pueda fundamentar una agenda termodinámica para la arquitectura, asunto que se aborda desde la firme creencia de que es posible una re-descripción crítica de la realidad. De acuerdo con intercambios de energía que se dan alrededor y a través de un edificio, esta cartografía se ha estructurado en tres entornos termodinámicos, que sintetizan mediante un corte transversal la variedad de intercambios de energía que se dan en la arquitectura: -Cualquier edificio, como constructo espacial y material inmerso en el medio, intercambia energía mediante un flujo bidireccional con su contexto, definiendo un primer entorno termodinámico al que se denomina atmósferas territoriales. -En el interior de los edificios, los flujos termodinámicos entre la arquitectura y su ambiente interior definen un segundo entorno termodinámico, atmósferas materiales, que explora las interacciones entre los sistemas materiales y la atmósfera interior. -El tercer entorno termodinámico, atmosferas fisiológicas, explora los intercambios de energía que se dan entre el cuerpo humano y el ambiente invisible que lo envuelve, desplazando el objeto de la arquitectura desde el marco físico hacia la interacción entre la atmósfera y los procesos somáticos y percepciones neurobiológicas de los usuarios. A través de estos tres entornos termodinámicos, esta cartografía mapea aquellos patrones climáticos que son relevantes para la arquitectura, definiendo tres situaciones espaciales y temporales sobre las que los arquitectos deben actuar. Estudiando las conexiones entre la atmósfera, la energía y la arquitectura, este mapa presenta un conjunto de ideas termodinámicas disponibles —desde los parámetros de confort definidos por la industria del aire acondicionado hasta las técnicas de acondicionamiento pasivo— que, para ser efectivas, necesitan ser evaluadas, sintetizadas y recombinadas a la luz de los retos de nuestro tiempo. El resultado es un manual que, mediando entre la arquitectura y la ciencia, y a través de este relato histórico, acorta la distancia entre la arquitectura y la termodinámica, preparando el terreno para la definición de una agenda termodinámica para el proyecto de arquitectura. A este respecto, este mapa se entiende como uno de los pasos necesarios para que la arquitectura recupere la capacidad de intervenir en la acuciante realidad a la que se enfrenta. ABSTRACT During the last five years, in the light of current challenges, several voices are urging to leave behind the modern energy paradigms —efficiency and performance— on which the so called sustainable practices are relying, and are posing the need to rethink, in the light of the scientific and cultural shifts, the thermodynamic and ecological models for architecture. The historical cartography this PhD dissertation presents aligns with this effort, providing the cultural background that this endeavor requires. The drive to ground architecture on a scientific basis needs to be complemented with a cultural discussion of the history of thermodynamic ideas in architecture. This cartography explores the history of thermodynamic ideas in architecture, from the turn of the 20th century until present day, focusing on the energy interactions between architecture and atmosphere. It surveys the evolution of thermodynamic ideas —the passage from equilibrium to far from equilibrium thermodynamics— and how these have gradually empowered within design and building practices. In doing so, it has posed a double-objective: first, to acquire a critical distance with modern practices which strengthens and recalibrates the intellectual framework and the tools in which contemporary architectural endeavors are unfolding; and second, to develop a projective approach for the development a thermodynamic agenda for architecture and atmosphere, with the firm belief that a critical re-imagination of reality is possible. According to the different systems which exchange energy across a building, the cartography has been structured in three particular thermodynamic environments, providing a synthetic cross-section of the range of thermodynamic exchanges which take place in architecture: -Buildings, as spatial and material constructs immersed in the environment, are subject to a contiuous bidirectional flow of energy with its context, defining a the first thermodynamic environment called territorial atmospheres. -Inside buildings, the thermodynamic flow between architecture and its indoor ambient defines a second thermodynamic environment, material atmospheres, which explores the energy interactions between the indoor atmosphere and its material systems. -The third thermodynamic environment, physiological atmospheres, explores the energy exchanges between the human body and the invisible environment which envelopes it, shifting design drivers from building to the interaction between the atmosphere and the somatic processes and neurobiological perceptions of users. Through these three thermodynamic environments, this cartography maps those climatic patterns which pertain to architecture, providing three situations on which designers need to take stock. Studying the connections between atmosphere, energy and architecture this map presents, not a historical paradigm shift from mechanical climate control to bioclimatic passive techniques, but a range of available thermodynamic ideas which need to be assessed, synthesized and recombined in the light of the emerging challenges of our time. The result is a manual which, mediating between architecture and science, and through this particular historical account, bridges the gap between architecture and thermodynamics, paving the way to a renewed approach to atmosphere, energy and architecture. In this regard this cartography is understood as one of the necessary steps to recuperate architecture’s lost capacity to intervene in the pressing reality of contemporary societies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A numerical simulation of the aerodynamic behavior of high-speed trains under synthetic crosswinds at a 90º yaw angle is presented. The train geometry is the aerodynamic train model (ATM). Flow description based on numerical simulations is obtained using large eddy simulation (LES) and the commercial code ANSYSFluent V14.5. A crosswind whose averaged velocity and turbulence characteristics change with distance to the ground is imposed. Turbulent fluctuations that vary temporally and spatially are simulated with TurbSim code. The crosswind boundary condition is calculated for the distance the train runs during a simulation period. The inlet streamwise velocity boundary condition is generated using Taylor?s frozen turbulence hypothesis. The model gives a time history of the force and moments acting on the train; this includes averaged values, standard deviations and extreme values. Of particular interest are the spectra of the forces and moments, and the admittance spectra. For comparison, results obtained with LES and a uniform wind velocity fluctuating in time, and results obtained with Reynolds averaged Navier Stokes equations (RANS), and the averaged wind conditions, are also presented.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Numerical simulations of flow surrounding a synthetic jet actuating device are presented. By modifying a dynamic mesh technique available in OpenFoam-a well-documented open-source solver for fluid dynamics, detailed computations of the sinusoidal motion of the synthetic jet diaphragm were possible. Numerical solutions were obtained by solving the two dimensional incompressible viscous N-S equations, with the use of a second order implicit time marching scheme and a central finite volume method for spatial discretization in both streamwise and crossflow directions. A systematic parametric study is reported here, in which the external Reynolds number, the diaphragm amplitude and frequency, and the slot dimensions are varied.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La tomografía axial computerizada (TAC) es la modalidad de imagen médica preferente para el estudio de enfermedades pulmonares y el análisis de su vasculatura. La segmentación general de vasos en pulmón ha sido abordada en profundidad a lo largo de los últimos años por la comunidad científica que trabaja en el campo de procesamiento de imagen; sin embargo, la diferenciación entre irrigaciones arterial y venosa es aún un problema abierto. De hecho, la separación automática de arterias y venas está considerado como uno de los grandes retos futuros del procesamiento de imágenes biomédicas. La segmentación arteria-vena (AV) permitiría el estudio de ambas irrigaciones por separado, lo cual tendría importantes consecuencias en diferentes escenarios médicos y múltiples enfermedades pulmonares o estados patológicos. Características como la densidad, geometría, topología y tamaño de los vasos sanguíneos podrían ser analizados en enfermedades que conllevan remodelación de la vasculatura pulmonar, haciendo incluso posible el descubrimiento de nuevos biomarcadores específicos que aún hoy en dípermanecen ocultos. Esta diferenciación entre arterias y venas también podría ayudar a la mejora y el desarrollo de métodos de procesamiento de las distintas estructuras pulmonares. Sin embargo, el estudio del efecto de las enfermedades en los árboles arterial y venoso ha sido inviable hasta ahora a pesar de su indudable utilidad. La extrema complejidad de los árboles vasculares del pulmón hace inabordable una separación manual de ambas estructuras en un tiempo realista, fomentando aún más la necesidad de diseñar herramientas automáticas o semiautomáticas para tal objetivo. Pero la ausencia de casos correctamente segmentados y etiquetados conlleva múltiples limitaciones en el desarrollo de sistemas de separación AV, en los cuales son necesarias imágenes de referencia tanto para entrenar como para validar los algoritmos. Por ello, el diseño de imágenes sintéticas de TAC pulmonar podría superar estas dificultades ofreciendo la posibilidad de acceso a una base de datos de casos pseudoreales bajo un entorno restringido y controlado donde cada parte de la imagen (incluyendo arterias y venas) está unívocamente diferenciada. En esta Tesis Doctoral abordamos ambos problemas, los cuales están fuertemente interrelacionados. Primero se describe el diseño de una estrategia para generar, automáticamente, fantomas computacionales de TAC de pulmón en humanos. Partiendo de conocimientos a priori, tanto biológicos como de características de imagen de CT, acerca de la topología y relación entre las distintas estructuras pulmonares, el sistema desarrollado es capaz de generar vías aéreas, arterias y venas pulmonares sintéticas usando métodos de crecimiento iterativo, que posteriormente se unen para formar un pulmón simulado con características realistas. Estos casos sintéticos, junto a imágenes reales de TAC sin contraste, han sido usados en el desarrollo de un método completamente automático de segmentación/separación AV. La estrategia comprende una primera extracción genérica de vasos pulmonares usando partículas espacio-escala, y una posterior clasificación AV de tales partículas mediante el uso de Graph-Cuts (GC) basados en la similitud con arteria o vena (obtenida con algoritmos de aprendizaje automático) y la inclusión de información de conectividad entre partículas. La validación de los fantomas pulmonares se ha llevado a cabo mediante inspección visual y medidas cuantitativas relacionadas con las distribuciones de intensidad, dispersión de estructuras y relación entre arterias y vías aéreas, los cuales muestran una buena correspondencia entre los pulmones reales y los generados sintéticamente. La evaluación del algoritmo de segmentación AV está basada en distintas estrategias de comprobación de la exactitud en la clasificación de vasos, las cuales revelan una adecuada diferenciación entre arterias y venas tanto en los casos reales como en los sintéticos, abriendo así un amplio abanico de posibilidades en el estudio clínico de enfermedades cardiopulmonares y en el desarrollo de metodologías y nuevos algoritmos para el análisis de imágenes pulmonares. ABSTRACT Computed tomography (CT) is the reference image modality for the study of lung diseases and pulmonary vasculature. Lung vessel segmentation has been widely explored by the biomedical image processing community, however, differentiation of arterial from venous irrigations is still an open problem. Indeed, automatic separation of arterial and venous trees has been considered during last years as one of the main future challenges in the field. Artery-Vein (AV) segmentation would be useful in different medical scenarios and multiple pulmonary diseases or pathological states, allowing the study of arterial and venous irrigations separately. Features such as density, geometry, topology and size of vessels could be analyzed in diseases that imply vasculature remodeling, making even possible the discovery of new specific biomarkers that remain hidden nowadays. Differentiation between arteries and veins could also enhance or improve methods processing pulmonary structures. Nevertheless, AV segmentation has been unfeasible until now in clinical routine despite its objective usefulness. The huge complexity of pulmonary vascular trees makes a manual segmentation of both structures unfeasible in realistic time, encouraging the design of automatic or semiautomatic tools to perform the task. However, this lack of proper labeled cases seriously limits in the development of AV segmentation systems, where reference standards are necessary in both algorithm training and validation stages. For that reason, the design of synthetic CT images of the lung could overcome these difficulties by providing a database of pseudorealistic cases in a constrained and controlled scenario where each part of the image (including arteries and veins) is differentiated unequivocally. In this Ph.D. Thesis we address both interrelated problems. First, the design of a complete framework to automatically generate computational CT phantoms of the human lung is described. Starting from biological and imagebased knowledge about the topology and relationships between structures, the system is able to generate synthetic pulmonary arteries, veins, and airways using iterative growth methods that can be merged into a final simulated lung with realistic features. These synthetic cases, together with labeled real CT datasets, have been used as reference for the development of a fully automatic pulmonary AV segmentation/separation method. The approach comprises a vessel extraction stage using scale-space particles and their posterior artery-vein classification using Graph-Cuts (GC) based on arterial/venous similarity scores obtained with a Machine Learning (ML) pre-classification step and particle connectivity information. Validation of pulmonary phantoms from visual examination and quantitative measurements of intensity distributions, dispersion of structures and relationships between pulmonary air and blood flow systems, show good correspondence between real and synthetic lungs. The evaluation of the Artery-Vein (AV) segmentation algorithm, based on different strategies to assess the accuracy of vessel particles classification, reveal accurate differentiation between arteries and vein in both real and synthetic cases that open a huge range of possibilities in the clinical study of cardiopulmonary diseases and the development of methodological approaches for the analysis of pulmonary images.