981 resultados para Standard Penetration Test


Relevância:

30.00% 30.00%

Publicador:

Resumo:

We determined the stable oxygen and carbon isotopic composition of live (Rose Bengal stained) tests belonging to different size classes of two benthic foraminiferal species from the Pakistan continental margin. Samples were taken at 2 sites, with water depth of about 135 and 275 m, corresponding to the upper boundary and upper part of the core region of the oxygen minimum zone (OMZ). For Uvigerina ex gr. U. semiornata and Bolivina aff. B. dilatata, delta13C and delta18O values increased significantly with increasing test size. In the case of U. ex gr. U. semiornata, delta13C increased linearly by about 0.105 per mil for each 100-µm increment in test size, whereas delta18O increased by 0.02 to 0.06 per mil per 100 µm increment. For B. aff. B. dilatata the relationship between test size and stable isotopic composition is better described by logarithmic equations. A strong positive linear correlation is observed between delta18O and delta13C values of both taxa, with a constant ratio of delta18O and delta13C values close to 2:1. This suggests that the strong ontogenetic effect is mainly caused by kinetic isotope fractionation during CO2 uptake. Our data underline the necessity to base longer delta18O and delta13C isotope records derived from benthic foraminifera on size windows of 100 µm or less. This is already common practice in down-core isotopic studies of planktonic foraminifera.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this study, we demonstrate the utility of amino acid geochronology based on single-foraminiferal tests in Quaternary sediment cores from the Queensland margin, Australia. The large planktonic foraminifer Pulleniatina obliquiloculata is ubiquitous in shelf, slope, and basin sediments of north Queensland as well as pantropical oceans. Fossil tests are resistant to dissolution, and retain substantial concentrations of amino acids (2-4 nmol/mg of shell) over hundreds of thousands of years. Amino acid D and L isomers of aspartic acid (Asp) and glutamic acid (Glu) were separated using reverse phase chromatography, which is sensitive enough to analyze individual foraminifera tests. In all, 462 Pulleniatina tests from 80 horizons in 11 cores exhibit a systematic increase in D/L ratios down core. D/L ratios were determined in 32 samples whose ages are known from AMS 14C analyses. In all cases, the Asp and Glu D/L ratios are concordant with 14C age. D/L ratios of equal-age samples are slightly lower for cores taken from deeper water sites, reflecting the sensitivity of the rate of racemization to bottom water temperature. Beyond the range of 14C dating, previously identified marine oxygen-isotope stage boundaries provide approximate ages of the sediments up to about 500,000 years. For this longer time frame, D/L ratios also vary systematically with isotope-correlated ages. The rate of racemization for Glu and Asp was modeled using power functions. These equations can be used to estimate ages of samples from the Queensland margin extending back at least 500,000 years. This analytical approach provides new opportunities for geochronological control necessary to understand fundamental sedimentary processes affecting a wide range of marine environments.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ocean surface CO2 levels are increasing in line with rising atmospheric CO2 and could exceed 900 µatm by year 2100, with extremes above 2000 µatm in some coastal habitats. The imminent increase in ocean pCO2 is predicted to have negative consequences for marine fishes, including reduced aerobic performance, but variability among species could be expected. Understanding interspecific responses to ocean acidification is important for predicting the consequences of ocean acidification on communities and ecosystems. In the present study, the effects of exposure to near-future seawater CO2 (860 µatm) on resting (M O2rest) and maximum (M O2max) oxygen consumption rates were determined for three tropical coral reef fish species interlinked through predator-prey relationships: juvenile Pomacentrus moluccensis and Pomacentrus amboinensis, and one of their predators: adult Pseudochromis fuscus. Contrary to predictions, one of the prey species, P. amboinensis, displayed a 28-39% increase in M O2max after both an acute and four-day exposure to near-future CO2 seawater, while maintaining M O2rest. By contrast, the same treatment had no significant effects on M O2rest or M O2max of the other two species. However, acute exposure of P. amboinensis to 1400 and 2400 µatm CO2 resulted in M O2max returning to control values. Overall, the findings suggest that: (1) the metabolic costs of living in a near-future CO2 seawater environment were insignificant for the species examined at rest; (2) the M O2max response of tropical reef species to near-future CO2 seawater can be dependent on the severity of external hypercapnia; and (3) near-future ocean pCO2 may not be detrimental to aerobic scope of all fish species and it may even augment aerobic scope of some species. The present results also highlight that close phylogenetic relatedness and living in the same environment, does not necessarily imply similar physiological responses to near-future CO2.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Carbon capture and storage is increasingly being considered one of the most efficient approaches to mitigate the increase of CO2 in the atmosphere associated with anthropogenic emissions. However, the environmental effects of potential CO2 leaks remain largely unknown. The amphipod Ampelisca brevicornis was exposed to environmental sediments collected in different areas of the Gulf of Cádiz and subjected to several pH treatments to study the effects of CO2-induced acidification on sediment toxicity. After 10 days of exposure, the results obtained indicated that high lethal effects were associated with the lowest pH treatments, except for the Ría of Huelva sediment test. The mobility of metals from sediment to the overlying seawater was correlated to a pH decrease. The data obtained revealed that CO2-related acidification would lead to lethal effects on amphipods as well as the mobility of metals, which could increase sediment toxicity.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Continuous anthropogenic CO2 emissions to the atmosphere and uptake by the oceans will cause a reduction of seawater pH and saturation state (Omega) of CaCO3 minerals from which marine calcifiers build their shells and skeletons. Sea urchins use the most soluble form of calcium carbonate, high-magnesium calcite, to build their skeleton, spines and grazing apparatus. In order to highlight the effects of increased pCO2 on the test thickness and carbonate elemental composition of juvenile sea urchins and potential differences in their responses linked to the diet, we performed a laboratory experiment on juvenile Paracentrotus lividus, grazing on calcifying (Corallina elongata) and non-calcifying (Cystoseira amentacea, Dictyota dichotoma) macroalgae, under different pH (corresponding to pCO2 values of 390, 550, 750 and 1000 µatm). Results highlighted the importance of the diet in determining sea urchin size irrespectively of the pCO2 level, and the relevance of macroalgal diet in modulating urchin Mg/Ca ratio. The present study provides relevant clues both in terms of the mechanism of mineral incorporation and in terms of bottom-up processes (algal diet) affecting top-down ones (fish predation) in rocky subtidal communities

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The present study investigated the combined effects of ocean acidification, temperature, and salinity on growth and test degradation of Ammonia aomoriensis. This species is one of the dominant benthic foraminifera in near-coastal habitats of the southwestern Baltic Sea that can be particularly sensitive to changes in seawater carbonate chemistry. To assess potential responses to ocean acidification and climate change, we performed a fully crossed experiment involving three temperatures (8, 13, and 18°C), three salinities (15, 20, and 25) and four pCO2 levels (566, 1195, 2108, and 3843 µatm) for six weeks. Our results highlight a sensitive response of A. aomoriensis to undersaturated seawater with respect to calcite. The specimens continued to grow and increase their test diameter in treatments with pCO2 <1200 µatm, when Omega calc >1. Growth rates declined when pCO2 exceeded 1200 µatm (Omega calc <1). A significant reduction in test diameter and number of tests due to dissolution was observed below a critical Omega calc of 0.5. Elevated temperature (18°C) led to increased Omega calc, larger test diameter, and lower test degradation. Maximal growth was observed at 18°C. No significant relationship was observed between salinity and test growth. Lowered and undersaturated Omega calc, which results from increasing pCO2 in bottom waters, may cause a significant future decline of the population density of A. aomoriensis in its natural environment. At the same time, this effect might be partially compensated by temperature rise due to global warming.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Underground coal mines explosions generally arise from the inflammation of a methane/air mixture. This explosion can also generate a subsequent coal dust explosion. Traditionally such explosions have being fought eliminating one or several of the factors needed by the explosion to take place. Although several preventive measures are taken to prevent explosions, other measures should be considered to reduce the effects or even to extinguish the flame front. Unlike other protection methods that remove one or two of the explosion triangle elements, namely; the ignition source, the oxidizing agent and the fuel, explosion barriers removes all of them: reduces the quantity of coal in suspension, cools the flame front and the steam generated by vaporization removes the oxygen present in the flame. Passive water barriers are autonomous protection systems against explosions that reduce to a satisfactory safety level the effects of methane and/or flammable dust explosions. The barriers are activated by the pressure wave provoked in the explosion destroying the barrier troughs and producing a uniform dispersion of the extinguishing agent throughout the gallery section in quantity enough to extinguish the explosion flame. Full scale tests have been carried out in Polish Barbara experimental mine at GIG Central Mining Institute in order to determine the requirements and the optimal installation conditions of these devices for small sections galleries which are very frequent in the Spanish coal mines. Full scale tests results have been analyzed to understand the explosion timing and development, in order to assess on the use of water barriers in the typical small crosssection Spanish galleries. Several arrangements of water barriers have been designed and tested to verify the effectiveness of the explosion suppression in each case. The results obtained demonstrate the efficiency of the water barriers in stopping the flame front even with smaller amounts of water than those established by the European standard. According to the tests realized, water barriers activation times are between 0.52 s and 0.78 s and the flame propagation speed are between 75 m/s and 80 m/s. The maximum pressures (Pmax) obtained in the full scale tests have varied between 0.2 bar and 1.8 bar. Passive barriers protect effectively against the spread of the flame but cannot be used as a safeguard of the gallery between the ignition source and the first row of water troughs or bags, or even after them, as the pressure could remain high after them even if the flame front has been extinguished.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Los sensores inerciales (acelerómetros y giróscopos) se han ido introduciendo poco a poco en dispositivos que usamos en nuestra vida diaria gracias a su minituarización. Hoy en día todos los smartphones contienen como mínimo un acelerómetro y un magnetómetro, siendo complementados en losmás modernos por giróscopos y barómetros. Esto, unido a la proliferación de los smartphones ha hecho viable el diseño de sistemas basados en las medidas de sensores que el usuario lleva colocados en alguna parte del cuerpo (que en un futuro estarán contenidos en tejidos inteligentes) o los integrados en su móvil. El papel de estos sensores se ha convertido en fundamental para el desarrollo de aplicaciones contextuales y de inteligencia ambiental. Algunos ejemplos son el control de los ejercicios de rehabilitación o la oferta de información referente al sitio turístico que se está visitando. El trabajo de esta tesis contribuye a explorar las posibilidades que ofrecen los sensores inerciales para el apoyo a la detección de actividad y la mejora de la precisión de servicios de localización para peatones. En lo referente al reconocimiento de la actividad que desarrolla un usuario, se ha explorado el uso de los sensores integrados en los dispositivos móviles de última generación (luz y proximidad, acelerómetro, giróscopo y magnetómetro). Las actividades objetivo son conocidas como ‘atómicas’ (andar a distintas velocidades, estar de pie, correr, estar sentado), esto es, actividades que constituyen unidades de actividades más complejas como pueden ser lavar los platos o ir al trabajo. De este modo, se usan algoritmos de clasificación sencillos que puedan ser integrados en un móvil como el Naïve Bayes, Tablas y Árboles de Decisión. Además, se pretende igualmente detectar la posición en la que el usuario lleva el móvil, no sólo con el objetivo de utilizar esa información para elegir un clasificador entrenado sólo con datos recogidos en la posición correspondiente (estrategia que mejora los resultados de estimación de la actividad), sino también para la generación de un evento que puede producir la ejecución de una acción. Finalmente, el trabajo incluye un análisis de las prestaciones de la clasificación variando el tipo de parámetros y el número de sensores usados y teniendo en cuenta no sólo la precisión de la clasificación sino también la carga computacional. Por otra parte, se ha propuesto un algoritmo basado en la cuenta de pasos utilizando informaiii ción proveniente de un acelerómetro colocado en el pie del usuario. El objetivo final es detectar la actividad que el usuario está haciendo junto con la estimación aproximada de la distancia recorrida. El algoritmo de cuenta pasos se basa en la detección de máximos y mínimos usando ventanas temporales y umbrales sin requerir información específica del usuario. El ámbito de seguimiento de peatones en interiores es interesante por la falta de un estándar de localización en este tipo de entornos. Se ha diseñado un filtro extendido de Kalman centralizado y ligeramente acoplado para fusionar la información medida por un acelerómetro colocado en el pie del usuario con medidas de posición. Se han aplicado también diferentes técnicas de corrección de errores como las de velocidad cero que se basan en la detección de los instantes en los que el pie está apoyado en el suelo. Los resultados han sido obtenidos en entornos interiores usando las posiciones estimadas por un sistema de triangulación basado en la medida de la potencia recibida (RSS) y GPS en exteriores. Finalmente, se han implementado algunas aplicaciones que prueban la utilidad del trabajo desarrollado. En primer lugar se ha considerado una aplicación de monitorización de actividad que proporciona al usuario información sobre el nivel de actividad que realiza durante un período de tiempo. El objetivo final es favorecer el cambio de comportamientos sedentarios, consiguiendo hábitos saludables. Se han desarrollado dos versiones de esta aplicación. En el primer caso se ha integrado el algoritmo de cuenta pasos en una plataforma OSGi móvil adquiriendo los datos de un acelerómetro Bluetooth colocado en el pie. En el segundo caso se ha creado la misma aplicación utilizando las implementaciones de los clasificadores en un dispositivo Android. Por otro lado, se ha planteado el diseño de una aplicación para la creación automática de un diario de viaje a partir de la detección de eventos importantes. Esta aplicación toma como entrada la información procedente de la estimación de actividad y de localización además de información almacenada en bases de datos abiertas (fotos, información sobre sitios) e información sobre sensores reales y virtuales (agenda, cámara, etc.) del móvil. Abstract Inertial sensors (accelerometers and gyroscopes) have been gradually embedded in the devices that people use in their daily lives thanks to their miniaturization. Nowadays all smartphones have at least one embedded magnetometer and accelerometer, containing the most upto- date ones gyroscopes and barometers. This issue, together with the fact that the penetration of smartphones is growing steadily, has made possible the design of systems that rely on the information gathered by wearable sensors (in the future contained in smart textiles) or inertial sensors embedded in a smartphone. The role of these sensors has become key to the development of context-aware and ambient intelligent applications. Some examples are the performance of rehabilitation exercises, the provision of information related to the place that the user is visiting or the interaction with objects by gesture recognition. The work of this thesis contributes to explore to which extent this kind of sensors can be useful to support activity recognition and pedestrian tracking, which have been proven to be essential for these applications. Regarding the recognition of the activity that a user performs, the use of sensors embedded in a smartphone (proximity and light sensors, gyroscopes, magnetometers and accelerometers) has been explored. The activities that are detected belong to the group of the ones known as ‘atomic’ activities (e.g. walking at different paces, running, standing), that is, activities or movements that are part of more complex activities such as doing the dishes or commuting. Simple, wellknown classifiers that can run embedded in a smartphone have been tested, such as Naïve Bayes, Decision Tables and Trees. In addition to this, another aim is to estimate the on-body position in which the user is carrying the mobile phone. The objective is not only to choose a classifier that has been trained with the corresponding data in order to enhance the classification but also to start actions. Finally, the performance of the different classifiers is analysed, taking into consideration different features and number of sensors. The computational and memory load of the classifiers is also measured. On the other hand, an algorithm based on step counting has been proposed. The acceleration information is provided by an accelerometer placed on the foot. The aim is to detect the activity that the user is performing together with the estimation of the distance covered. The step counting strategy is based on detecting minima and its corresponding maxima. Although the counting strategy is not innovative (it includes time windows and amplitude thresholds to prevent under or overestimation) no user-specific information is required. The field of pedestrian tracking is crucial due to the lack of a localization standard for this kind of environments. A loosely-coupled centralized Extended Kalman Filter has been proposed to perform the fusion of inertial and position measurements. Zero velocity updates have been applied whenever the foot is detected to be placed on the ground. The results have been obtained in indoor environments using a triangulation algorithm based on RSS measurements and GPS outdoors. Finally, some applications have been designed to test the usefulness of the work. The first one is called the ‘Activity Monitor’ whose aim is to prevent sedentary behaviours and to modify habits to achieve desired objectives of activity level. Two different versions of the application have been implemented. The first one uses the activity estimation based on the step counting algorithm, which has been integrated in an OSGi mobile framework acquiring the data from a Bluetooth accelerometer placed on the foot of the individual. The second one uses activity classifiers embedded in an Android smartphone. On the other hand, the design of a ‘Travel Logbook’ has been planned. The input of this application is the information provided by the activity and localization modules, external databases (e.g. pictures, points of interest, weather) and mobile embedded and virtual sensors (agenda, camera, etc.). The aim is to detect important events in the journey and gather the information necessary to store it as a journal page.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The International Standard ISO 140-5 on field measurements of airborne sound insulation of façades establishes that the directivity of the measurement loudspeaker should be such that the variation in the local direct sound pressure level (ΔSPL) on the sample is ΔSPL < 5 dB (or ΔSPL < 10 dB for large façades). This condition is usually not very easy to accomplish nor is it easy to verify whether the loudspeaker produces such a uniform level. Direct sound pressure levels on the ISO standard façade essentially depend on the distance and directivity of the loudspeaker used. This paper presents a comprehensive analysis of the test geometry for measuring sound insulation and explains how the loudspeaker directivity, combined with distance, affects the acoustic level distribution on the façade. The first sections of the paper are focused on analysing the measurement geometry and its influence on the direct acoustic level variations on the façade. The most favourable and least favourable positions to minimise these direct acoustic level differences are found, and the angles covered by the façade in the reference system of the loudspeaker are also determined. Then, the maximum dimensions of the façade that meet the conditions of the ISO 140-5 standard are obtained for the ideal omnidirectional sound source and the piston radiating in an infinite baffle, which is chosen as the typical radiation pattern for loudspeakers. Finally, a complete study of the behaviour of different loudspeaker radiation models (such as those usually utilised in the ISO 140-5 measurements) is performed, comparing their radiation maps on the façade for searching their maximum dimensions and the most appropriate radiation configurations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Las pruebas de software (Testing) son en la actualidad la técnica más utilizada para la validación y la evaluación de la calidad de un programa. El testing está integrado en todas las metodologías prácticas de desarrollo de software y juega un papel crucial en el éxito de cualquier proyecto de software. Desde las unidades de código más pequeñas a los componentes más complejos, su integración en un sistema de software y su despliegue a producción, todas las piezas de un producto de software deben ser probadas a fondo antes de que el producto de software pueda ser liberado a un entorno de producción. La mayor limitación del testing de software es que continúa siendo un conjunto de tareas manuales, representando una buena parte del coste total de desarrollo. En este escenario, la automatización resulta fundamental para aliviar estos altos costes. La generación automática de casos de pruebas (TCG, del inglés test case generation) es el proceso de generar automáticamente casos de prueba que logren un alto recubrimiento del programa. Entre la gran variedad de enfoques hacia la TCG, esta tesis se centra en un enfoque estructural de caja blanca, y más concretamente en una de las técnicas más utilizadas actualmente, la ejecución simbólica. En ejecución simbólica, el programa bajo pruebas es ejecutado con expresiones simbólicas como argumentos de entrada en lugar de valores concretos. Esta tesis se basa en un marco general para la generación automática de casos de prueba dirigido a programas imperativos orientados a objetos (Java, por ejemplo) y basado en programación lógica con restricciones (CLP, del inglés constraint logic programming). En este marco general, el programa imperativo bajo pruebas es primeramente traducido a un programa CLP equivalente, y luego dicho programa CLP es ejecutado simbólicamente utilizando los mecanismos de evaluación estándar de CLP, extendidos con operaciones especiales para el tratamiento de estructuras de datos dinámicas. Mejorar la escalabilidad y la eficiencia de la ejecución simbólica constituye un reto muy importante. Es bien sabido que la ejecución simbólica resulta impracticable debido al gran número de caminos de ejecución que deben ser explorados y a tamaño de las restricciones que se deben manipular. Además, la generación de casos de prueba mediante ejecución simbólica tiende a producir un número innecesariamente grande de casos de prueba cuando es aplicada a programas de tamaño medio o grande. Las contribuciones de esta tesis pueden ser resumidas como sigue. (1) Se desarrolla un enfoque composicional basado en CLP para la generación de casos de prueba, el cual busca aliviar el problema de la explosión de caminos interprocedimiento analizando de forma separada cada componente (p.ej. método) del programa bajo pruebas, almacenando los resultados y reutilizándolos incrementalmente hasta obtener resultados para el programa completo. También se ha desarrollado un enfoque composicional basado en especialización de programas (evaluación parcial) para la herramienta de ejecución simbólica Symbolic PathFinder (SPF). (2) Se propone una metodología para usar información del consumo de recursos del programa bajo pruebas para guiar la ejecución simbólica hacia aquellas partes del programa que satisfacen una determinada política de recursos, evitando la exploración de aquellas partes del programa que violan dicha política. (3) Se propone una metodología genérica para guiar la ejecución simbólica hacia las partes más interesantes del programa, la cual utiliza abstracciones como generadores de trazas para guiar la ejecución de acuerdo a criterios de selección estructurales. (4) Se propone un nuevo resolutor de restricciones, el cual maneja eficientemente restricciones sobre el uso de la memoria dinámica global (heap) durante ejecución simbólica, el cual mejora considerablemente el rendimiento de la técnica estándar utilizada para este propósito, la \lazy initialization". (5) Todas las técnicas propuestas han sido implementadas en el sistema PET (el enfoque composicional ha sido también implementado en la herramienta SPF). Mediante evaluación experimental se ha confirmado que todas ellas mejoran considerablemente la escalabilidad y eficiencia de la ejecución simbólica y la generación de casos de prueba. ABSTRACT Testing is nowadays the most used technique to validate software and assess its quality. It is integrated into all practical software development methodologies and plays a crucial role towards the success of any software project. From the smallest units of code to the most complex components and their integration into a software system and later deployment; all pieces of a software product must be tested thoroughly before a software product can be released. The main limitation of software testing is that it remains a mostly manual task, representing a large fraction of the total development cost. In this scenario, test automation is paramount to alleviate such high costs. Test case generation (TCG) is the process of automatically generating test inputs that achieve high coverage of the system under test. Among a wide variety of approaches to TCG, this thesis focuses on structural (white-box) TCG, where one of the most successful enabling techniques is symbolic execution. In symbolic execution, the program under test is executed with its input arguments being symbolic expressions rather than concrete values. This thesis relies on a previously developed constraint-based TCG framework for imperative object-oriented programs (e.g., Java), in which the imperative program under test is first translated into an equivalent constraint logic program, and then such translated program is symbolically executed by relying on standard evaluation mechanisms of Constraint Logic Programming (CLP), extended with special treatment for dynamically allocated data structures. Improving the scalability and efficiency of symbolic execution constitutes a major challenge. It is well known that symbolic execution quickly becomes impractical due to the large number of paths that must be explored and the size of the constraints that must be handled. Moreover, symbolic execution-based TCG tends to produce an unnecessarily large number of test cases when applied to medium or large programs. The contributions of this dissertation can be summarized as follows. (1) A compositional approach to CLP-based TCG is developed which overcomes the inter-procedural path explosion by separately analyzing each component (method) in a program under test, stowing the results as method summaries and incrementally reusing them to obtain whole-program results. A similar compositional strategy that relies on program specialization is also developed for the state-of-the-art symbolic execution tool Symbolic PathFinder (SPF). (2) Resource-driven TCG is proposed as a methodology to use resource consumption information to drive symbolic execution towards those parts of the program under test that comply with a user-provided resource policy, avoiding the exploration of those parts of the program that violate such policy. (3) A generic methodology to guide symbolic execution towards the most interesting parts of a program is proposed, which uses abstractions as oracles to steer symbolic execution through those parts of the program under test that interest the programmer/tester most. (4) A new heap-constraint solver is proposed, which efficiently handles heap-related constraints and aliasing of references during symbolic execution and greatly outperforms the state-of-the-art standard technique known as lazy initialization. (5) All techniques above have been implemented in the PET system (and some of them in the SPF tool). Experimental evaluation has confirmed that they considerably help towards a more scalable and efficient symbolic execution and TCG.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Foliage Penetration (FOPEN) radar systems were introduced in 1960, and have been constantly improved by several organizations since that time. The use of Synthetic Aperture Radar (SAR) approaches for this application has important advantages, due to the need for high resolution in two dimensions. The design of this type of systems, however, includes some complications that are not present in standard SAR systems. FOPEN SAR systems need to operate with a low central frequency (VHF or UHF bands) in order to be able to penetrate the foliage. High bandwidth is also required to obtain high resolution. Due to the low central frequency, large integration angles are required during SAR image formation, and therefore the Range Migration Algorithm (RMA) is used. This project thesis identifies the three main complications that arise due to these requirements. First, a high fractional bandwidth makes narrowband propagation models no longer valid. Second, the VHF and UHF bands are used by many communications systems. The transmitted signal spectrum needs to be notched to avoid interfering them. Third, those communications systems cause Radio Frequency Interference (RFI) on the received signal. The thesis carries out a thorough analysis of the three problems, their degrading effects and possible solutions to compensate them. The UWB model is applied to the SAR signal, and the degradation induced by it is derived. The result is tested through simulation of both a single pulse stretch processor and the complete RMA image formation. Both methods show that the degradation is negligible, and therefore the UWB propagation effect does not need compensation. A technique is derived to design a notched transmitted signal. Then, its effect on the SAR image formation is evaluated analytically. It is shown that the stretch processor introduces a processing gain that reduces the degrading effects of the notches. The remaining degrading effect after processing gain is assessed through simulation, and an experimental graph of degradation as a function of percentage of nulled frequencies is obtained. The RFI is characterized and its effect on the SAR processor is derived. Once again, a processing gain is found to be introduced by the receiver. As the RFI power can be much higher than that of the desired signal, an algorithm is proposed to remove the RFI from the received signal before RMA processing. This algorithm is a modification of the Chirp Least Squares Algorithm (CLSA) explained in [4], which adapts it to deramped signals. The algorithm is derived analytically and then its performance is evaluated through simulation, showing that it is effective in removing the RFI and reducing the degradation caused by both RFI and notching. Finally, conclusions are drawn as to the importance of each one of the problems in SAR system design.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

For safety barriers the load bearing capacity of the glass when subjected to the soft body impact should be verified. The soft body pendulum test became a testing standard to classify safety glass plates. The classification of the safety glass do not consider the structural behavior when one sheet of a laminated glass is broken; in situations when the replacement of the plate could not be very urgent, structural behavior should be evaluated. The main objective of this paper is to present the structural behavior o laminated glass plates, though modal test and human impact test, including the post fracture behavior for the laminated cases. A god reproducibility and repeatability is obtained. Two main aspects of the structural behavior can be observed: the increment of the rupture load for laminated plates after the failure of the first sheet, and some similarities with a tempered monolithic behavior of equivalent thickness.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

En este documento se detalla, la planificación y elaboración de un paquete que respeta el estándar S4 de programación en lenguaje R. El paquete consiste en una serie de métodos y clases para la generación de exámenes tipos test y soluciones a partir de un archivo xls, que hace las funciones de una base de datos. El diseño propuesto está orientado a objetos y desarrolla un conjunto de clases que representan los contenidos de una prueba de evaluación tipo test: enunciados, peguntas y respuestas. Se ha realizado una implementación sencilla de un prototipo con las funciones básicas necesarias para generar los tests. Además se ha generado la documentación necesaria para crear el paquete, esto significa que cada método tiene una página de ayuda, que se podrá consultar desde un terminal con R, dicha documentación incluye ejemplos de ejecución de cada método.---ABSTRACT---In this document is detailed the elaboration and development of a package that meets the standard S4 of programming language R. This package consists of a group of methods and classes used for the generation of test exams and their solutions starting from a xls format file wich plays the role of a data base at the same time. These classes have been grouped in a way that the user could have a complete and easy vision of them. This division has been done by using data storage and functions whose tasks are more or less the same. Furthermore, the necessary documentation to create this package has also been generated, that means that every method has a help page wich can be called from a R terminal if necessary. This documentation has examples of the execution of every method.