76 resultados para Propagation waves


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nonparametric belief propagation (NBP) is a well-known particle-based method for distributed inference in wireless networks. NBP has a large number of applications, including cooperative localization. However, in loopy networks NBP suffers from similar problems as standard BP, such as over-confident beliefs and possible nonconvergence. Tree-reweighted NBP (TRW-NBP) can mitigate these problems, but does not easily lead to a distributed implementation due to the non-local nature of the required so-called edge appearance probabilities. In this paper, we propose a variation of TRWNBP, suitable for cooperative localization in wireless networks. Our algorithm uses a fixed edge appearance probability for every edge, and can outperform standard NBP in dense wireless networks.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Belief propagation (BP) is a technique for distributed inference in wireless networks and is often used even when the underlying graphical model contains cycles. In this paper, we propose a uniformly reweighted BP scheme that reduces the impact of cycles by weighting messages by a constant ?edge appearance probability? rho ? 1. We apply this algorithm to distributed binary hypothesis testing problems (e.g., distributed detection) in wireless networks with Markov random field models. We demonstrate that in the considered setting the proposed method outperforms standard BP, while maintaining similar complexity. We then show that the optimal ? can be approximated as a simple function of the average node degree, and can hence be computed in a distributed fashion through a consensus algorithm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Tree-reweighted belief propagation is a message passing method that has certain advantages compared to traditional belief propagation (BP). However, it fails to outperform BP in a consistent manner, does not lend itself well to distributed implementation, and has not been applied to distributions with higher-order interactions. We propose a method called uniformly-reweighted belief propagation that mitigates these drawbacks. After having shown in previous works that this method can substantially outperform BP in distributed inference with pairwise interaction models, in this paper we extend it to higher-order interactions and apply it to LDPC decoding, leading performance gains over BP.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, a new countermeasure against power and electromagnetic (EM) Side Channel Attacks (SCA) on FPGA implemented cryptographic algorithms is proposed. This structure mainly focuses on a critical vulnerability, Early Evaluation, also known as Early Propagation Effect (EPE), which exists in most conventional SCA-hardened DPL (Dual-rail with Precharge Logic) solutions. The main merit of this proposal is that the EPE can be effectively prevented by using a synchronized non regular precharge network, which maintains identical routing between the original and mirror parts, where costs and design complexity compared with previous EPE-resistant countermeasures are reduced, while security level is not sacrificed. Another advantage for our Precharge Absorbed(PA) - DPL method is that its Dual-Core style (independent architecture for true and false parts) could be generated using partial reconfiguration. This helps to get a dynamic security protection with better energy planning. That means system only keeps the true part which fulfills the normal en/decryption task in low security level, and reconfigures the false parts once high security level is required. A relatively limited clock speed is a compromise, since signal propagation is restricted to a portion of the clock period. In this paper, we explain the principles of PA-DPL and provide the guidelines to design this structure. We experimentally validate our methods in a minimized AES co-processor on Xilinx Virtex-5 board using electromagnetic (EM) attacks.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We develop a novel remote sensing technique for the observation of waves on the ocean surface. Our method infers the 3-D waveform and radiance of oceanic sea states via a variational stereo imagery formulation. In this setting, the shape and radiance of the wave surface are given by minimizers of a composite energy functional that combines a photometric matching term along with regularization terms involving the smoothness of the unknowns. The desired ocean surface shape and radiance are the solution of a system of coupled partial differential equations derived from the optimality conditions of the energy functional. The proposed method is naturally extended to study the spatiotemporal dynamics of ocean waves and applied to three sets of stereo video data. Statistical and spectral analysis are carried out. Our results provide evidence that the observed omnidirectional wavenumber spectrum S(k) decays as k-2.5 is in agreement with Zakharov's theory (1999). Furthermore, the 3-D spectrum of the reconstructed wave surface is exploited to estimate wave dispersion and currents.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Strong high-order Rayleigh or Sezawa modes, in addition to the fundamental Rayleigh mode, have been observed in ZnO/GaAs(001) systems along the [110] propagation direction of GaAs. The dispersion of the different acoustic waves has been calculated and compared to the experimental data. The bandwidth and impedance matching characteristics of the multimode SAW delay lines operating at high frequencies (2.5-3.5 GHz regime) have been investigated.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There is no unanimous consensus yet on the propagation mechanism before the break point inside tunnels. Some deem that the propagation mechanism follows the free space model, others argue that it should be described by the multimode waveguide model. Firstly, this paper analyzes the propagation loss in two mechanisms. Then, by conjunctively using the propagation theory and the three-dimensional solid geometry, a generic analytical model for the boundary between the free space mechanism and the multi-mode waveguide mechanism inside tunnels has been presented. Three measurement campaigns validate the model in different tunnels at different frequencies. Furthermore, the condition of the validity of the free space model used in tunnel environment has been discussed in some specific situations. Finally, through mathematical derivation, the seemingly conflicting viewpoints on the free space mechanism and the multi-mode waveguide mechanism have been unified in some specific situations by the presented generic model. The results in this paper can be helpful to gain deeper insight and better understanding of the propagation mechanism inside tunnels

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper investigates the propagation of airblast from quarry blasting. Peak overpressure is calculated as a function of blasting parameters (explosive mass per delay and velocity at which the detonation sequence proceeds along the bench) and polar coordinates of the point of interest (distance to the blast and azimuth with respect to the free face of the blast). The model is in the form of the product of a classical scaled distance attenuation law times a directional correction factor. The latter considers the influence of the bench face, and attenuates overpressure at the top level and amplifies it at the bottom. Such factor also accounts for the effect of the delay by amplifying the pressure in the direction of the initiation sequence if the velocity of initiation exceeds half the speed of sound and up to an initiation velocity in the range of the speed of sound. The model has been fitted to an empirical data set composed by 134 airblast records monitored in 47 blasts at two quarries. The measurements were made at distances to the blast less than 450 m. The model is statistically significant and has a determination coefficient of 0.869

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The propagation losses (PL) of lithium niobate optical planar waveguides fabricated by swift heavy-ion irradiation (SHI), an alternative to conventional ion implantation, have been investigated and optimized. For waveguide fabrication, congruently melting LiNbO3 substrates were irradiated with F ions at 20 MeV or 30 MeV and fluences in the range 1013–1014 cm−2. The influence of the temperature and time of post-irradiation annealing treatments has been systematically studied. Optimum propagation losses lower than 0.5 dB/cm have been obtained for both TE and TM modes, after a two-stage annealing treatment at 350 and 375∘C. Possible loss mechanisms are discussed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We report synchronization of networked excitable nodes embedded in a metric space, where the connectivity properties are mostly determined by the distance between units. Such a high clustered structure, combined with the lack of long-range connections, prevents full synchronization and yields instead the emergence of synchronization waves. We show that this regime is optimal for information transmission through the system, as it enhances the options of reconstructing the topology from the dynamics. Measurements of topological and functional centralities reveal that the wave-synchronization state allows detection of the most structurally relevant nodes from a single observation of the dynamics, without any a priori information on the model equations ruling the evolution of the ensemble

Relevância:

20.00% 20.00%

Publicador:

Resumo:

El estudio de vibraciones tiene por objetivo fijar una banda de seguridad para proteger la nueva línea ferroviaria de alta velocidad Madrid – Vitoria – San Sebastián (tramo Andoain – Urnieta).Las tomas de datos han permitido el cálculo de unas leyes de amortiguación de la propagación de las ondas de vibración que permiten predecir cuál será la magnitud de vibración generada a una cierta distancia de la explotación cuando se detona una carga de explosivo determinada. Así mismo, se lleva a cabo un análisis de las concentraciones de energía (vibraciones) que podrían producirse durante una voladura tipo de producción atendiendo a diferentes secuencias de disparo. Con este método, basado en el disparo de un “barreno semilla” de similares condiciones de carga y arranque que un barreno típico de producción, se pretende resaltar la importancia que tiene el rango de las frecuencias en el proceso de generación y transmisión de vibraciones en el terreno, y consecuentemente en los criterios de protección de estructuras frente a tales vibraciones. ABSTRACT The aim of the vibration study is to set a security band to protect the Spanish high speed train (AVE) which goes Madrid-Vitoria-San Sebastian (section Andoain- Urnieta). The data gathered has made possible the calculation of absorption laws of propagation of vibrating waves, that allow the prediction of the magnitude created at a certain distance of the detonation of a known explosive. In addition, an analysis of the energy concentrations that could appear during a blasting in different sequences is made. With this method based on the firing of a seed drill with similar load and boot conditions of a common production drill, is intended to highlight the importance of the frequency range in the process of vibrations transmission and generation in the field, and consequently the importance in the structures protection criteria against these vibrations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

El retroceso de las costas acantiladas es un fenómeno muy extendido sobre los litorales rocosos expuestos a la incidencia combinada de los procesos marinos y meteorológicos que se dan en la franja costera. Este fenómeno se revela violentamente como movimientos gravitacionales del terreno esporádicos, pudiendo causar pérdidas materiales y/o humanas. Aunque el conocimiento de estos riesgos de erosión resulta de vital importancia para la correcta gestión de la costa, el desarrollo de modelos predictivos se encuentra limitado desde el punto de vista geomorfológico debido a la complejidad e interacción de los procesos de desarrollo espacio-temporal que tienen lugar en la zona costera. Los modelos de predicción publicados son escasos y con importantes inconvenientes: a) extrapolación, extienden la información de registros históricos; b) empíricos, sobre registros históricos estudian la respuesta al cambio de un parámetro; c) estocásticos, determinan la cadencia y magnitud de los eventos futuros extrapolando las distribuciones de probabilidad extraídas de catálogos históricos; d) proceso-respuesta, de estabilidad y propagación del error inexplorada; e) en Ecuaciones en Derivadas Parciales, computacionalmente costosos y poco exactos. La primera parte de esta tesis detalla las principales características de los modelos más recientes de cada tipo y, para los más habitualmente utilizados, se indican sus rangos de aplicación, ventajas e inconvenientes. Finalmente como síntesis de los procesos más relevantes que contemplan los modelos revisados, se presenta un diagrama conceptual de la recesión costera, donde se recogen los procesos más influyentes que deben ser tenidos en cuenta, a la hora de utilizar o crear un modelo de recesión costera con el objetivo de evaluar la peligrosidad (tiempo/frecuencia) del fenómeno a medio-corto plazo. En esta tesis se desarrolla un modelo de proceso-respuesta de retroceso de acantilados costeros que incorpora el comportamiento geomecánico de materiales cuya resistencia a compresión no supere los 5 MPa. El modelo simula la evolución espaciotemporal de un perfil-2D del acantilado que puede estar formado por materiales heterogéneos. Para ello, se acoplan la dinámica marina: nivel medio del mar, cambios en el nivel medio del lago, mareas y oleaje; con la evolución del terreno: erosión, desprendimiento rocoso y formación de talud de derrubios. El modelo en sus diferentes variantes es capaz de incluir el análisis de la estabilidad geomecánica de los materiales, el efecto de los derrubios presentes al pie del acantilado, el efecto del agua subterránea, la playa, el run-up, cambios en el nivel medio del mar o cambios (estacionales o interanuales) en el nivel medio de la masa de agua (lagos). Se ha estudiado el error de discretización del modelo y su propagación en el tiempo a partir de las soluciones exactas para los dos primeros periodos de marea para diferentes aproximaciones numéricas tanto en tiempo como en espacio. Los resultados obtenidos han permitido justificar las elecciones que minimizan el error y los métodos de aproximación más adecuados para su posterior uso en la modelización. El modelo ha sido validado frente a datos reales en la costa de Holderness, Yorkshire, Reino Unido; y en la costa norte del lago Erie, Ontario, Canadá. Los resultados obtenidos presentan un importante avance en los modelos de recesión costera, especialmente en su relación con las condiciones geomecánicas del medio, la influencia del agua subterránea, la verticalización de los perfiles rocosos y su respuesta ante condiciones variables producidas por el cambio climático (por ejemplo, nivel medio del mar, cambios en los niveles de lago, etc.). The recession of coastal cliffs is a widespread phenomenon on the rocky shores that are exposed to the combined incidence of marine and meteorological processes that occur in the shoreline. This phenomenon is revealed violently and occasionally, as gravitational movements of the ground and can cause material or human losses. Although knowledge of the risks of erosion is vital for the proper management of the coast, the development of cliff erosion predictive models is limited by the complex interactions between environmental processes and material properties over a range of temporal and spatial scales. Published prediction models are scarce and present important drawbacks: extrapolation, that extend historical records to the future; empirical, that based on historical records studies the system response against the change in one parameter; stochastic, that represent of cliff behaviour based on assumptions regarding the magnitude and frequency of events in a probabilistic framework based on historical records; process-response, stability and error propagation unexplored; PDE´s, highly computationally expensive and not very accurate. The first part of this thesis describes the main features of the latest models of each type and, for the most commonly used, their ranges of application, advantages and disadvantages are given. Finally as a synthesis of the most relevant processes that include the revised models, a conceptual diagram of coastal recession is presented. This conceptual model includes the most influential processes that must be taken into account when using or creating a model of coastal recession to evaluate the dangerousness (time/frequency) of the phenomenon to medium-short term. A new process-response coastal recession model developed in this thesis has been designed to incorporate the behavioural and mechanical characteristics of coastal cliffs which are composed of with materials whose compressive strength is less than 5 MPa. The model simulates the spatial and temporal evolution of a cliff-2D profile that can consist of heterogeneous materials. To do so, marine dynamics: mean sea level, waves, tides, lake seasonal changes; is coupled with the evolution of land recession: erosion, cliff face failure and associated protective colluvial wedge. The model in its different variants can include analysis of material geomechanical stability, the effect of debris present at the cliff foot, groundwater effects, beach and run-up effects, changes in the mean sea level or changes (seasonal or inter-annual) in the mean lake level. Computational implementation and study of different numerical resolution techniques, in both time and space approximations, and the produced errors are exposed and analysed for the first two tidal periods. The results obtained in the errors analysis allow us to operate the model with a configuration that minimizes the error of the approximation methods. The model is validated through profile evolution assessment at various locations of coastline retreat on the Holderness Coast, Yorkshire, UK and on the north coast of Lake Erie, Ontario, Canada. The results represent an important stepforward in linking material properties to the processes of cliff recession, in considering the effect of groundwater charge and the slope oversteeping and their response to changing conditions caused by climate change (i.e. sea level, changes in lakes levels, etc.).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

One important task in the design of an antenna is to carry out an analysis to find out the characteristics of the antenna that best fulfills the specifications fixed by the application. After that, a prototype is manufactured and the next stage in design process is to check if the radiation pattern differs from the designed one. Besides the radiation pattern, other radiation parameters like directivity, gain, impedance, beamwidth, efficiency, polarization, etc. must be also evaluated. For this purpose, accurate antenna measurement techniques are needed in order to know exactly the actual electromagnetic behavior of the antenna under test. Due to this fact, most of the measurements are performed in anechoic chambers, which are closed areas, normally shielded, covered by electromagnetic absorbing material, that simulate free space propagation conditions, due to the absorption of the radiation absorbing material. Moreover, these facilities can be employed independently of the weather conditions and allow measurements free from interferences. Despite all the advantages of the anechoic chambers, the results obtained both from far-field measurements and near-field measurements are inevitably affected by errors. Thus, the main objective of this Thesis is to propose algorithms to improve the quality of the results obtained in antenna measurements by using post-processing techniques and without requiring additional measurements. First, a deep revision work of the state of the art has been made in order to give a general vision of the possibilities to characterize or to reduce the effects of errors in antenna measurements. Later, new methods to reduce the unwanted effects of four of the most commons errors in antenna measurements are described and theoretical and numerically validated. The basis of all them is the same, to perform a transformation from the measurement surface to another domain where there is enough information to easily remove the contribution of the errors. The four errors analyzed are noise, reflections, truncation errors and leakage and the tools used to suppress them are mainly source reconstruction techniques, spatial and modal filtering and iterative algorithms to extrapolate functions. Therefore, the main idea of all the methods is to modify the classical near-field-to-far-field transformations by including additional steps with which errors can be greatly suppressed. Moreover, the proposed methods are not computationally complex and, because they are applied in post-processing, additional measurements are not required. The noise is the most widely studied error in this Thesis, proposing a total of three alternatives to filter out an important noise contribution before obtaining the far-field pattern. The first one is based on a modal filtering. The second alternative uses a source reconstruction technique to obtain the extreme near-field where it is possible to apply a spatial filtering. The last one is to back-propagate the measured field to a surface with the same geometry than the measurement surface but closer to the AUT and then to apply also a spatial filtering. All the alternatives are analyzed in the three most common near-field systems, including comprehensive noise statistical analyses in order to deduce the signal-to-noise ratio improvement achieved in each case. The method to suppress reflections in antenna measurements is also based on a source reconstruction technique and the main idea is to reconstruct the field over a surface larger than the antenna aperture in order to be able to identify and later suppress the virtual sources related to the reflective waves. The truncation error presents in the results obtained from planar, cylindrical and partial spherical near-field measurements is the third error analyzed in this Thesis. The method to reduce this error is based on an iterative algorithm to extrapolate the reliable region of the far-field pattern from the knowledge of the field distribution on the AUT plane. The proper termination point of this iterative algorithm as well as other critical aspects of the method are also studied. The last part of this work is dedicated to the detection and suppression of the two most common leakage sources in antenna measurements. A first method tries to estimate the leakage bias constant added by the receiver’s quadrature detector to every near-field data and then suppress its effect on the far-field pattern. The second method can be divided into two parts; the first one to find the position of the faulty component that radiates or receives unwanted radiation, making easier its identification within the measurement environment and its later substitution; and the second part of this method is able to computationally remove the leakage effect without requiring the substitution of the faulty component. Resumen Una tarea importante en el diseño de una antena es llevar a cabo un análisis para averiguar las características de la antena que mejor cumple las especificaciones fijadas por la aplicación. Después de esto, se fabrica un prototipo de la antena y el siguiente paso en el proceso de diseño es comprobar si el patrón de radiación difiere del diseñado. Además del patrón de radiación, otros parámetros de radiación como la directividad, la ganancia, impedancia, ancho de haz, eficiencia, polarización, etc. deben ser también evaluados. Para lograr este propósito, se necesitan técnicas de medida de antenas muy precisas con el fin de saber exactamente el comportamiento electromagnético real de la antena bajo prueba. Debido a esto, la mayoría de las medidas se realizan en cámaras anecoicas, que son áreas cerradas, normalmente revestidas, cubiertas con material absorbente electromagnético. Además, estas instalaciones se pueden emplear independientemente de las condiciones climatológicas y permiten realizar medidas libres de interferencias. A pesar de todas las ventajas de las cámaras anecoicas, los resultados obtenidos tanto en medidas en campo lejano como en medidas en campo próximo están inevitablemente afectados por errores. Así, el principal objetivo de esta Tesis es proponer algoritmos para mejorar la calidad de los resultados obtenidos en medida de antenas mediante el uso de técnicas de post-procesado. Primeramente, se ha realizado un profundo trabajo de revisión del estado del arte con el fin de dar una visión general de las posibilidades para caracterizar o reducir los efectos de errores en medida de antenas. Después, se han descrito y validado tanto teórica como numéricamente nuevos métodos para reducir el efecto indeseado de cuatro de los errores más comunes en medida de antenas. La base de todos ellos es la misma, realizar una transformación de la superficie de medida a otro dominio donde hay suficiente información para eliminar fácilmente la contribución de los errores. Los cuatro errores analizados son ruido, reflexiones, errores de truncamiento y leakage y las herramientas usadas para suprimirlos son principalmente técnicas de reconstrucción de fuentes, filtrado espacial y modal y algoritmos iterativos para extrapolar funciones. Por lo tanto, la principal idea de todos los métodos es modificar las transformaciones clásicas de campo cercano a campo lejano incluyendo pasos adicionales con los que los errores pueden ser enormemente suprimidos. Además, los métodos propuestos no son computacionalmente complejos y dado que se aplican en post-procesado, no se necesitan medidas adicionales. El ruido es el error más ampliamente estudiado en esta Tesis, proponiéndose un total de tres alternativas para filtrar una importante contribución de ruido antes de obtener el patrón de campo lejano. La primera está basada en un filtrado modal. La segunda alternativa usa una técnica de reconstrucción de fuentes para obtener el campo sobre el plano de la antena donde es posible aplicar un filtrado espacial. La última es propagar el campo medido a una superficie con la misma geometría que la superficie de medida pero más próxima a la antena y luego aplicar también un filtrado espacial. Todas las alternativas han sido analizadas en los sistemas de campo próximos más comunes, incluyendo detallados análisis estadísticos del ruido con el fin de deducir la mejora de la relación señal a ruido lograda en cada caso. El método para suprimir reflexiones en medida de antenas está también basado en una técnica de reconstrucción de fuentes y la principal idea es reconstruir el campo sobre una superficie mayor que la apertura de la antena con el fin de ser capaces de identificar y después suprimir fuentes virtuales relacionadas con las ondas reflejadas. El error de truncamiento que aparece en los resultados obtenidos a partir de medidas en un plano, cilindro o en la porción de una esfera es el tercer error analizado en esta Tesis. El método para reducir este error está basado en un algoritmo iterativo para extrapolar la región fiable del patrón de campo lejano a partir de información de la distribución del campo sobre el plano de la antena. Además, se ha estudiado el punto apropiado de terminación de este algoritmo iterativo así como otros aspectos críticos del método. La última parte de este trabajo está dedicado a la detección y supresión de dos de las fuentes de leakage más comunes en medida de antenas. El primer método intenta realizar una estimación de la constante de fuga del leakage añadido por el detector en cuadratura del receptor a todos los datos en campo próximo y después suprimir su efecto en el patrón de campo lejano. El segundo método se puede dividir en dos partes; la primera de ellas para encontrar la posición de elementos defectuosos que radian o reciben radiación indeseada, haciendo más fácil su identificación dentro del entorno de medida y su posterior substitución. La segunda parte del método es capaz de eliminar computacionalmente el efector del leakage sin necesidad de la substitución del elemento defectuoso.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Esta Tesis Doctoral aborda el estudio de algunas técnicas no destructivas para la clasificación de madera de pino silvestre (Pinus sylvestris L.) de procedencia española y de gruesa escuadría para uso estructural. Para la estimación del módulo de elasticidad y de la resistencia se han aplicado técnicas basadas en la propagación de una onda a través de la madera: onda de ultrasonidos (Sylvatest) o de impacto (Microsecond Timer) en dirección longitudinal, o vibración en dirección longitudinal y transversal (PLG). Para la estimación de la densidad se han utilizado métodos puntuales basados en el penetrómetro (Pilodyn) y en la resistencia al arranque de un tornillo. Las variables obtenidas han sido relacionadas con los resultados de la clasificación visual y con las propiedades de la madera determinadas mediante ensayo mecánico. Además, se ha estudiado la influencia de la humedad de la madera en la velocidad de propagación de la onda para definir factores de corrección a los equipos comerciales utilizados en esta Tesis Doctoral. La muestra de estudio está formada por 244 piezas procedentes de El Espinar, Segovia, con dimensiones nominales 150 x 200 x 4000 mm (218 piezas) y 100 x 150 x 3000 mm (26 piezas). De todas las piezas se tomaron datos de dimensiones, contenido de humedad y clasificación visual según la norma UNE 56544. En las primeras 218 vigas se aplicaron las técnicas de ultrasonidos, onda de impacto y vibraciones, se determinó la densidad de cada pieza completa y se ensayaron según la norma UNE-EN 408 para obtener el módulo de elasticidad global (en todos los casos) y local (en un porcentaje), así como de la tensión de rotura. Se extrajeron tres rebanadas para los ensayos puntuales y para el cálculo de la densidad. En las otras 26 piezas se repitieron los ensayos (transmisión de onda, vibración y clasificación visual) durante el proceso de secado natural, desde que la madera se encontraba húmeda (en torno al 40 %) hasta la humedad de equilibrio higroscópico (en torno al 9%). Respecto a la clasificación visual no se han observado diferencias significativas entre la calidad MEG o las rechazadas. Se han estudiado las consecuencias del secado (principalmente las deformaciones) y no se ha encontrado justificación para que estos defectos penalicen la clasificación. Para la densidad, el mayor R2 obtenido ha sido de un 47% a partir del uso combinado de los dos equipos puntuales (penetrómetro y arranque de tornillo). Para el módulo de elasticidad y la tensión de rotura, la mejor relación se ha obtenido a partir de la técnica de vibración longitudinal, con unos coeficientes de 79% y un 52% respectivamente. Se ha estimado que el aumento de un punto porcentual en el contenido de humedad de la madera produce una pérdida de velocidad de onda del 0,58% para Sylvatest y Microsecond Timer, y del 0,71% para PLG. Estos valores son generalizables para un rango de humedades entre 9 y 25 %. Abstract This Doctoral Thesis approach the study of some non-destructive techniques as a classification method for structural use of Scots pine wood of Spanish origin with large cross section. To estimate the modulus of elasticity and strength have been used techniques based on the propagation of a wave through the timber: ultrasonic wave (Sylvatest) or stress wave (Microsecond Timer) in longitudinal direction, or vibration in longitudinal and transversal direction (PLG). Local probing methods have been applied to estimate the density, based on penetrometer (Pilodyn) and the screw withdrawal resistance meter. The different variables obtained were compared with the results of the visual grading and the values of the properties of the wood determined by the standardized test of the pieces. Furthermore, the influence of the moisture content of the wood on the velocity of propagation of the waves through the timber has been analyzed in order to establish a correction factor for the commercial devices used in this Doctoral Thesis. The sample tested consists of 244 pieces from El Espinar, Segovia, with nominal dimensions 150 x 200 x 4000 mm (218 pieces) and 100 x 150 x 3000 mm (26 pieces). Data collection about dimensions, moisture content and visual grading according to the UNE 56544 standard were carried out on all the pieces. The first 218 pieces were tested by non destructive techniques based on ultrasonic wave, stress wave and vibration, the density was measured on each piece and bending test according to the UNE-EN 408 standard was carried out for calculating the global modulus of elasticity (all the pieces) and the local one (only a representative group), as well as the bending strength. Three slices were removed for implementing the local probing and to calculate the density. In the other 26 pieces the tests (wave transmission, vibration and visual grading) were repeated during the natural drying process, from wet timber (around 40 % moisture content) up to the equilibrium moisture content (around 9%). Regarding the visual grading no significant differences were observed between MEG or rejected pieces. The effects of drying (deformations) have been studied, and justification for the specification hasn't been found. To estimate the density, the greater R2 obtained was 47% by using both penetrometer and screw withdrawal. For the modulus of elasticity and bending strength, the best relationship has been found with the longitudinal vibration, with coefficients of 79% and 52% respectively. It has been estimated that an increase of a point of the moisture content of the wood produces a decrease on the velocity obtained from ultrasonic or stress wave of 0,58%, and 0,71 % for the one obtained from vibration. Those values can be generalized for a range of moisture content from 9 to 25 %.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Corrosion of reinforcing steel in concrete due to chloride ingress is one of the main causes of the deterioration of reinforced concrete structures. Structures most affected by such a corrosion are marine zone buildings and structures exposed to de-icing salts like highways and bridges. Such process is accompanied by an increase in volume of the corrosión products on the rebarsconcrete interface. Depending on the level of oxidation, iron can expand as much as six times its original volume. This increase in volume exerts tensile stresses in the surrounding concrete which result in cracking and spalling of the concrete cover if the concrete tensile strength is exceeded. The mechanism by which steel embedded in concrete corrodes in presence of chloride is the local breakdown of the passive layer formed in the highly alkaline condition of the concrete. It is assumed that corrosion initiates when a critical chloride content reaches the rebar surface. The mathematical formulation idealized the corrosion sequence as a two-stage process: an initiation stage, during which chloride ions penetrate to the reinforcing steel surface and depassivate it, and a propagation stage, in which active corrosion takes place until cracking of the concrete cover has occurred. The aim of this research is to develop computer tools to evaluate the duration of the service life of reinforced concrete structures, considering both the initiation and propagation periods. Such tools must offer a friendly interface to facilitate its use by the researchers even though their background is not in numerical simulation. For the evaluation of the initiation period different tools have been developed: Program TavProbabilidade: provides means to carry out a probability analysis of a chloride ingress model. Such a tool is necessary due to the lack of data and general uncertainties associated with the phenomenon of the chloride diffusion. It differs from the deterministic approach because it computes not just a chloride profile at a certain age, but a range of chloride profiles for each probability or occurrence. Program TavProbabilidade_Fiabilidade: carries out reliability analyses of the initiation period. It takes into account the critical value of the chloride concentration on the steel that causes breakdown of the passive layer and the beginning of the propagation stage. It differs from the deterministic analysis in that it does not predict if the corrosion is going to begin or not, but to quantifies the probability of corrosion initiation. Program TavDif_1D: was created to do a one dimension deterministic analysis of the chloride diffusion process by the finite element method (FEM) which numerically solves Fick’second Law. Despite of the different FEM solver already developed in one dimension, the decision to create a new code (TavDif_1D) was taken because of the need to have a solver with friendly interface for pre- and post-process according to the need of IETCC. An innovative tool was also developed with a systematic method devised to compare the ability of the different 1D models to predict the actual evolution of chloride ingress based on experimental measurements, and also to quantify the degree of agreement of the models with each others. For the evaluation of the entire service life of the structure: a computer program has been developed using finite elements method to do the coupling of both service life periods: initiation and propagation. The program for 2D (TavDif_2D) allows the complementary use of two external programs in a unique friendly interface: • GMSH - an finite element mesh generator and post-processing viewer • OOFEM – a finite element solver. This program (TavDif_2D) is responsible to decide in each time step when and where to start applying the boundary conditions of fracture mechanics module in function of the amount of chloride concentration and corrosion parameters (Icorr, etc). This program is also responsible to verify the presence and the degree of fracture in each element to send the Information of diffusion coefficient variation with the crack width. • GMSH - an finite element mesh generator and post-processing viewer • OOFEM – a finite element solver. The advantages of the FEM with the interface provided by the tool are: • the flexibility to input the data such as material property and boundary conditions as time dependent function. • the flexibility to predict the chloride concentration profile for different geometries. • the possibility to couple chloride diffusion (initiation stage) with chemical and mechanical behavior (propagation stage). The OOFEM code had to be modified to accept temperature, humidity and the time dependent values for the material properties, which is necessary to adequately describe the environmental variations. A 3-D simulation has been performed to simulate the behavior of the beam on both, action of the external load and the internal load caused by the corrosion products, using elements of imbedded fracture in order to plot the curve of the deflection of the central region of the beam versus the external load to compare with the experimental data.