917 resultados para post-processing method


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstract The proliferation of wireless sensor networks and the variety of envisioned applications associated with them has motivated the development of distributed algorithms for collaborative processing over networked systems. One of the applications that has attracted the attention of the researchers is that of target localization where the nodes of the network try to estimate the position of an unknown target that lies within its coverage area. Particularly challenging is the problem of estimating the target’s position when we use received signal strength indicator (RSSI) due to the nonlinear relationship between the measured signal and the true position of the target. Many of the existing approaches suffer either from high computational complexity (e.g., particle filters) or lack of accuracy. Further, many of the proposed solutions are centralized which make their application to a sensor network questionable. Depending on the application at hand and, from a practical perspective it could be convenient to find a balance between localization accuracy and complexity. Into this direction we approach the maximum likelihood location estimation problem by solving a suboptimal (and more tractable) problem. One of the main advantages of the proposed scheme is that it allows for a decentralized implementation using distributed processing tools (e.g., consensus and convex optimization) and therefore, it is very suitable to be implemented in real sensor networks. If further accuracy is needed an additional refinement step could be performed around the found solution. Under the assumption of independent noise among the nodes such local search can be done in a fully distributed way using a distributed version of the Gauss-Newton method based on consensus. Regardless of the underlying application or function of the sensor network it is al¬ways necessary to have a mechanism for data reporting. While some approaches use a special kind of nodes (called sink nodes) for data harvesting and forwarding to the outside world, there are however some scenarios where such an approach is impractical or even impossible to deploy. Further, such sink nodes become a bottleneck in terms of traffic flow and power consumption. To overcome these issues instead of using sink nodes for data reporting one could use collaborative beamforming techniques to forward directly the generated data to a base station or gateway to the outside world. In a dis-tributed environment like a sensor network nodes cooperate in order to form a virtual antenna array that can exploit the benefits of multi-antenna communications. In col-laborative beamforming nodes synchronize their phases in order to add constructively at the receiver. Some of the inconveniences associated with collaborative beamforming techniques is that there is no control over the radiation pattern since it is treated as a random quantity. This may cause interference to other coexisting systems and fast bat-tery depletion at the nodes. Since energy-efficiency is a major design issue we consider the development of a distributed collaborative beamforming scheme that maximizes the network lifetime while meeting some quality of service (QoS) requirement at the re¬ceiver side. Using local information about battery status and channel conditions we find distributed algorithms that converge to the optimal centralized beamformer. While in the first part we consider only battery depletion due to communications beamforming, we extend the model to account for more realistic scenarios by the introduction of an additional random energy consumption. It is shown how the new problem generalizes the original one and under which conditions it is easily solvable. By formulating the problem under the energy-efficiency perspective the network’s lifetime is significantly improved. Resumen La proliferación de las redes inalámbricas de sensores junto con la gran variedad de posi¬bles aplicaciones relacionadas, han motivado el desarrollo de herramientas y algoritmos necesarios para el procesado cooperativo en sistemas distribuidos. Una de las aplicaciones que suscitado mayor interés entre la comunidad científica es la de localization, donde el conjunto de nodos de la red intenta estimar la posición de un blanco localizado dentro de su área de cobertura. El problema de la localization es especialmente desafiante cuando se usan niveles de energía de la seal recibida (RSSI por sus siglas en inglés) como medida para la localization. El principal inconveniente reside en el hecho que el nivel de señal recibida no sigue una relación lineal con la posición del blanco. Muchas de las soluciones actuales al problema de localization usando RSSI se basan en complejos esquemas centralizados como filtros de partículas, mientas que en otras se basan en esquemas mucho más simples pero con menor precisión. Además, en muchos casos las estrategias son centralizadas lo que resulta poco prácticos para su implementación en redes de sensores. Desde un punto de vista práctico y de implementation, es conveniente, para ciertos escenarios y aplicaciones, el desarrollo de alternativas que ofrezcan un compromiso entre complejidad y precisión. En esta línea, en lugar de abordar directamente el problema de la estimación de la posición del blanco bajo el criterio de máxima verosimilitud, proponemos usar una formulación subóptima del problema más manejable analíticamente y que ofrece la ventaja de permitir en¬contrar la solución al problema de localization de una forma totalmente distribuida, convirtiéndola así en una solución atractiva dentro del contexto de redes inalámbricas de sensores. Para ello, se usan herramientas de procesado distribuido como los algorit¬mos de consenso y de optimización convexa en sistemas distribuidos. Para aplicaciones donde se requiera de un mayor grado de precisión se propone una estrategia que con¬siste en la optimización local de la función de verosimilitud entorno a la estimación inicialmente obtenida. Esta optimización se puede realizar de forma descentralizada usando una versión basada en consenso del método de Gauss-Newton siempre y cuando asumamos independencia de los ruidos de medida en los diferentes nodos. Independientemente de la aplicación subyacente de la red de sensores, es necesario tener un mecanismo que permita recopilar los datos provenientes de la red de sensores. Una forma de hacerlo es mediante el uso de uno o varios nodos especiales, llamados nodos “sumidero”, (sink en inglés) que actúen como centros recolectores de información y que estarán equipados con hardware adicional que les permita la interacción con el exterior de la red. La principal desventaja de esta estrategia es que dichos nodos se convierten en cuellos de botella en cuanto a tráfico y capacidad de cálculo. Como alter¬nativa se pueden usar técnicas cooperativas de conformación de haz (beamforming en inglés) de manera que el conjunto de la red puede verse como un único sistema virtual de múltiples antenas y, por tanto, que exploten los beneficios que ofrecen las comu¬nicaciones con múltiples antenas. Para ello, los distintos nodos de la red sincronizan sus transmisiones de manera que se produce una interferencia constructiva en el recep¬tor. No obstante, las actuales técnicas se basan en resultados promedios y asintóticos, cuando el número de nodos es muy grande. Para una configuración específica se pierde el control sobre el diagrama de radiación causando posibles interferencias sobre sis¬temas coexistentes o gastando más potencia de la requerida. La eficiencia energética es una cuestión capital en las redes inalámbricas de sensores ya que los nodos están equipados con baterías. Es por tanto muy importante preservar la batería evitando cambios innecesarios y el consecuente aumento de costes. Bajo estas consideraciones, se propone un esquema de conformación de haz que maximice el tiempo de vida útil de la red, entendiendo como tal el máximo tiempo que la red puede estar operativa garantizando unos requisitos de calidad de servicio (QoS por sus siglas en inglés) que permitan una decodificación fiable de la señal recibida en la estación base. Se proponen además algoritmos distribuidos que convergen a la solución centralizada. Inicialmente se considera que la única causa de consumo energético se debe a las comunicaciones con la estación base. Este modelo de consumo energético es modificado para tener en cuenta otras formas de consumo de energía derivadas de procesos inherentes al funcionamiento de la red como la adquisición y procesado de datos, las comunicaciones locales entre nodos, etc. Dicho consumo adicional de energía se modela como una variable aleatoria en cada nodo. Se cambia por tanto, a un escenario probabilístico que generaliza el caso determinista y se proporcionan condiciones bajo las cuales el problema se puede resolver de forma eficiente. Se demuestra que el tiempo de vida de la red mejora de forma significativa usando el criterio propuesto de eficiencia energética.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Look-up tables are collected and analysed for 12 European National Travel Surveys (NTS) in a harmonized way covering the age group 13-84 year. Travel behaviour measured as kilometres, time use and trips per traveller is compared. Trips per traveller are very similar over the countries whereas kilometres differ most, from minus 28% for Spain to plus 19% and 14% for Sweden and Finland. It is shown that two main factors for differences are GDP per capita and density in the urban areas. The latter is the main reason for the low level in Spain. Mode share is except for Spain with a very high level of walking trips rather similar with a higher level of cycling in the Netherlands, more public transport in Switzerland, and more air traffic in Sweden. Normally kilometres per respondent/inhabitant is used for national planning purpose and this is very affected by the share of mobile travellers. The immobile share is varying between 8 and 28% with 6 NTS at a 15-17% level. These differences are analysed and discussed and it is concluded that the immobile share should be a little less than 15-17% because it is assessed that some short trips might have been forgotten in these 6 countries. The share has a downward tendency with higher density. The resulting immobile share is very dependent on data collection methodology, sampling method, quality of interviewer felt-work etc. The paper shows other possibilities to improve local surveys based on comparison with other countries.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Digital chaotic behavior in an optically processing element is reported. It is obtained as the result of processing two fixed trains of bits. The process is performed with an optically programmable logic gate, previously reported as a possible main block for optical computing. Outputs for some specific conditions of the circuit are given. Digital chaos is obtained using a feedback configuration. Period doublings in a Feigenbaum‐like scenario are obtained. A new method to characterize this type of digital chaos is reported.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Processing tomato industry has a high potential in Spain. Variety testing and mechanization studies and applications have been performed during the last 15 years. Many factors affect the quality and product losses during post-harvest handling which may be classified as: main or external factors: those related to the systems, procedures and devices; and fruit factors: those related to fruit properties. A research project is being carried on in the area of Vegas del Guadiana (Badajoz, Spain) to study these factors and to estimate costs, and to develop improved post-harvest handling practices.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Digital chaotic behaviour in an Optical-Processing Element is reported. It is obtained as the result of processing two fixed trains of bits. Period doublings in a Feigenbaum-like scenario have been obtained. A new method to characterize digital chaos is reported

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Stochastic model updating must be considered for quantifying uncertainties inherently existing in real-world engineering structures. By this means the statistical properties,instead of deterministic values, of structural parameters can be sought indicating the parameter variability. However, the implementation of stochastic model updating is much more complicated than that of deterministic methods particularly in the aspects of theoretical complexity and low computational efficiency. This study attempts to propose a simple and cost-efficient method by decomposing a stochastic updating process into a series of deterministic ones with the aid of response surface models and Monte Carlo simulation. The response surface models are used as surrogates for original FE models in the interest of programming simplification, fast response computation and easy inverse optimization. Monte Carlo simulation is adopted for generating samples from the assumed or measured probability distributions of responses. Each sample corresponds to an individual deterministic inverse process predicting the deterministic values of parameters. Then the parameter means and variances can be statistically estimated based on all the parameter predictions by running all the samples. Meanwhile, the analysis of variance approach is employed for the evaluation of parameter variability significance. The proposed method has been demonstrated firstly on a numerical beam and then a set of nominally identical steel plates tested in the laboratory. It is found that compared with the existing stochastic model updating methods, the proposed method presents similar accuracy while its primary merits consist in its simple implementation and cost efficiency in response computation and inverse optimization.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In living bodies, the correct perceptual representation of size constancy requires that an object's size appear the same when it changes its location with respect to the observer. At the same time, it is necessary that objects at different locations appear to be the same size if they are. In order to do that, the perceptual system must recover from the stimuli impinging on the individual, from the light falling on the retina, a representation of the relative sizes of objects in the environment. Moreover, at the same time, image perception is related to another type of phenomena. It corresponds to the well known perceptual illusions. To analyze this facts, we propose a system based on a particular arrays of receptive points composed by optical fibers and dummy fibers. The structure is based on the first layers of the mammalians primary visual cortex. At that part of the brain, the neurons located at certain columns, respond to particular directions. This orientation changes in a systematic way as one moves across the cortical surface. In our case, the signals from the above-mentioned array are analyzed and information concerning orientation and size of a particular line is obtained. With this system, the Muelle-Lyer illusion has been studied and some rules to interpret why equal length objects give rise to different interpretations are presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A method to reduce the bruise susceptibility of apples by controlling the moisture loss of the fruit was evaluated. Previous research indicates that reduction of the relative humidity of the storage air leads to an immediate effect on the weight loss and on skin properties and to a lower bruise susceptibility of apples. The diffusion equation is used to determine the waterpotential profile inside the fruit during storage. Characteristics of the waterpotential distribution in the fruit are related to measured bruise volumes. The results indicate how /this model can be used to control bruise susceptibility.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of this work is to present the Exercise I-1b “pin-cell burn-up benchmark” proposed in the framework of OECD LWR UAM. Its objective is to address the uncertainty due to the basic nuclear data as well as the impact of processing the nuclear and covariance data in a pin-cell depletion calculation. Four different sensitivity/uncertainty propagation methodologies participate in this benchmark (GRS, NRG, UPM, and SNU&KAERI). The paper describes the main features of the UPM model (hybrid method) compared with other methodologies. The requested output provided by UPM is presented, and it is discussed regarding the results of other methodologies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the last years, many analyses from acoustic signal processing have been used for different applications. In most cases, these sensor systems are based on the determination of times of flight for signals from every transducer. This paper presents a flat plate generalization method for impact detection and location over linear links or bars-based structures. The use of three piezoelectric sensors allow to achieve the position and impact time while the use of additional sensors lets cover a larger area of detection and avoid wrong timing difference measurements. An experimental setup and some experimental results are briefly presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

En este proyecto se han analizado distintas imágenes de fragmentos de rocas de distintas granulometrías correspondientes a una serie de voladuras de una misma cantera. Cada una de las voladuras se componen de 20 imágenes. A posteriori utilizando el programa Split Desktop en su versión 3.1, se delimitaron los fragmentos de roca de los que está compuesta la imagen, obteniéndose posteriormente la curva granulométrica correspondiente a dicha imagen. Una vez se calculan las curvas granulométricas correspondientes a cada imagen, se calcula la curva media de todas ellas, pudiéndose considerar por tanto la curva media de cada voladura. Se han utilizado las distintas soluciones del software, manual, online y automático, para realizar los análisis de dichas imágenes y a posteriori comparar sus resultados. Dichos resultados se muestran a través de una serie de gráficos y tablas que se explican con detalle para la comprensión del estudio. De dichos resultados es posible afirmar que, el tratamiento de imágenes realizado de manera online y automático por Split, desemboca en el mismo resultado, al no haber una diferencia estadística significativa. Por el contrario, el sistema manual es diferente de los otros dos, no pudiéndose afirmar cual es mejor de los dos. El manual depende del operario que trabaje las imágenes y el online de los ajustes realizados y por tanto, ambos tienen ciertas incertidumbres difíciles de solucionar. Abstract In this project, different images of rock fragments of different grain sizes corresponding to a series of blasts from the same quarry have been analyzed. To study each blast, 20 images has been used and studied with the software Split Desktop 3.1. Rock fragments from each image has been delimitated with the software, obtaining a grading curve of each one. Once these curves are calculated, the mean curve of these data set is obtained and can be considered the mean curve of each blast. Different software solutions as manual, online and automatic, has been used for the analysis of these images. Then the results has been compared between them. These results are shown through a series of graphs and tables, that are explained in detail, to enhance the understanding of the study. From these results, it can be said that the image processing with online and automatic options from Split, leads to the same result, after an statistical study. On the contrary, the manual Split mode is different from the others; however is not possible to assert what will be the best. The manual Split mode depends on the operator ability and dedication, although the online mode depends on the software settings, so therefore, both have some uncertainties that are difficult to solve.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Directionally solidified Al2O3–Er3Al5O12–ZrO2 eutectic rods were processed using the laser floating zone method at growth rates of 25, 350and 750 mm/h to obtain microstructures with different domain size. The mechanical properties were investigated as a function of the processing rate. The hardness, 15.6 GPa, and the fracture toughness, 4 MPa m1/2, obtained from Vickers indentation at room temperature were practically independent of the size of the eutectic phases. However, the flexural strength increased as the domain size decreased, reaching outstanding strength values close to 3 GPa in the samples grown at 750 mm/h. A high retention of the flexural strength was observed up to 1500 K in the materials processed at 25 and 350 mm/h, while superplastic behaviour was observed at 1700 K in the eutectic rods solidified at the highest rate of 750 mm/h

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the last decade, Object Based Image Analysis (OBIA) has been accepted as an effective method for processing high spatial resolution multiband images. This image analysis method is an approach that starts with the segmentation of the image. Image segmentation in general is a procedure to partition an image into homogenous groups (segments). In practice, visual interpretation is often used to assess the quality of segmentation and the analysis relies on the experience of an analyst. In an effort to address the issue, in this study, we evaluate several seed selection strategies for an automatic image segmentation methodology based on a seeded region growing-merging approach. In order to evaluate the segmentation quality, segments were subjected to spatial autocorrelation analysis using Moran's I index and intra-segment variance analysis. We apply the algorithm to image segmentation using an aerial multiband image.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A new technology is being proposed as a solution to the problem of unintentional facial detection and recognition in pictures in which the individuals appearing want to express their privacy preferences, through the use of different tags. The existing methods for face de-identification were mostly ad hoc solutions that only provided an absolute binary solution in a privacy context such as pixelation, or a bar mask. As the number and users of social networks are increasing, our preferences regarding our privacy may become more complex, leaving these absolute binary solutions as something obsolete. The proposed technology overcomes this problem by embedding information in a tag which will be placed close to the face without being disruptive. Through a decoding method the tag will provide the preferences that will be applied to the images in further stages.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents an adaptation of the Cross-Entropy (CE) method to optimize fuzzy logic controllers. The CE is a recently developed optimization method based on a general Monte-Carlo approach to combinatorial and continuous multi-extremal optimization and importance sampling. This work shows the application of this optimization method to optimize the inputs gains, the location and size of the different membership functions' sets of each variable, as well as the weight of each rule from the rule's base of a fuzzy logic controller (FLC). The control system approach presented in this work was designed to command the orientation of an unmanned aerial vehicle (UAV) to modify its trajectory for avoiding collisions. An onboard looking forward camera was used to sense the environment of the UAV. The information extracted by the image processing algorithm is the only input of the fuzzy control approach to avoid the collision with a predefined object. Real tests with a quadrotor have been done to corroborate the improved behavior of the optimized controllers at different stages of the optimization process.