934 resultados para Compressed Objects
Resumo:
In this work, we consider a simple model problem for the electromagnetic exploration of small perfectly conducting objects buried within the lower halfspace of an unbounded two–layered background medium. In possible applications, such as, e.g., humanitarian demining, the two layers would correspond to air and soil. Moving a set of electric devices parallel to the surface of ground to generate a time–harmonic field, the induced field is measured within the same devices. The goal is to retrieve information about buried scatterers from these data. In mathematical terms, we are concerned with the analysis and numerical solution of the inverse scattering problem to reconstruct the number and the positions of a collection of finitely many small perfectly conducting scatterers buried within the lower halfspace of an unbounded two–layered background medium from near field measurements of time–harmonic electromagnetic waves. For this purpose, we first study the corresponding direct scattering problem in detail and derive an asymptotic expansion of the scattered field as the size of the scatterers tends to zero. Then, we use this expansion to justify a noniterative MUSIC–type reconstruction method for the solution of the inverse scattering problem. We propose a numerical implementation of this reconstruction method and provide a series of numerical experiments.
Resumo:
In computer systems, specifically in multithread, parallel and distributed systems, a deadlock is both a very subtle problem - because difficult to pre- vent during the system coding - and a very dangerous one: a deadlocked system is easily completely stuck, with consequences ranging from simple annoyances to life-threatening circumstances, being also in between the not negligible scenario of economical losses. Then, how to avoid this problem? A lot of possible solutions has been studied, proposed and implemented. In this thesis we focus on detection of deadlocks with a static program analysis technique, i.e. an analysis per- formed without actually executing the program. To begin, we briefly present the static Deadlock Analysis Model devel- oped for coreABS−− in chapter 1, then we proceed by detailing the Class- based coreABS−− language in chapter 2. Then, in Chapter 3 we lay the foundation for further discussions by ana- lyzing the differences between coreABS−− and ASP, an untyped Object-based calculi, so as to show how it can be possible to extend the Deadlock Analysis to Object-based languages in general. In this regard, we explicit some hypotheses in chapter 4 first by present- ing a possible, unproven type system for ASP, modeled after the Deadlock Analysis Model developed for coreABS−−. Then, we conclude our discussion by presenting a simpler hypothesis, which may allow to circumvent the difficulties that arises from the definition of the ”ad-hoc” type system discussed in the aforegoing chapter.
Resumo:
We consider a simple (but fully three-dimensional) mathematical model for the electromagnetic exploration of buried, perfect electrically conducting objects within the soil underground. Moving an electric device parallel to the ground at constant height in order to generate a magnetic field, we measure the induced magnetic field within the device, and factor the underlying mathematics into a product of three operations which correspond to the primary excitation, some kind of reflection on the surface of the buried object(s) and the corresponding secondary excitation, respectively. Using this factorization we are able to give a justification of the so-called sampling method from inverse scattering theory for this particular set-up.
Resumo:
n the last few years, the vision of our connected and intelligent information society has evolved to embrace novel technological and research trends. The diffusion of ubiquitous mobile connectivity and advanced handheld portable devices, amplified the importance of the Internet as the communication backbone for the fruition of services and data. The diffusion of mobile and pervasive computing devices, featuring advanced sensing technologies and processing capabilities, triggered the adoption of innovative interaction paradigms: touch responsive surfaces, tangible interfaces and gesture or voice recognition are finally entering our homes and workplaces. We are experiencing the proliferation of smart objects and sensor networks, embedded in our daily living and interconnected through the Internet. This ubiquitous network of always available interconnected devices is enabling new applications and services, ranging from enhancements to home and office environments, to remote healthcare assistance and the birth of a smart environment. This work will present some evolutions in the hardware and software development of embedded systems and sensor networks. Different hardware solutions will be introduced, ranging from smart objects for interaction to advanced inertial sensor nodes for motion tracking, focusing on system-level design. They will be accompanied by the study of innovative data processing algorithms developed and optimized to run on-board of the embedded devices. Gesture recognition, orientation estimation and data reconstruction techniques for sensor networks will be introduced and implemented, with the goal to maximize the tradeoff between performance and energy efficiency. Experimental results will provide an evaluation of the accuracy of the presented methods and validate the efficiency of the proposed embedded systems.
Resumo:
Introduzione al metodo del Compressed Sensing per il campionamento di segnali sparsi.
Resumo:
The idea of balancing the resources spent in the acquisition and encoding of natural signals strictly to their intrinsic information content has interested nearly a decade of research under the name of compressed sensing. In this doctoral dissertation we develop some extensions and improvements upon this technique's foundations, by modifying the random sensing matrices on which the signals of interest are projected to achieve different objectives. Firstly, we propose two methods for the adaptation of sensing matrix ensembles to the second-order moments of natural signals. These techniques leverage the maximisation of different proxies for the quantity of information acquired by compressed sensing, and are efficiently applied in the encoding of electrocardiographic tracks with minimum-complexity digital hardware. Secondly, we focus on the possibility of using compressed sensing as a method to provide a partial, yet cryptanalysis-resistant form of encryption; in this context, we show how a random matrix generation strategy with a controlled amount of perturbations can be used to distinguish between multiple user classes with different quality of access to the encrypted information content. Finally, we explore the application of compressed sensing in the design of a multispectral imager, by implementing an optical scheme that entails a coded aperture array and Fabry-Pérot spectral filters. The signal recoveries obtained by processing real-world measurements show promising results, that leave room for an improvement of the sensing matrix calibration problem in the devised imager.
Resumo:
Tecniche per l'acquisizione a basso consumo di segnali sparsi tramite compressed sensing
Resumo:
Il compressed sensing è un’innovativa tecnica per l’acquisizione dei dati, che mira all'estrazione del solo contenuto informativo intrinseco di un segnale. Ciò si traduce nella possibilità di acquisire informazione direttamente in forma compressa, riducendo la quantità di risorse richieste per tale operazione. In questa tesi è sviluppata un'architettura hardware per l'acquisizione di segnali analogici basata sul compressed sensing, specializzata al campionamento con consumo di potenza ridotto di segnali biomedicali a basse frequenze. Lo studio è svolto a livello di sistema mediante l'integrazione della modulazione richiesta dal compressed sensing in un convertitore analogico-digitale ad approssimazioni successive, modificandone la logica di controllo. Le prestazioni risultanti sono misurate tramite simulazioni numeriche e circuitali. Queste confermano la possibilità di ridurre la complessità hardware del sistema di acquisizione rispetto allo stato dell'arte, senza alterarne le prestazioni.
Resumo:
In this thesis the results of the multifrequency VLBA observations of the GPS 1944+5448 and the HFP J0111+3906 are presented. They are compact objects smaller than about 100 pc, completely embedded in the host galaxy. The availability of multi-epoch VLBI observations spanning more than 10 years, allowed us to compute the hot spot advance speed in order to obtain the kinematic age of both sources. Both radio sources are young, in agreement with the idea that they are in an early evolutionary stage. The spectral analysis of each source component, such as the lobes, the hot spots, the core and the jets, making a comparison with the theoretical ones is described. In addition the physical parameters derived from VLBA images as the magnetic field, the luminosity, the energy and the ambient medium density of both sources are discussed.
Resumo:
In this work we study a model for the breast image reconstruction in Digital Tomosynthesis, that is a non-invasive and non-destructive method for the three-dimensional visualization of the inner structures of an object, in which the data acquisition includes measuring a limited number of low-dose two-dimensional projections of an object by moving a detector and an X-ray tube around the object within a limited angular range. The problem of reconstructing 3D images from the projections provided in the Digital Tomosynthesis is an ill-posed inverse problem, that leads to a minimization problem with an object function that contains a data fitting term and a regularization term. The contribution of this thesis is to use the techniques of the compressed sensing, in particular replacing the standard least squares problem of data fitting with the problem of minimizing the 1-norm of the residuals, and using as regularization term the Total Variation (TV). We tested two different algorithms: a new alternating minimization algorithm (ADM), and a version of the more standard scaled projected gradient algorithm (SGP) that involves the 1-norm. We perform some experiments and analyse the performance of the two methods comparing relative errors, iterations number, times and the qualities of the reconstructed images. In conclusion we noticed that the use of the 1-norm and the Total Variation are valid tools in the formulation of the minimization problem for the image reconstruction resulting from Digital Tomosynthesis and the new algorithm ADM has reached a relative error comparable to a version of the classic algorithm SGP and proved best in speed and in the early appearance of the structures representing the masses.
Resumo:
The aims of this study were to determine the effects of pH and acid concentration on the dissolution of enamel, dentine, and compressed hydroxyapatite (HA) in citric acid solutions (15.6 and 52.1 mmol l(-1) ; pH 2.45, 3.2, and 3.9), using a pH-stat system. After an initial adjustment period, the dissolution rates of enamel and HA were constant, while that of dentine decreased with time. The dissolution rate increased as the pH decreased, and this was most marked for enamel. To compare substrates, the rate of mineral dissolution was normalized to the area occupied by mineral at the specimen surface. For a given acid concentration, the normalized dissolution rate of HA was always less than that for either dentine or enamel. The dissolution rate for dentine mineral was similar to that for enamel at pH 2.45 and greater at pH 3.2 and pH 3.9. The concentration of acid significantly affected the enamel dissolution rate at pH 2.45 and pH 3.2, but not at pH 3.9, and did not significantly affect the dissolution rates of dentine or HA at any pH. The variation in response of the dissolution rate to acid concentration/buffer capacity with respect to pH and tissue type might complicate attempts to predict erosive potential from solution composition.
Resumo:
To study the effect of a nonlinear noise filter on the detection of simulated endoleaks in a phantom with 80- and 100-kVp multidetector computed tomographic (CT) angiography.