996 resultados para Image sensor


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Purpose: This study aims to investigate the influence of tube potential (kVp) variation in relation to perceptual image quality and effective dose for pelvis using automatic exposure control (AEC) and non-AEC in a computed radiography (CR) system. Methods and Materials: To determine the effects of using AEC and non-AEC by applying the 10 kVp rule in two experiments using an anthropomorphic pelvis phantom. Images were acquired using 10 kVp increments (60-120 kVp) for both experiments. The first experiment, based on seven AEC combinations, produced 49 images. The mean mAs from each kVp increment were used as a baseline for the second experiment producing 35 images. A total of 84 images were produced and a panel of 5 experienced observers participated for the image scoring using the 2 AFC visual grading software. PCXMC software was used to estimate the effective dose. Results: A decrease in perceptual image quality as the kVp increases was observed both in non-AEC and AEC experiments, however no significant statistical differences (p> 0.05) were found. Image quality scores from all observers at 10 kVp increments for all mAs values using non-AEC mode demonstrates a better score up to 90 kVp. Effective dose results show a statistical significant decrease (p=0.000) on the 75th quartile from 0.3 mSv at 60 kVp to 0.1 mSv at 120 kVp when applying the 10 kVp rule in non-AEC mode. Conclusion: No significant reduction in perceptual image quality is observed when increasing kVp whilst a marked and significant effective dose reduction is observed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In a time of fierce competition between regions, an image serve as a basis to develop a strong sense of community, which fosters trust and cooperation that can be mobilized for regional growth. A positive image and reputation could be used in the promotional activities of the region benefiting all the stakeholders as a whole. Mega cultural events are frequently used to attract tourists and investments to a region, but also to enhance the city’s image. This study adopts a marketing/communication perspective of city’s image, and intends to explain how the image of the city is perceived by their residents. Specifically, we intend to compare the perceptions of residents that effectively participated in the Guimarães European Capital of Culture (ECOC) 2012 (engaged residents), and the residents that only assisted to the event (attendees). Several significant findings are reported and their implications for event managers and public policy administrators presented, along with the limitations of the study

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Measurements in civil engineering load tests usually require considerable time and complex procedures. Therefore, measurements are usually constrained by the number of sensors resulting in a restricted monitored area. Image processing analysis is an alternative way that enables the measurement of the complete area of interest with a simple and effective setup. In this article photo sequences taken during load displacement tests were captured by a digital camera and processed with image correlation algorithms. Three different image processing algorithms were used with real images taken from tests using specimens of PVC and Plexiglas. The data obtained from the image processing algorithms were also compared with the data from physical sensors. A complete displacement and strain map were obtained. Results show that the accuracy of the measurements obtained by photogrammetry is equivalent to that from the physical sensors but with much less equipment and fewer setup requirements. © 2015Computer-Aided Civil and Infrastructure Engineering.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do grau de Mestre em Engenharia Electrotécnica e de Computadores

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nowadays the incredible grow of mobile devices market led to the need for location-aware applications. However, sometimes person location is difficult to obtain, since most of these devices only have a GPS (Global Positioning System) chip to retrieve location. In order to suppress this limitation and to provide location everywhere (even where a structured environment doesn’t exist) a wearable inertial navigation system is proposed, which is a convenient way to track people in situations where other localization systems fail. The system combines pedestrian dead reckoning with GPS, using widely available, low-cost and low-power hardware components. The system innovation is the information fusion and the use of probabilistic methods to learn persons gait behavior to correct, in real-time, the drift errors given by the sensors.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nowadays there is an increase of location-aware mobile applications. However, these applications only retrieve location with a mobile device's GPS chip. This means that in indoor or in more dense environments these applications don't work properly. To provide location information everywhere a pedestrian Inertial Navigation System (INS) is typically used, but these systems can have a large estimation error since, in order to turn the system wearable, they use low-cost and low-power sensors. In this work a pedestrian INS is proposed, where force sensors were included to combine with the accelerometer data in order to have a better detection of the stance phase of the human gait cycle, which leads to improvements in location estimation. Besides sensor fusion an information fusion architecture is proposed, based on the information from GPS and several inertial units placed on the pedestrian body, that will be used to learn the pedestrian gait behavior to correct, in real-time, the inertial sensors errors, thus improving location estimation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hyperspectral imaging has become one of the main topics in remote sensing applications, which comprise hundreds of spectral bands at different (almost contiguous) wavelength channels over the same area generating large data volumes comprising several GBs per flight. This high spectral resolution can be used for object detection and for discriminate between different objects based on their spectral characteristics. One of the main problems involved in hyperspectral analysis is the presence of mixed pixels, which arise when the spacial resolution of the sensor is not able to separate spectrally distinct materials. Spectral unmixing is one of the most important task for hyperspectral data exploitation. However, the unmixing algorithms can be computationally very expensive, and even high power consuming, which compromises the use in applications under on-board constraints. In recent years, graphics processing units (GPUs) have evolved into highly parallel and programmable systems. Specifically, several hyperspectral imaging algorithms have shown to be able to benefit from this hardware taking advantage of the extremely high floating-point processing performance, compact size, huge memory bandwidth, and relatively low cost of these units, which make them appealing for onboard data processing. In this paper, we propose a parallel implementation of an augmented Lagragian based method for unsupervised hyperspectral linear unmixing on GPUs using CUDA. The method called simplex identification via split augmented Lagrangian (SISAL) aims to identify the endmembers of a scene, i.e., is able to unmix hyperspectral data sets in which the pure pixel assumption is violated. The efficient implementation of SISAL method presented in this work exploits the GPU architecture at low level, using shared memory and coalesced accesses to memory.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hyperspectral imaging sensors provide image data containing both spectral and spatial information from the Earth surface. The huge data volumes produced by these sensors put stringent requirements on communications, storage, and processing. This paper presents a method, termed hyperspectral signal subspace identification by minimum error (HySime), that infer the signal subspace and determines its dimensionality without any prior knowledge. The identification of this subspace enables a correct dimensionality reduction yielding gains in algorithm performance and complexity and in data storage. HySime method is unsupervised and fully-automatic, i.e., it does not depend on any tuning parameters. The effectiveness of the proposed method is illustrated using simulated data based on U.S.G.S. laboratory spectra and real hyperspectral data collected by the AVIRIS sensor over Cuprite, Nevada.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dissertation presented to obtain the degree of Doctor of Philosophy in Electrical Engineering, speciality on Perceptional Systems, by the Universidade Nova de Lisboa, Faculty of Sciences and Technology

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Given an hyperspectral image, the determination of the number of endmembers and the subspace where they live without any prior knowledge is crucial to the success of hyperspectral image analysis. This paper introduces a new minimum mean squared error based approach to infer the signal subspace in hyperspectral imagery. The method, termed hyperspectral signal identification by minimum error (HySime), is eigendecomposition based and it does not depend on any tuning parameters. It first estimates the signal and noise correlation matrices and then selects the subset of eigenvalues that best represents the signal subspace in the least squared error sense. The effectiveness of the proposed method is illustrated using simulated data based on U.S.G.S. laboratory spectra and real hyperspectral data collected by the AVIRIS sensor over Cuprite, Nevada.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper introduces a new toolbox for hyperspectral imagery, developed under the MATLAB environment. This toolbox provides easy access to different supervised and unsupervised classification methods. This new application is also versatile and fully dynamic since the user can embody their own methods, that can be reused and shared. This toolbox, while extends the potentiality of MATLAB environment, it also provides a user-friendly platform to assess the results of different methodologies. In this paper it is also presented, under the new application, a study of several different supervised and unsupervised classification methods on real hyperspectral data.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper introduces a new method to blindly unmix hyperspectral data, termed dependent component analysis (DECA). This method decomposes a hyperspectral images into a collection of reflectance (or radiance) spectra of the materials present in the scene (endmember signatures) and the corresponding abundance fractions at each pixel. DECA assumes that each pixel is a linear mixture of the endmembers signatures weighted by the correspondent abundance fractions. These abudances are modeled as mixtures of Dirichlet densities, thus enforcing the constraints on abundance fractions imposed by the acquisition process, namely non-negativity and constant sum. The mixing matrix is inferred by a generalized expectation-maximization (GEM) type algorithm. This method overcomes the limitations of unmixing methods based on Independent Component Analysis (ICA) and on geometrical based approaches. The effectiveness of the proposed method is illustrated using simulated data based on U.S.G.S. laboratory spectra and real hyperspectral data collected by the AVIRIS sensor over Cuprite, Nevada.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A Norfloxacina (NFX) é um antibiótico antibacteriano indicado para combater bactérias Gram-negativas e amplamente utilizado para o tratamento de infeções no trato respiratório e urinário. Com a necessidade de realizar estudos clínicos e farmacológicos esenvolveram-se métodos de análise rápida e sensitiva para a determinação da Norfloxacina. Neste trabalho foi desenvolvido um novo sensor eletroquímico sensível e seletivo para a deteção da NFX. O sensor foi construído a partir de modificações efetuadas num elétrodo de carbono vítreo. Inicialmente o elétrodo foi modificado com a deposição de uma suspensão de nanotubos de carbono de paredes múltiplas (MWCNT) de modo a aumentar a sensibilidade de resposta analítica. De seguida um filme polímerico molecularmente impresso (MIP) foi preparado por eletrodeposição, a partir de uma solução contendo pirrol (monómero funcional) e NFX (template). Um elétrodo de controlo não impresso foi também preparado (NIP). Estudouse e caraterizou-se a resposta eletroquímica do sensor para a oxidação da NFX por voltametria de onda quadrada. Foram optimizados diversos parâmetros experimentais, tais como, condições ótimas de polimerização, condições de incubação e condições de extração. O sensor apresenta um comportamento linear entre a intensidade da corrente do pico e o logaritmo da concentração de NFX na gama entre 0,1 e 8μM. Os resultados obtidos apresentam boa precisão, com repetibilidade inferior a 6% e reprodutibilidade inferior a 9%. Foi calculado a partir da curva de calibração um limite de deteção de 0,2 μM O método desenvolvido é seletivo, rápido e de fácil manuseamento. O sensor molecularmente impresso foi aplicado com sucesso na deteção da NFX em amostras de urina real e água.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The visual image is a fundamental component of epiphany, stressing its immediacy and vividness, corresponding to the enargeia of the traditional ekphrasis and also playing with cultural and social meanings. Morris Beja in his seminal book Epiphany in the Modern Novel, draws our attention to the distinction made by Joyce between the epiphany originated in a common object, in a discourse or gesture and the one arising in “a memorable phase of the mind itself”. This type materializes in the “dream-epiphany” and in the epiphany based in memory. On the other hand, Robert Langbaum in his study of the epiphanic mode, suggests that the category of “visionary epiphany” could account for the modern effect of an internally glowing vision like Blake’s “The Tyger”, which projects the vitality of a real tyger. The short story, whose length renders it a fitting genre for the use of different types of epiphany, has dealt with the impact of the visual image in this technique, to convey different effects and different aesthetic aims. This paper will present some examples of this occurrence in short stories of authors in whose work epiphany is a fundamental concept and literary technique: Walter Pater, Joseph Conrad, K. Mansfield, Clarice Lispector. Pater’s “imaginary portraits” concentrate on “priviledged moments” of the lives of the characters depicting their impressions through pictorial language; Conrad tries to show “moments of awakening” that can be remembered by the eye; Mansfield suggests that epiphany, the “glimpse”, should replace plot as an internal ordering principle of her impressionist short-stories; in C. Lispector the visualization of some situations is so aggressive that it causes nausea and a radical revelation on the protagonist’s.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dissertação apresentada para obtenção do grau de Mestre em Bioquímica Estrutural e Funcional, pela Universidade Nova de Lisboa, Faculdade de Ciências e Tecnologia