29 resultados para Iterative decoding
Resumo:
Introduction: The purpose of this review is to gather and analyse current research publications to evaluate Sinogram-Affirmed Iterative Reconstruction (SAFIRE). The aim of this review is to investigate whether this algorithm is capable of reducing the dose delivered during CT imaging while maintaining image quality. Recent research shows that children have a greater risk per unit dose due to increased radiosensitivity and longer life expectancies, which means it is particularly important to reduce the radiation dose received by children. Discussion: Recent publications suggest that SAFIRE is capable of reducing image noise in CT images, thereby enabling the potential to reduce dose. Some publications suggest a decrease in dose, by up to 64% compared to filtered back projection, can be accomplished without a change in image quality. However, literature suggests that using a higher SAFIRE strength may alter the image texture, creating an overly ‘smoothed’ image that lacks contrast. Some literature reports SAFIRE gives decreased low contrast detectability as well as spatial resolution. Publications tend to agree that SAFIRE strength three is optimal for an acceptable level of visual image quality, but more research is required. The importance of creating a balance between dose reduction and image quality is stressed. In this literature review most of the publications were completed using adults or phantoms, and a distinct lack of literature for paediatric patients is noted. Conclusion: It is necessary to find an optimal way to balance dose reduction and image quality. More research relating to SAFIRE and paediatric patients is required to fully investigate dose reduction potential in this population, for a range of different SAFIRE strengths.
Resumo:
Finding the structure of a confined liquid crystal is a difficult task since both the density and order parameter profiles are nonuniform. Starting from a microscopic model and density-functional theory, one has to either (i) solve a nonlinear, integral Euler-Lagrange equation, or (ii) perform a direct multidimensional free energy minimization. The traditional implementations of both approaches are computationally expensive and plagued with convergence problems. Here, as an alternative, we introduce an unsupervised variant of the multilayer perceptron (MLP) artificial neural network for minimizing the free energy of a fluid of hard nonspherical particles confined between planar substrates of variable penetrability. We then test our algorithm by comparing its results for the structure (density-orientation profiles) and equilibrium free energy with those obtained by standard iterative solution of the Euler-Lagrange equations and with Monte Carlo simulation results. Very good agreement is found and the MLP method proves competitively fast, flexible, and refinable. Furthermore, it can be readily generalized to the richer experimental patterned-substrate geometries that are now experimentally realizable but very problematic to conventional theoretical treatments.
Resumo:
We prove existence, uniqueness, and stability of solutions of the prescribed curvature problem (u'/root 1 + u'(2))' = au - b/root 1 + u'(2) in [0, 1], u'(0) = u(1) = 0, for any given a > 0 and b > 0. We also develop a linear monotone iterative scheme for approximating the solution. This equation has been proposed as a model of the corneal shape in the recent paper (Okrasinski and Plociniczak in Nonlinear Anal., Real World Appl. 13:1498-1505, 2012), where a simplified version obtained by partial linearization has been investigated.
Resumo:
Objectives: Children have a greater risk from radiation, per unit dose, due to increased radiosensitivity and longer life expectancies. It is of paramount importance to reduce the radiation dose received by children. This research concerns chest CT examinations on paediatric patients. The purpose of this study was to compare the image quality and the dose received from imaging with images reconstructed with filtered back projection (FBP) and five strengths of Sinogram-Affirmed Iterative Reconstruction (SAFIRE). Methods: Using a multi-slice CT scanner, six series of images were taken of a paediatric phantom. Two kVp values (80 and 110), 3 mAs values (25, 50 and 100) and 2 slice thicknesses (1 mm and 3 mm) were used. All images were reconstructed with FBP and five strengths of SAFIRE. Ten observers evaluated visual image quality. Dose was measured using CT-Expo. Results: FBP required a higher dose than all SAFIRE strengths to obtain the same image quality for sharpness and noise. For sharpness and contrast image quality ratings of 4, FBP required doses of 6.4 and 6.8 mSv respectively. SAFIRE 5 required doses of 3.4 and 4.3 mSv respectively. Clinical acceptance rate was improved by the higher voltage (110 kV) for all images in comparison to 80 kV, which required a higher dose for acceptable image quality. 3 mm images were typically better quality than 1 mm images. Conclusion: SAFIRE 5 was optimal for dose reduction and image quality.
Resumo:
Projeto de Intervenção apresentado à Escola Superior de Educação de Lisboa para obtenção do grau de Mestre em Educação Especial, Especialização em Problemas de Aprendizagem e Comportamento
Resumo:
Nas últimas décadas, tem se observado uma crescente demanda pelo domínio da linguagem escrita em todas as áreas da vida social (Olson 2009). Essa demanda não se limita apenas ao contexto brasileiro, mas refere-se a um contexto mundial, que hoje coloca o domínio das diversas capacidades de linguagem, em especial das capacidades de leitura e escrita, como condição para acesso ao conhecimento, à participação social e o exercício efetivo da cidadania. O domínio de tais capacidades refere-se a um certo tipo ou nível de literacia que vai além da descodificação, diz respeito às diversas capacidades de leitura e escrita necessárias em diferentes práticas sociais. Partindo de uma parceria de investigação entre instituições de ensino superior em Portugal e no Brasil, temos vindo a reflectir sobre questões de currículo nos dois países.
Resumo:
The acquisition of a Myocardial Perfusion image (MPI) is of great importance for the diagnosis of the coronary artery disease, since it allows to evaluate which areas of the heart aren’t being properly perfused, in rest and stress situations. This exam is greatly influenced by photon attenuation which creates image artifacts and affects quantification. The acquisition of a Computerized Tomography (CT) image makes it possible to get an atomic images which can be used to perform high-quality attenuation corrections of the radiopharmaceutical distribution, in the MPI image. Studies show that by using hybrid imaging to perform diagnosis of the coronary artery disease, there is an increase on the specificity when evaluating the perfusion of the right coronary artery (RCA). Using an iterative algorithm with a resolution recovery software for the reconstruction, which balances the image quality, the administered activity and the scanning time, we aim to evaluate the influence of attenuation correction on the MPI image and the outcome in perfusion quantification and imaging quality.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Electrónica e telecomunicações
Resumo:
Dissertação apresentada à Escola Superior de Comunicação Social como parte dos requisitos para obtenção de grau de mestre em Audiovisual e Multimédia.
Resumo:
The iterative simulation of the Brownian bridge is well known. In this article, we present a vectorial simulation alternative based on Gaussian processes for machine learning regression that is suitable for interpreted programming languages implementations. We extend the vectorial simulation of path-dependent trajectories to other Gaussian processes, namely, sequences of Brownian bridges, geometric Brownian motion, fractional Brownian motion, and Ornstein-Ulenbeck mean reversion process.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Mecânica
Resumo:
ABSTRACT - Starting with the explanation of metanarrative as a sort of self-reflexive storytelling (as defended by Kenneth Weaver Hope in his unpublished PhD. thesis), I propose to talk about enunciative practices that stress the telling more than the told. In line with some metaficcional practices applied to cinema, such as the ‘mindfuck’ film (Jonathan Eig, 2003), the ‘psychological puzzle film’ (Elliot Panek, 2003) and the ‘mind-game film’ (Thomas Elsaesser, 2009), I will address the manipulations that a narrative film endures in order to produce a more fruitful and complex experience for the viewer. I will particularly concentrate on the misrepresentation of time as a way to produce a labyrinthine work of fiction where the linear description of events is replaced by a game of time disclosure. The viewer is thus called upon to reconstruct the order of the various situations portrayed in a process that I call ‘temporal mapping’. However, as the viewer attempts to do this, the film, ironically, because of the intricate nature of the plot and the uncertain status of the characters, resists the attempt. There is a sort of teasing taking place between the film and its spectator: an invitation of decoding that is half-denied until the end, where the puzzle is finally solved. I will use three of Alejandro Iñárritu’s films to better convey my point: Amores perros (2000), 21 Grams (2003) and Babel (2006). I will consider Iñárritu’s methods to produce a non-linear storytelling as a way to stress the importance of time and its validity as one of the elements that make up for a metanarrative experience in films. I will focus especially on 21 Grams, which I consider to be a paragon of the labyrinth.
Resumo:
Partial dynamic reconfiguration of FPGAs can be used to implement complex applications using the concept of virtual hardware. In this work we have used partial dynamic reconfiguration to implement a JPEG decoder with reduced area. The image decoding process was adapted to be implemented on the FPGA fabric using this technique. The architecture was tested in a low cost ZYNQ-7020 FPGA that supports dynamic reconfiguration. The results show that the proposed solution needs only 40% of the resources utilized by a static implementation. The performance of the dynamic solution is about 9X slower than the static solution by trading-off internal resources of the FPGA. A throughput of 7 images per second is achievable with the proposed partial dynamic reconfiguration solution.
Resumo:
The application of compressive sensing (CS) to hyperspectral images is an active area of research over the past few years, both in terms of the hardware and the signal processing algorithms. However, CS algorithms can be computationally very expensive due to the extremely large volumes of data collected by imaging spectrometers, a fact that compromises their use in applications under real-time constraints. This paper proposes four efficient implementations of hyperspectral coded aperture (HYCA) for CS, two of them termed P-HYCA and P-HYCA-FAST and two additional implementations for its constrained version (CHYCA), termed P-CHYCA and P-CHYCA-FAST on commodity graphics processing units (GPUs). HYCA algorithm exploits the high correlation existing among the spectral bands of the hyperspectral data sets and the generally low number of endmembers needed to explain the data, which largely reduces the number of measurements necessary to correctly reconstruct the original data. The proposed P-HYCA and P-CHYCA implementations have been developed using the compute unified device architecture (CUDA) and the cuFFT library. Moreover, this library has been replaced by a fast iterative method in the P-HYCA-FAST and P-CHYCA-FAST implementations that leads to very significant speedup factors in order to achieve real-time requirements. The proposed algorithms are evaluated not only in terms of reconstruction error for different compressions ratios but also in terms of computational performance using two different GPU architectures by NVIDIA: 1) GeForce GTX 590; and 2) GeForce GTX TITAN. Experiments are conducted using both simulated and real data revealing considerable acceleration factors and obtaining good results in the task of compressing remotely sensed hyperspectral data sets.