968 resultados para Blurred and noisy images


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Image collections are growing at a rapid rate and hence visual information is becoming more and more important. Clearly, these image repositories need to be managed, and tools for effectively and efficiently searching image databases are highly sought after, especially on mobile devices where more and more images are being stored. In this paper, we present an image browsing system for interactive exploration of image collections on mobile devices. Images are arranged so that visually similar images are grouped together while large image repositories become accessible through a hierarchical, browsable tree structure, arranged on a hexagonal lattice. The developed system provides an intuitive and fast interface for navigating through image databases using a variety of touch gestures.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This book is an urgent and compelling account of the Occupy movements: from the M15 movement in Spain, to the wave of Occupations flooding across cities in American, Europe and Australia, to the harsh reality of evictions as corporations and governments attempted to reassert exclusive control over public space. Across a vast range of international examples over twenty authors analyse, explain and helps us understand the movement. These movements were a novel and noisy intervention into the recent capitalist crisis in developed economies, developing an exceptionally broad identity through a call to arms addressed to 'the 99%', and emphasizing the importance of public space in the creation and maintenance of opposition. The novelties of these movements, along with their radical positioning and the urgency of their claims all demand analysis. This book investigates the crucial questions of how and why this form of action spread so rapidly and so widely, how the inclusive discourse of 'the 99%' matched up to the reality of the practice. It is vital to understand not just the choice of tactics and the vitality of protest camps in public spaces, but also how the myriad of challenges and problems were negotiated. This book was published as a special issue of Social Movement Studies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Mathematical Morphology presents a systematic approach to extract geometric features of binary images, using morphological operators that transform the original image into another by means of a third image called structuring element and came out in 1960 by researchers Jean Serra and George Matheron. Fuzzy mathematical morphology extends the operators towards grayscale and color images and was initially proposed by Goetherian using fuzzy logic. Using this approach it is possible to make a study of fuzzy connectives, which allows some scope for analysis for the construction of morphological operators and their applicability in image processing. In this paper, we propose the development of morphological operators fuzzy using the R-implications for aid and improve image processing, and then to build a system with these operators to count the spores mycorrhizal fungi and red blood cells. It was used as the hypothetical-deductive methodologies for the part formal and incremental-iterative for the experimental part. These operators were applied in digital and microscopic images. The conjunctions and implications of fuzzy morphology mathematical reasoning will be used in order to choose the best adjunction to be applied depending on the problem being approached, i.e., we will use automorphisms on the implications and observe their influence on segmenting images and then on their processing. In order to validate the developed system, it was applied to counting problems in microscopic images, extending to pathological images. It was noted that for the computation of spores the best operator was the erosion of Gödel. It developed three groups of morphological operators fuzzy, Lukasiewicz, And Godel Goguen that can have a variety applications

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis aims to analyze a corpus containing some hybrid poems and some images related to these poems because the omnipresence of some of them in different semiotic systems and even different books of Arnaldo Antunes. The poems in question will be drawn from two of his books: Things and Name; We will also consider some excerpts from songs, which are present in the same author career discs; as well as observe the embodiment of the poem video coming out of the paper support and enters on the TV screen through the VHS / DVD project Name. Our work focuses on this corpus, mainly observing a recurrent feature already observed at Masters level that is the hallmark of firstness, theoretical category developed by Charles Sanders Peirce. In addition to observe the semiotic aspect, we will also be a discussion of the relationship of verbal texts with visual and its nuances with changing media. The semiotic theory is basically anchored in Peirce vision studied by Lucia Santaella on the headquarters of language and thought (noise, verbal and visual). And with regard to the study of the songs, we use the theory of Luiz Tatit, which discusses the verbal intonation and musical indices as the party responsible for global understanding of Song.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This work presents an analysis of the behavior of some algorithms usually available in stereo correspondence literature, with full HD images (1920x1080 pixels) to establish, within the precision dilemma versus runtime applications which these methods can be better used. The images are obtained by a system composed of a stereo camera coupled to a computer via a capture board. The OpenCV library is used for computer vision operations and processing images involved. The algorithms discussed are an overall method of search for matching blocks with the Sum of the Absolute Value of the difference (Sum of Absolute Differences - SAD), a global technique based on cutting energy graph cuts, and a so-called matching technique semi -global. The criteria for analysis are processing time, the consumption of heap memory and the mean absolute error of disparity maps generated.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: To evaluate in vitro the surface roughness and bacterial adhesion of nanoparticle composites, after being subjected to different finishing and polishing systems. Materials and Methods: 66 specimens were prepared, and 30 with Filtek Z350 XT (3M ESPE, USA) and 30 with the resin IPS Empress Direct (Ivoclar Vivadent, USA), divided into 6 groups (n = 10 ). Six specimens were prepared for analysis in scanning electron microscopy (SEM) .Each kind of resin was subjected to finishing and polishing systems: Sof-Lex Pop-On discs (3M ESPE, USA) and AstropolTM system (Ivoclar Vivadent , USA), featuring the experimental group. The control group did not undergo any kind of finishing and polishing technique. The average roughness (Ra) in both groups was measured using a roughness in the setting of 0.25 mm (cut off) and surface images obtained with photomicrographs taken with a scanning electron microscope (SEM) magnified 500 times. Bacterial adherence was evaluated by determining the absorbance (OD) of the suspension of adhered cells by spectrophotometer at 570 nm. The results were submitted for analyzed with 2-way ANOVA at α=.05 and Tukey multiple comparison tests. Results: Statistically significant differences were found between the groups in terms of roughness and bacterial adhesion. Filtek Z350 XT for resin were no differences between the tested finishing and polishing systems, where the system of lowest surface roughness was the Sof-Lex Pop-On. To the resin IPS Empress Direct, the finishing and polishing system Astropol, had lower results of surface roughness. As for bacterial adhesion, the lowest optical density value for Filtek Z350 XT was for the group that used the finishing and polishing system Sof-Lex Pop-On and the resin IPS Empress Direct the group that used the Astropol system. In addition, there was a positive correlation between surface roughness and bacterial adhesion on polished surfaces (r = 0.612) Conclusions: surface roughness and bacterial adhesion are closely related. The finishing and polishing Sof-Lex Pop-On system is more suitable for nanoparticulate Filtek Z350 XT and the finishing and polishing system Astropol for resin nanohíbrida IPS Empress Direct.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

X-ray computed tomography (CT) imaging constitutes one of the most widely used diagnostic tools in radiology today with nearly 85 million CT examinations performed in the U.S in 2011. CT imparts a relatively high amount of radiation dose to the patient compared to other x-ray imaging modalities and as a result of this fact, coupled with its popularity, CT is currently the single largest source of medical radiation exposure to the U.S. population. For this reason, there is a critical need to optimize CT examinations such that the dose is minimized while the quality of the CT images is not degraded. This optimization can be difficult to achieve due to the relationship between dose and image quality. All things being held equal, reducing the dose degrades image quality and can impact the diagnostic value of the CT examination.

A recent push from the medical and scientific community towards using lower doses has spawned new dose reduction technologies such as automatic exposure control (i.e., tube current modulation) and iterative reconstruction algorithms. In theory, these technologies could allow for scanning at reduced doses while maintaining the image quality of the exam at an acceptable level. Therefore, there is a scientific need to establish the dose reduction potential of these new technologies in an objective and rigorous manner. Establishing these dose reduction potentials requires precise and clinically relevant metrics of CT image quality, as well as practical and efficient methodologies to measure such metrics on real CT systems. The currently established methodologies for assessing CT image quality are not appropriate to assess modern CT scanners that have implemented those aforementioned dose reduction technologies.

Thus the purpose of this doctoral project was to develop, assess, and implement new phantoms, image quality metrics, analysis techniques, and modeling tools that are appropriate for image quality assessment of modern clinical CT systems. The project developed image quality assessment methods in the context of three distinct paradigms, (a) uniform phantoms, (b) textured phantoms, and (c) clinical images.

The work in this dissertation used the “task-based” definition of image quality. That is, image quality was broadly defined as the effectiveness by which an image can be used for its intended task. Under this definition, any assessment of image quality requires three components: (1) A well defined imaging task (e.g., detection of subtle lesions), (2) an “observer” to perform the task (e.g., a radiologists or a detection algorithm), and (3) a way to measure the observer’s performance in completing the task at hand (e.g., detection sensitivity/specificity).

First, this task-based image quality paradigm was implemented using a novel multi-sized phantom platform (with uniform background) developed specifically to assess modern CT systems (Mercury Phantom, v3.0, Duke University). A comprehensive evaluation was performed on a state-of-the-art CT system (SOMATOM Definition Force, Siemens Healthcare) in terms of noise, resolution, and detectability as a function of patient size, dose, tube energy (i.e., kVp), automatic exposure control, and reconstruction algorithm (i.e., Filtered Back-Projection– FPB vs Advanced Modeled Iterative Reconstruction– ADMIRE). A mathematical observer model (i.e., computer detection algorithm) was implemented and used as the basis of image quality comparisons. It was found that image quality increased with increasing dose and decreasing phantom size. The CT system exhibited nonlinear noise and resolution properties, especially at very low-doses, large phantom sizes, and for low-contrast objects. Objective image quality metrics generally increased with increasing dose and ADMIRE strength, and with decreasing phantom size. The ADMIRE algorithm could offer comparable image quality at reduced doses or improved image quality at the same dose (increase in detectability index by up to 163% depending on iterative strength). The use of automatic exposure control resulted in more consistent image quality with changing phantom size.

Based on those results, the dose reduction potential of ADMIRE was further assessed specifically for the task of detecting small (<=6 mm) low-contrast (<=20 HU) lesions. A new low-contrast detectability phantom (with uniform background) was designed and fabricated using a multi-material 3D printer. The phantom was imaged at multiple dose levels and images were reconstructed with FBP and ADMIRE. Human perception experiments were performed to measure the detection accuracy from FBP and ADMIRE images. It was found that ADMIRE had equivalent performance to FBP at 56% less dose.

Using the same image data as the previous study, a number of different mathematical observer models were implemented to assess which models would result in image quality metrics that best correlated with human detection performance. The models included naïve simple metrics of image quality such as contrast-to-noise ratio (CNR) and more sophisticated observer models such as the non-prewhitening matched filter observer model family and the channelized Hotelling observer model family. It was found that non-prewhitening matched filter observers and the channelized Hotelling observers both correlated strongly with human performance. Conversely, CNR was found to not correlate strongly with human performance, especially when comparing different reconstruction algorithms.

The uniform background phantoms used in the previous studies provided a good first-order approximation of image quality. However, due to their simplicity and due to the complexity of iterative reconstruction algorithms, it is possible that such phantoms are not fully adequate to assess the clinical impact of iterative algorithms because patient images obviously do not have smooth uniform backgrounds. To test this hypothesis, two textured phantoms (classified as gross texture and fine texture) and a uniform phantom of similar size were built and imaged on a SOMATOM Flash scanner (Siemens Healthcare). Images were reconstructed using FBP and a Sinogram Affirmed Iterative Reconstruction (SAFIRE). Using an image subtraction technique, quantum noise was measured in all images of each phantom. It was found that in FBP, the noise was independent of the background (textured vs uniform). However, for SAFIRE, noise increased by up to 44% in the textured phantoms compared to the uniform phantom. As a result, the noise reduction from SAFIRE was found to be up to 66% in the uniform phantom but as low as 29% in the textured phantoms. Based on this result, it clear that further investigation was needed into to understand the impact that background texture has on image quality when iterative reconstruction algorithms are used.

To further investigate this phenomenon with more realistic textures, two anthropomorphic textured phantoms were designed to mimic lung vasculature and fatty soft tissue texture. The phantoms (along with a corresponding uniform phantom) were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Scans were repeated a total of 50 times in order to get ensemble statistics of the noise. A novel method of estimating the noise power spectrum (NPS) from irregularly shaped ROIs was developed. It was found that SAFIRE images had highly locally non-stationary noise patterns with pixels near edges having higher noise than pixels in more uniform regions. Compared to FBP, SAFIRE images had 60% less noise on average in uniform regions for edge pixels, noise was between 20% higher and 40% lower. The noise texture (i.e., NPS) was also highly dependent on the background texture for SAFIRE. Therefore, it was concluded that quantum noise properties in the uniform phantoms are not representative of those in patients for iterative reconstruction algorithms and texture should be considered when assessing image quality of iterative algorithms.

The move beyond just assessing noise properties in textured phantoms towards assessing detectability, a series of new phantoms were designed specifically to measure low-contrast detectability in the presence of background texture. The textures used were optimized to match the texture in the liver regions actual patient CT images using a genetic algorithm. The so called “Clustured Lumpy Background” texture synthesis framework was used to generate the modeled texture. Three textured phantoms and a corresponding uniform phantom were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Images were reconstructed with FBP and SAFIRE and analyzed using a multi-slice channelized Hotelling observer to measure detectability and the dose reduction potential of SAFIRE based on the uniform and textured phantoms. It was found that at the same dose, the improvement in detectability from SAFIRE (compared to FBP) was higher when measured in a uniform phantom compared to textured phantoms.

The final trajectory of this project aimed at developing methods to mathematically model lesions, as a means to help assess image quality directly from patient images. The mathematical modeling framework is first presented. The models describe a lesion’s morphology in terms of size, shape, contrast, and edge profile as an analytical equation. The models can be voxelized and inserted into patient images to create so-called “hybrid” images. These hybrid images can then be used to assess detectability or estimability with the advantage that the ground truth of the lesion morphology and location is known exactly. Based on this framework, a series of liver lesions, lung nodules, and kidney stones were modeled based on images of real lesions. The lesion models were virtually inserted into patient images to create a database of hybrid images to go along with the original database of real lesion images. ROI images from each database were assessed by radiologists in a blinded fashion to determine the realism of the hybrid images. It was found that the radiologists could not readily distinguish between real and virtual lesion images (area under the ROC curve was 0.55). This study provided evidence that the proposed mathematical lesion modeling framework could produce reasonably realistic lesion images.

Based on that result, two studies were conducted which demonstrated the utility of the lesion models. The first study used the modeling framework as a measurement tool to determine how dose and reconstruction algorithm affected the quantitative analysis of liver lesions, lung nodules, and renal stones in terms of their size, shape, attenuation, edge profile, and texture features. The same database of real lesion images used in the previous study was used for this study. That database contained images of the same patient at 2 dose levels (50% and 100%) along with 3 reconstruction algorithms from a GE 750HD CT system (GE Healthcare). The algorithms in question were FBP, Adaptive Statistical Iterative Reconstruction (ASiR), and Model-Based Iterative Reconstruction (MBIR). A total of 23 quantitative features were extracted from the lesions under each condition. It was found that both dose and reconstruction algorithm had a statistically significant effect on the feature measurements. In particular, radiation dose affected five, three, and four of the 23 features (related to lesion size, conspicuity, and pixel-value distribution) for liver lesions, lung nodules, and renal stones, respectively. MBIR significantly affected 9, 11, and 15 of the 23 features (including size, attenuation, and texture features) for liver lesions, lung nodules, and renal stones, respectively. Lesion texture was not significantly affected by radiation dose.

The second study demonstrating the utility of the lesion modeling framework focused on assessing detectability of very low-contrast liver lesions in abdominal imaging. Specifically, detectability was assessed as a function of dose and reconstruction algorithm. As part of a parallel clinical trial, images from 21 patients were collected at 6 dose levels per patient on a SOMATOM Flash scanner. Subtle liver lesion models (contrast = -15 HU) were inserted into the raw projection data from the patient scans. The projections were then reconstructed with FBP and SAFIRE (strength 5). Also, lesion-less images were reconstructed. Noise, contrast, CNR, and detectability index of an observer model (non-prewhitening matched filter) were assessed. It was found that SAFIRE reduced noise by 52%, reduced contrast by 12%, increased CNR by 87%. and increased detectability index by 65% compared to FBP. Further, a 2AFC human perception experiment was performed to assess the dose reduction potential of SAFIRE, which was found to be 22% compared to the standard of care dose.

In conclusion, this dissertation provides to the scientific community a series of new methodologies, phantoms, analysis techniques, and modeling tools that can be used to rigorously assess image quality from modern CT systems. Specifically, methods to properly evaluate iterative reconstruction have been developed and are expected to aid in the safe clinical implementation of dose reduction technologies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Este artículo analiza la existencia de un posible ritual en la antigua Macedonia que fuese similar al Akîtu babilonio. Pensamos que la victoria de Zeus sobre sus enemigos en la gigantomaquia y tifonomaquia fue celebrada por los reyes macedonios para fortalecer su posición en el poder. Aunque no hay evidencias directas de ello, tenemos una serie de diferentes fuentes en las cuales los macedonios están relacionados directa o indirectamente con Tifón y los gigantes. Además, la estratégica importancia del monte Olimpo para los macedonios nos hace pensar que la victoria de Zeus sobre las fuerzas del mal fue conmemorada de alguna forma por ellos.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A travers l’étude de la Correspondance Jules Supervielle-Etiemble et sa confrontation avec les manifestes du poète que sont  « Eléments d’une poétique » et « En songeant à un art poétique », notre article tente de définir la spécificité de la conception supervilienne de la poésie tout en évoquant les difficultés du poète à extérioriser, dans les premières années de sa jeunesse, les images inhérentes à son univers intérieur, et en faisant l’inventaire de quelques principes qui régissent son activité poétique : le refus de toutes les esthétiques et de toutes les techniques qui régissent préalablement la création poétique, l’adoption d’une technique a posteriori ou d’une « technique mouvante qui ne se fixe qu’à chaque poème », le travail de corrections, de retouches et de ratures, et le culte de l’oubli considéré moins comme une défaillance de la mémoire que comme le catalyseur des associations et images inattendues.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We address the problem of 3D-assisted 2D face recognition in scenarios when the input image is subject to degradations or exhibits intra-personal variations not captured by the 3D model. The proposed solution involves a novel approach to learn a subspace spanned by perturbations caused by the missing modes of variation and image degradations, using 3D face data reconstructed from 2D images rather than 3D capture. This is accomplished by modelling the difference in the texture map of the 3D aligned input and reference images. A training set of these texture maps then defines a perturbation space which can be represented using PCA bases. Assuming that the image perturbation subspace is orthogonal to the 3D face model space, then these additive components can be recovered from an unseen input image, resulting in an improved fit of the 3D face model. The linearity of the model leads to efficient fitting. Experiments show that our method achieves very competitive face recognition performance on Multi-PIE and AR databases. We also present baseline face recognition results on a new data set exhibiting combined pose and illumination variations as well as occlusion.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Person re-identification involves recognizing a person across non-overlapping camera views, with different pose, illumination, and camera characteristics. We propose to tackle this problem by training a deep convolutional network to represent a person’s appearance as a low-dimensional feature vector that is invariant to common appearance variations encountered in the re-identification problem. Specifically, a Siamese-network architecture is used to train a feature extraction network using pairs of similar and dissimilar images. We show that use of a novel multi-task learning objective is crucial for regularizing the network parameters in order to prevent over-fitting due to the small size the training dataset. We complement the verification task, which is at the heart of re-identification, by training the network to jointly perform verification, identification, and to recognise attributes related to the clothing and pose of the person in each image. Additionally, we show that our proposed approach performs well even in the challenging cross-dataset scenario, which may better reflect real-world expected performance. 

Relevância:

100.00% 100.00%

Publicador:

Resumo:

AIRES, Kelson R. T. ; ARAÚJO, Hélder J. ; MEDEIROS, Adelardo A. D. . Plane Detection from Monocular Image Sequences. In: VISUALIZATION, IMAGING AND IMAGE PROCESSING, 2008, Palma de Mallorca, Spain. Proceedings..., Palma de Mallorca: VIIP, 2008

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Les maladies cardiovasculaires sont la première cause de mortalité dans le monde et les anévrismes de l’aorte abdominale (AAAs) font partie de ce lot déplorable. Un anévrisme est la dilatation d’une artère pouvant conduire à la mort. Une rupture d’AAA s’avère fatale près de 80% du temps. Un moyen de traiter les AAAs est l’insertion d’une endoprothèse (SG) dans l’aorte, communément appelée la réparation endovasculaire (EVAR), afin de réduire la pression exercée par le flux sanguin sur la paroi. L’efficacité de ce traitement est compromise par la survenue d’endofuites (flux sanguins entre la prothèse et le sac anévrismal) pouvant conduire à la rupture de l’anévrisme. Ces flux sanguins peuvent survenir à n’importe quel moment après le traitement EVAR. Une surveillance par tomodensitométrie (CT-scan) annuelle est donc requise, augmentant ainsi le coût du suivi post-EVAR et exposant le patient à la radiation ionisante et aux complications des contrastes iodés. L’endotension est le concept de dilatation de l’anévrisme sans la présence d’une endofuite apparente au CT-scan. Après le traitement EVAR, le sang dans le sac anévrismal coagule pour former un thrombus frais, qui deviendra progressivement un thrombus plus fibreux et plus organisé, donnant lieu à un rétrécissement de l’anévrisme. Il y a très peu de données dans la littérature pour étudier ce processus temporel et la relation entre le thrombus frais et l’endotension. L’étalon d’or du suivi post-EVAR, le CT-scan, ne peut pas détecter la présence de thrombus frais. Il y a donc un besoin d’investir dans une technique sécuritaire et moins coûteuse pour le suivi d’AAAs après EVAR. Une méthode récente, l’élastographie dynamique, mesure l’élasticité des tissus en temps réel. Le principe de cette technique repose sur la génération d’ondes de cisaillement et l’étude de leur propagation afin de remonter aux propriétés mécaniques du milieu étudié. Cette thèse vise l’application de l’élastographie dynamique pour la détection des endofuites ainsi que de la caractérisation mécanique des tissus du sac anévrismal après le traitement EVAR. Ce projet dévoile le potentiel de l’élastographie afin de réduire les dangers de la radiation, de l’utilisation d’agent de contraste ainsi que des coûts du post-EVAR des AAAs. L’élastographie dynamique utilisant le « Shear Wave Imaging » (SWI) est prometteuse. Cette modalité pourrait complémenter l’échographie-Doppler (DUS) déjà utilisée pour le suivi d’examen post-EVAR. Le SWI a le potentiel de fournir des informations sur l’organisation fibreuse du thrombus ainsi que sur la détection d’endofuites. Tout d’abord, le premier objectif de cette thèse consistait à tester le SWI sur des AAAs dans des modèles canins pour la détection d’endofuites et la caractérisation du thrombus. Des SGs furent implantées dans un groupe de 18 chiens avec un anévrisme créé au moyen de la veine jugulaire. 4 anévrismes avaient une endofuite de type I, 13 avaient une endofuite de type II et un anévrisme n’avait pas d’endofuite. Des examens échographiques, DUS et SWI ont été réalisés à l’implantation, puis 1 semaine, 1 mois, 3 mois et 6 mois après le traitement EVAR. Une angiographie, un CT-scan et des coupes macroscopiques ont été produits au sacrifice. Les régions d’endofuites, de thrombus frais et de thrombus organisé furent identifiées et segmentées. Les valeurs de rigidité données par le SWI des différentes régions furent comparées. Celles-ci furent différentes de façon significative (P < 0.001). Également, le SWI a pu détecter la présence d’endofuites où le CT-scan (1) et le DUS (3) ont échoué. Dans la continuité de ces travaux, le deuxième objectif de ce projet fut de caractériser l’évolution du thrombus dans le temps, de même que l’évolution des endofuites après embolisation dans des modèles canins. Dix-huit anévrismes furent créés dans les artères iliaques de neuf modèles canins, suivis d’une endofuite de type I après EVAR. Deux gels embolisants (Chitosan (Chi) ou Chitosan-Sodium-Tetradecyl-Sulfate (Chi-STS)) furent injectés dans le sac anévrismal pour promouvoir la guérison. Des examens échographiques, DUS et SWI ont été effectués à l’implantation et après 1 semaine, 1 mois, 3 mois et 6 mois. Une angiographie, un CT-scan et un examen histologique ont été réalisés au sacrifice afin d’évaluer la présence, le type et la grosseur de l’endofuite. Les valeurs du module d’élasticité des régions d’intérêts ont été identifiées et segmentées sur les données pathologiques. Les régions d’endofuites et de thrombus frais furent différentes de façon significative comparativement aux autres régions (P < 0.001). Les valeurs d’élasticité du thrombus frais à 1 semaine et à 3 mois indiquent que le SWI peut évaluer la maturation du thrombus, de même que caractériser l’évolution et la dégradation des gels embolisants dans le temps. Le SWI a pu détecter des endofuites où le DUS a échoué (2) et, contrairement au CT-scan, détecter la présence de thrombus frais. Finalement, la dernière étape du projet doctoral consistait à appliquer le SWI dans une phase clinique, avec des patients humains ayant déjà un AAA, pour la détection d’endofuite et la caractérisation de l’élasticité des tissus. 25 patients furent sélectionnés pour participer à l’étude. Une comparaison d’imagerie a été produite entre le SWI, le CT-scan et le DUS. Les valeurs de rigidité données par le SWI des différentes régions (endofuite, thrombus) furent identifiées et segmentées. Celles-ci étaient distinctes de façon significative (P < 0.001). Le SWI a détecté 5 endofuites sur 6 (sensibilité de 83.3%) et a eu 6 faux positifs (spécificité de 76%). Le SWI a pu détecter la présence d’endofuites où le CT-scan (2) ainsi que le DUS (2) ont échoué. Il n’y avait pas de différence statistique notable entre la rigidité du thrombus pour un AAA avec endofuite et un AAA sans endofuite. Aucune corrélation n’a pu être établie de façon significative entre les diamètres des AAAs ainsi que leurs variations et l’élasticité du thrombus. Le SWI a le potentiel de détecter les endofuites et caractériser le thrombus selon leurs propriétés mécaniques. Cette technique pourrait être combinée au suivi des AAAs post-EVAR, complémentant ainsi l’imagerie DUS et réduisant le coût et l’exposition à la radiation ionisante et aux agents de contrastes néphrotoxiques.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

During sleep, humans experience the offline images and sensations that we call dreams, which are typically emotional and lacking in rational judgment of their bizarreness. However, during lucid dreaming (LD), subjects know that they are dreaming, and may control oneiric content. Dreaming and LD features have been studied in North Americans, Europeans and Asians, but not among Brazilians, the largest population in Latin America. Here we investigated dreams and LD characteristics in a Brazilian sample (n=3,427; median age=25 years) through an online survey. The subjects reported recalling dreams at least once a week (76%), and that dreams typically depicted actions (93%), known people (92%), sounds/voices (78%), and colored images (76%). The oneiric content was associated with plans for the upcoming days (37%), memories of the previous day (13%), or unrelated to the dreamer (30%). Nightmares usually depicted anxiety/fear (65%), being stalked (48%), or other unpleasant sensations(47%). These data corroborate Freudian notion of day residue in dreams, and suggest that dreams and nightmares are simulations of life situations that are related to our psychobiological integrity. Regarding LD, we observed that 77% of the subjects experienced LD at least once in life (44% up to 10 episodes ever), and for 48% LD subjectively lasted less than 1 min. LD frequency correlated weakly with dream recall frequency (r =0.20,p< 0.01), and LD control was rare (29%). LD occurrence was facilitated when subjects did not need to wake up early (38%), a situation that increases rapid eye movement sleep (REMS) duration, or when subjects were under stress (30%), which increases REMS transitions into waking. These results indicate that LD is relatively ubiquitous but rare, unstable, difficult to control, and facilitated by increases in REMS duration and transitions to wake state. Together with LD incidence in USA, Europe and Asia, our data from Latin America strengthen the notion that LD is a general phenomenon of the human species.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Thesis (Ph.D.)--University of Washington, 2016-07