744 resultados para body-image perception
Resumo:
One developing theme in consciousness research is that consciousness is not the product of any specific component of the brain, rather it is an emergent property of the changing patterns of connectivity between different specialised functional components. For example, the dynamic core hypothesis proposes that conscious experience requires high levels of neural complexity, where complexity is defined in terms of functional connectivity. To test this hypothesis, electroencephalography was recorded while participants were shown random dot-stereograms. Consistent with the dynamic core hypothesis, neural complexity increased as the participants changed from simply viewing the stereogram to consciously perceiving the hidden 3D image.
Resumo:
With the progress of computer technology, computers are expected to be more intelligent in the interaction with humans, presenting information according to the user's psychological and physiological characteristics. However, computer users with visual problems may encounter difficulties on the perception of icons, menus, and other graphical information displayed on the screen, limiting the efficiency of their interaction with computers. In this dissertation, a personalized and dynamic image precompensation method was developed to improve the visual performance of the computer users with ocular aberrations. The precompensation was applied on the graphical targets before presenting them on the screen, aiming to counteract the visual blurring caused by the ocular aberration of the user's eye. A complete and systematic modeling approach to describe the retinal image formation of the computer user was presented, taking advantage of modeling tools, such as Zernike polynomials, wavefront aberration, Point Spread Function and Modulation Transfer Function. The ocular aberration of the computer user was originally measured by a wavefront aberrometer, as a reference for the precompensation model. The dynamic precompensation was generated based on the resized aberration, with the real-time pupil diameter monitored. The potential visual benefit of the dynamic precompensation method was explored through software simulation, with the aberration data from a real human subject. An "artificial eye'' experiment was conducted by simulating the human eye with a high-definition camera, providing objective evaluation to the image quality after precompensation. In addition, an empirical evaluation with 20 human participants was also designed and implemented, involving image recognition tests performed under a more realistic viewing environment of computer use. The statistical analysis results of the empirical experiment confirmed the effectiveness of the dynamic precompensation method, by showing significant improvement on the recognition accuracy. The merit and necessity of the dynamic precompensation were also substantiated by comparing it with the static precompensation. The visual benefit of the dynamic precompensation was further confirmed by the subjective assessments collected from the evaluation participants.
Resumo:
A BODY OUTSIDE THE KREMLIN is a historical mystery novel set in the Northern Camps of Special Significance, a Soviet Russian penal institution based in the Solovetsky Archipelago during the 1920s. The protagonist, working first with the camp authorities, then in spite of their disapproval, solves the murder of a fellow prisoner. In the process he improves his position within the camp, while also becoming hardened to the brutal necessities of camp life. Prior to the establishment of the penal camp, the Solovetsky Archipelago was the site of an important Russian Orthodox monastery, and the mystery proves to involve valuables, particularly icons, seized from the monks by the Soviet secret police. Thus the novel treats themes not only of statist repression, but also religious epiphany and the problems of true perception in a world of symbols.
Resumo:
The objective of this research is to describe and analyze in literary corpus, the way we conceptualize emotions, especially anger. Using the assumptions of the call Cognitive Theory of Metaphor, present a general overview of cognition metaphor on the basis of Cognitive Linguistics, and in a deeper way, we analyze the metaphorical conceptualization of anger. The proposal embodied mind, prevalent in current cognitive science, is fundamental for studies involving mental simulation. Recent research shows that the metaphor is the result of cognitive processes that involve our perception sensorimotor combined with socio-cultural experiences. The ability to build via frequency standards for our experiments is crucial to our language, including metaphorical constructions. Such constructions are the result of cognitive processes that involve the relationship between image schemas and frames. Image schemas comes from our sensorimotor experience, which lists the limits of our bodies to the limits of our surroundings, and frames, in turn, comes from our ability to stock sociocultural events. The metaphorical construction is the result of this constant relationship between body, mind and culture, situating us in bodily experiences and cultural. By analyzing five national literary works, we created an analytical framework on how anger is understood, specifically in Portuguese language. The results are important to understand, through language, how culture is part of our cognition, in conjunction with the sensorimotor aspects.
Resumo:
X-ray computed tomography (CT) imaging constitutes one of the most widely used diagnostic tools in radiology today with nearly 85 million CT examinations performed in the U.S in 2011. CT imparts a relatively high amount of radiation dose to the patient compared to other x-ray imaging modalities and as a result of this fact, coupled with its popularity, CT is currently the single largest source of medical radiation exposure to the U.S. population. For this reason, there is a critical need to optimize CT examinations such that the dose is minimized while the quality of the CT images is not degraded. This optimization can be difficult to achieve due to the relationship between dose and image quality. All things being held equal, reducing the dose degrades image quality and can impact the diagnostic value of the CT examination.
A recent push from the medical and scientific community towards using lower doses has spawned new dose reduction technologies such as automatic exposure control (i.e., tube current modulation) and iterative reconstruction algorithms. In theory, these technologies could allow for scanning at reduced doses while maintaining the image quality of the exam at an acceptable level. Therefore, there is a scientific need to establish the dose reduction potential of these new technologies in an objective and rigorous manner. Establishing these dose reduction potentials requires precise and clinically relevant metrics of CT image quality, as well as practical and efficient methodologies to measure such metrics on real CT systems. The currently established methodologies for assessing CT image quality are not appropriate to assess modern CT scanners that have implemented those aforementioned dose reduction technologies.
Thus the purpose of this doctoral project was to develop, assess, and implement new phantoms, image quality metrics, analysis techniques, and modeling tools that are appropriate for image quality assessment of modern clinical CT systems. The project developed image quality assessment methods in the context of three distinct paradigms, (a) uniform phantoms, (b) textured phantoms, and (c) clinical images.
The work in this dissertation used the “task-based” definition of image quality. That is, image quality was broadly defined as the effectiveness by which an image can be used for its intended task. Under this definition, any assessment of image quality requires three components: (1) A well defined imaging task (e.g., detection of subtle lesions), (2) an “observer” to perform the task (e.g., a radiologists or a detection algorithm), and (3) a way to measure the observer’s performance in completing the task at hand (e.g., detection sensitivity/specificity).
First, this task-based image quality paradigm was implemented using a novel multi-sized phantom platform (with uniform background) developed specifically to assess modern CT systems (Mercury Phantom, v3.0, Duke University). A comprehensive evaluation was performed on a state-of-the-art CT system (SOMATOM Definition Force, Siemens Healthcare) in terms of noise, resolution, and detectability as a function of patient size, dose, tube energy (i.e., kVp), automatic exposure control, and reconstruction algorithm (i.e., Filtered Back-Projection– FPB vs Advanced Modeled Iterative Reconstruction– ADMIRE). A mathematical observer model (i.e., computer detection algorithm) was implemented and used as the basis of image quality comparisons. It was found that image quality increased with increasing dose and decreasing phantom size. The CT system exhibited nonlinear noise and resolution properties, especially at very low-doses, large phantom sizes, and for low-contrast objects. Objective image quality metrics generally increased with increasing dose and ADMIRE strength, and with decreasing phantom size. The ADMIRE algorithm could offer comparable image quality at reduced doses or improved image quality at the same dose (increase in detectability index by up to 163% depending on iterative strength). The use of automatic exposure control resulted in more consistent image quality with changing phantom size.
Based on those results, the dose reduction potential of ADMIRE was further assessed specifically for the task of detecting small (<=6 mm) low-contrast (<=20 HU) lesions. A new low-contrast detectability phantom (with uniform background) was designed and fabricated using a multi-material 3D printer. The phantom was imaged at multiple dose levels and images were reconstructed with FBP and ADMIRE. Human perception experiments were performed to measure the detection accuracy from FBP and ADMIRE images. It was found that ADMIRE had equivalent performance to FBP at 56% less dose.
Using the same image data as the previous study, a number of different mathematical observer models were implemented to assess which models would result in image quality metrics that best correlated with human detection performance. The models included naïve simple metrics of image quality such as contrast-to-noise ratio (CNR) and more sophisticated observer models such as the non-prewhitening matched filter observer model family and the channelized Hotelling observer model family. It was found that non-prewhitening matched filter observers and the channelized Hotelling observers both correlated strongly with human performance. Conversely, CNR was found to not correlate strongly with human performance, especially when comparing different reconstruction algorithms.
The uniform background phantoms used in the previous studies provided a good first-order approximation of image quality. However, due to their simplicity and due to the complexity of iterative reconstruction algorithms, it is possible that such phantoms are not fully adequate to assess the clinical impact of iterative algorithms because patient images obviously do not have smooth uniform backgrounds. To test this hypothesis, two textured phantoms (classified as gross texture and fine texture) and a uniform phantom of similar size were built and imaged on a SOMATOM Flash scanner (Siemens Healthcare). Images were reconstructed using FBP and a Sinogram Affirmed Iterative Reconstruction (SAFIRE). Using an image subtraction technique, quantum noise was measured in all images of each phantom. It was found that in FBP, the noise was independent of the background (textured vs uniform). However, for SAFIRE, noise increased by up to 44% in the textured phantoms compared to the uniform phantom. As a result, the noise reduction from SAFIRE was found to be up to 66% in the uniform phantom but as low as 29% in the textured phantoms. Based on this result, it clear that further investigation was needed into to understand the impact that background texture has on image quality when iterative reconstruction algorithms are used.
To further investigate this phenomenon with more realistic textures, two anthropomorphic textured phantoms were designed to mimic lung vasculature and fatty soft tissue texture. The phantoms (along with a corresponding uniform phantom) were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Scans were repeated a total of 50 times in order to get ensemble statistics of the noise. A novel method of estimating the noise power spectrum (NPS) from irregularly shaped ROIs was developed. It was found that SAFIRE images had highly locally non-stationary noise patterns with pixels near edges having higher noise than pixels in more uniform regions. Compared to FBP, SAFIRE images had 60% less noise on average in uniform regions for edge pixels, noise was between 20% higher and 40% lower. The noise texture (i.e., NPS) was also highly dependent on the background texture for SAFIRE. Therefore, it was concluded that quantum noise properties in the uniform phantoms are not representative of those in patients for iterative reconstruction algorithms and texture should be considered when assessing image quality of iterative algorithms.
The move beyond just assessing noise properties in textured phantoms towards assessing detectability, a series of new phantoms were designed specifically to measure low-contrast detectability in the presence of background texture. The textures used were optimized to match the texture in the liver regions actual patient CT images using a genetic algorithm. The so called “Clustured Lumpy Background” texture synthesis framework was used to generate the modeled texture. Three textured phantoms and a corresponding uniform phantom were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Images were reconstructed with FBP and SAFIRE and analyzed using a multi-slice channelized Hotelling observer to measure detectability and the dose reduction potential of SAFIRE based on the uniform and textured phantoms. It was found that at the same dose, the improvement in detectability from SAFIRE (compared to FBP) was higher when measured in a uniform phantom compared to textured phantoms.
The final trajectory of this project aimed at developing methods to mathematically model lesions, as a means to help assess image quality directly from patient images. The mathematical modeling framework is first presented. The models describe a lesion’s morphology in terms of size, shape, contrast, and edge profile as an analytical equation. The models can be voxelized and inserted into patient images to create so-called “hybrid” images. These hybrid images can then be used to assess detectability or estimability with the advantage that the ground truth of the lesion morphology and location is known exactly. Based on this framework, a series of liver lesions, lung nodules, and kidney stones were modeled based on images of real lesions. The lesion models were virtually inserted into patient images to create a database of hybrid images to go along with the original database of real lesion images. ROI images from each database were assessed by radiologists in a blinded fashion to determine the realism of the hybrid images. It was found that the radiologists could not readily distinguish between real and virtual lesion images (area under the ROC curve was 0.55). This study provided evidence that the proposed mathematical lesion modeling framework could produce reasonably realistic lesion images.
Based on that result, two studies were conducted which demonstrated the utility of the lesion models. The first study used the modeling framework as a measurement tool to determine how dose and reconstruction algorithm affected the quantitative analysis of liver lesions, lung nodules, and renal stones in terms of their size, shape, attenuation, edge profile, and texture features. The same database of real lesion images used in the previous study was used for this study. That database contained images of the same patient at 2 dose levels (50% and 100%) along with 3 reconstruction algorithms from a GE 750HD CT system (GE Healthcare). The algorithms in question were FBP, Adaptive Statistical Iterative Reconstruction (ASiR), and Model-Based Iterative Reconstruction (MBIR). A total of 23 quantitative features were extracted from the lesions under each condition. It was found that both dose and reconstruction algorithm had a statistically significant effect on the feature measurements. In particular, radiation dose affected five, three, and four of the 23 features (related to lesion size, conspicuity, and pixel-value distribution) for liver lesions, lung nodules, and renal stones, respectively. MBIR significantly affected 9, 11, and 15 of the 23 features (including size, attenuation, and texture features) for liver lesions, lung nodules, and renal stones, respectively. Lesion texture was not significantly affected by radiation dose.
The second study demonstrating the utility of the lesion modeling framework focused on assessing detectability of very low-contrast liver lesions in abdominal imaging. Specifically, detectability was assessed as a function of dose and reconstruction algorithm. As part of a parallel clinical trial, images from 21 patients were collected at 6 dose levels per patient on a SOMATOM Flash scanner. Subtle liver lesion models (contrast = -15 HU) were inserted into the raw projection data from the patient scans. The projections were then reconstructed with FBP and SAFIRE (strength 5). Also, lesion-less images were reconstructed. Noise, contrast, CNR, and detectability index of an observer model (non-prewhitening matched filter) were assessed. It was found that SAFIRE reduced noise by 52%, reduced contrast by 12%, increased CNR by 87%. and increased detectability index by 65% compared to FBP. Further, a 2AFC human perception experiment was performed to assess the dose reduction potential of SAFIRE, which was found to be 22% compared to the standard of care dose.
In conclusion, this dissertation provides to the scientific community a series of new methodologies, phantoms, analysis techniques, and modeling tools that can be used to rigorously assess image quality from modern CT systems. Specifically, methods to properly evaluate iterative reconstruction have been developed and are expected to aid in the safe clinical implementation of dose reduction technologies.
Resumo:
Social decision-making is often complex, requiring the decision-maker to make social inferences about another person in addition to engaging traditional decision-making processes. However, until recently, much research in neuroeconomics and behavioral economics has examined social decision-making while failing to take into account the importance of the social context and social cognitive processes that are engaged when viewing another person. Using social psychological theory to guide our hypotheses, four research studies investigate the role of social cognition and person perception in guiding economic decisions made in social contexts. The first study (Chapter 2) demonstrates that only specific types of social information engage brain regions implicated in social cognition and hinder learning in social contexts. Study 2 (Chapter 3) extends these findings and examines contexts in which this social information is used to generalize across contexts to form predictions about another person’s behavior. Study 3 (Chapter 4) demonstrates that under certain contexts these social cognitive processes may be withheld in order to more effectively complete the task at hand. Last, Study 4 (Chapter 5) examines how this knowledge of social cognitive processing can be used to change behavior in a prosocial group context. Taken together, these studies add to the growing body of literature examining decision-making in social contexts and highlight the importance of social cognitive processing in guiding these decisions. Although social cognitive processing typically facilitates social interactions, these processes may alter economic decision-making in social contexts.
Resumo:
X-ray computed tomography (CT) is a non-invasive medical imaging technique that generates cross-sectional images by acquiring attenuation-based projection measurements at multiple angles. Since its first introduction in the 1970s, substantial technical improvements have led to the expanding use of CT in clinical examinations. CT has become an indispensable imaging modality for the diagnosis of a wide array of diseases in both pediatric and adult populations [1, 2]. Currently, approximately 272 million CT examinations are performed annually worldwide, with nearly 85 million of these in the United States alone [3]. Although this trend has decelerated in recent years, CT usage is still expected to increase mainly due to advanced technologies such as multi-energy [4], photon counting [5], and cone-beam CT [6].
Despite the significant clinical benefits, concerns have been raised regarding the population-based radiation dose associated with CT examinations [7]. From 1980 to 2006, the effective dose from medical diagnostic procedures rose six-fold, with CT contributing to almost half of the total dose from medical exposure [8]. For each patient, the risk associated with a single CT examination is likely to be minimal. However, the relatively large population-based radiation level has led to enormous efforts among the community to manage and optimize the CT dose.
As promoted by the international campaigns Image Gently and Image Wisely, exposure to CT radiation should be appropriate and safe [9, 10]. It is thus a responsibility to optimize the amount of radiation dose for CT examinations. The key for dose optimization is to determine the minimum amount of radiation dose that achieves the targeted image quality [11]. Based on such principle, dose optimization would significantly benefit from effective metrics to characterize radiation dose and image quality for a CT exam. Moreover, if accurate predictions of the radiation dose and image quality were possible before the initiation of the exam, it would be feasible to personalize it by adjusting the scanning parameters to achieve a desired level of image quality. The purpose of this thesis is to design and validate models to quantify patient-specific radiation dose prospectively and task-based image quality. The dual aim of the study is to implement the theoretical models into clinical practice by developing an organ-based dose monitoring system and an image-based noise addition software for protocol optimization.
More specifically, Chapter 3 aims to develop an organ dose-prediction method for CT examinations of the body under constant tube current condition. The study effectively modeled the anatomical diversity and complexity using a large number of patient models with representative age, size, and gender distribution. The dependence of organ dose coefficients on patient size and scanner models was further evaluated. Distinct from prior work, these studies use the largest number of patient models to date with representative age, weight percentile, and body mass index (BMI) range.
With effective quantification of organ dose under constant tube current condition, Chapter 4 aims to extend the organ dose prediction system to tube current modulated (TCM) CT examinations. The prediction, applied to chest and abdominopelvic exams, was achieved by combining a convolution-based estimation technique that quantifies the radiation field, a TCM scheme that emulates modulation profiles from major CT vendors, and a library of computational phantoms with representative sizes, ages, and genders. The prospective quantification model is validated by comparing the predicted organ dose with the dose estimated based on Monte Carlo simulations with TCM function explicitly modeled.
Chapter 5 aims to implement the organ dose-estimation framework in clinical practice to develop an organ dose-monitoring program based on a commercial software (Dose Watch, GE Healthcare, Waukesha, WI). In the first phase of the study we focused on body CT examinations, and so the patient’s major body landmark information was extracted from the patient scout image in order to match clinical patients against a computational phantom in the library. The organ dose coefficients were estimated based on CT protocol and patient size as reported in Chapter 3. The exam CTDIvol, DLP, and TCM profiles were extracted and used to quantify the radiation field using the convolution technique proposed in Chapter 4.
With effective methods to predict and monitor organ dose, Chapters 6 aims to develop and validate improved measurement techniques for image quality assessment. Chapter 6 outlines the method that was developed to assess and predict quantum noise in clinical body CT images. Compared with previous phantom-based studies, this study accurately assessed the quantum noise in clinical images and further validated the correspondence between phantom-based measurements and the expected clinical image quality as a function of patient size and scanner attributes.
Chapter 7 aims to develop a practical strategy to generate hybrid CT images and assess the impact of dose reduction on diagnostic confidence for the diagnosis of acute pancreatitis. The general strategy is (1) to simulate synthetic CT images at multiple reduced-dose levels from clinical datasets using an image-based noise addition technique; (2) to develop quantitative and observer-based methods to validate the realism of simulated low-dose images; (3) to perform multi-reader observer studies on the low-dose image series to assess the impact of dose reduction on the diagnostic confidence for multiple diagnostic tasks; and (4) to determine the dose operating point for clinical CT examinations based on the minimum diagnostic performance to achieve protocol optimization.
Chapter 8 concludes the thesis with a summary of accomplished work and a discussion about future research.
Resumo:
Depuis le début du XXIe siècle, un type particulier d’images a envahi l’espace public constitué par Internet : il s’agit des images principales de profil, ces images que les utilisateurs de sites de réseaux sociaux choisissent pour les représenter auprès des autres individus connectés. Comme le plus souvent il s’agit d’une image du corps de celui ou celle qui s’affiche ainsi, il est intéressant de s’intéresser à cette pratique en la rattachant à des pratiques plus anciennes. Dans un premier temps, cette étude présente donc une perspective socio-historique en notant la ressemblance de la pratique de l’image principale de profil avec celle de l’autoportrait et du portrait commandé. Cela permet de remarquer plusieurs points de rupture ou d’inflexion dans l’usage de ce type d’images, mais aussi d’en dégager les usages sociaux typiques. Ensuite, l’observation d’un lieu particulier d’Internet permet de tirer les conclusions suivantes : si l’usage principal de ces images est facile à expliquer, elles servent à symboliser une présence dans des lieux non accessibles aux corps sensibles, ces images montrent toujours des éléments qui permettent de déduire une position sociale et elles sont fondamentalement identiques aux images produites avant Internet. Ensuite, l’étude de ces images montre qu’il y a un véritable continuum dans la manière de dévoiler son intimité qui permet d’affirmer que la frontière entre public et privé n’existe pas sur Internet. Finalement, ces images montrent une absence de canon quant à leur production et une multiplicité des façons de se mettre en scène qui laissent à penser qu’elles sont devenues des symboles à part entière dans la communication qui peut s’établir entre des étrangers sur Internet.
Resumo:
El Psicodrama, creado por Jacob Levy Moreno, es un método de abordaje del individuo, la familia, los grupos y la comunidad, que tuvo sus orígenes en el teatro, la sociología y la psicología. Desde su trabajo con niños y grupos marginales, su creador va gestando su teoría y su técnica. Se desarrolla en diferentes ámbitos: clínica, psicoterapia, pedagogía y comunidad. El cuerpo, el juego, el espacio y el tiempo confluyen de una manera poética y contribuyen a la libertad creadora y a la tan añorada espontaneidad. El Psicodrama se desarrolla en ese espacio transicional, que no siendo ni interno, ni externo, se transforma en ese espacio pluripotencial de infinitas posibilidades creadoras.
Resumo:
The aim of art, as transformer of the individual, has numerous sides which give the human beings a sense of enhancement and growth. It is considered by university the need for our students to take part of this process of social transformation in which they feel the need of helping the community when its members are at risk of social exclusion. Art is considered to be a means, a tool and a purpose for an artis-pedagogue to be used as a guide for the renewal. And the university is also considered as a focus of commitment by means of the development of good practices as well as adopting an open and innovative attitude to any changes aimed at living harmoniously within a more just society.
Resumo:
The essay intends to demonstrate the need to postulate an (invisible) spiritual reality in the human being as a condition of possibility of the world’s visibility. For that purpose, it employs the phenomenological method. It describes the structure of human sensitive perception. This description firstly presents the world’s structure (first distance). Secondly, it manifests the importance of the body as a border element between the soul and the world and as an organ by means of which the soul structures the world. It is the second distance the distance between the body and the world. But, thirdly, the description states that, no matter how intimately the soul interpenetrates the body and requires it to know the world, the soul is not the body. It is the third distance or the primordial distance between the soul and everything else (including the body itself).
Resumo:
Digital Image Processing is a rapidly evolving eld with growing applications in Science and Engineering. It involves changing the nature of an image in order to either improve its pictorial information for human interpretation or render it more suitable for autonomous machine perception. One of the major areas of image processing for human vision applications is image enhancement. The principal goal of image enhancement is to improve visual quality of an image, typically by taking advantage of the response of human visual system. Image enhancement methods are carried out usually in the pixel domain. Transform domain methods can often provide another way to interpret and understand image contents. A suitable transform, thus selected, should have less computational complexity. Sequency ordered arrangement of unique MRT (Mapped Real Transform) coe cients can give rise to an integer-to-integer transform, named Sequency based unique MRT (SMRT), suitable for image processing applications. The development of the SMRT from UMRT (Unique MRT), forward & inverse SMRT algorithms and the basis functions are introduced. A few properties of the SMRT are explored and its scope in lossless text compression is presented.
Resumo:
Au cours des dernières décennies, l’effort sur les applications de capteurs infrarouges a largement progressé dans le monde. Mais, une certaine difficulté demeure, en ce qui concerne le fait que les objets ne sont pas assez clairs ou ne peuvent pas toujours être distingués facilement dans l’image obtenue pour la scène observée. L’amélioration de l’image infrarouge a joué un rôle important dans le développement de technologies de la vision infrarouge de l’ordinateur, le traitement de l’image et les essais non destructifs, etc. Cette thèse traite de la question des techniques d’amélioration de l’image infrarouge en deux aspects, y compris le traitement d’une seule image infrarouge dans le domaine hybride espacefréquence, et la fusion d’images infrarouges et visibles employant la technique du nonsubsampled Contourlet transformer (NSCT). La fusion d’images peut être considérée comme étant la poursuite de l’exploration du modèle d’amélioration de l’image unique infrarouge, alors qu’il combine les images infrarouges et visibles en une seule image pour représenter et améliorer toutes les informations utiles et les caractéristiques des images sources, car une seule image ne pouvait contenir tous les renseignements pertinents ou disponibles en raison de restrictions découlant de tout capteur unique de l’imagerie. Nous examinons et faisons une enquête concernant le développement de techniques d’amélioration d’images infrarouges, et ensuite nous nous consacrons à l’amélioration de l’image unique infrarouge, et nous proposons un schéma d’amélioration de domaine hybride avec une méthode d’évaluation floue de seuil amélioré, qui permet d’obtenir une qualité d’image supérieure et améliore la perception visuelle humaine. Les techniques de fusion d’images infrarouges et visibles sont établies à l’aide de la mise en oeuvre d’une mise en registre précise des images sources acquises par différents capteurs. L’algorithme SURF-RANSAC est appliqué pour la mise en registre tout au long des travaux de recherche, ce qui conduit à des images mises en registre de façon très précise et des bénéfices accrus pour le traitement de fusion. Pour les questions de fusion d’images infrarouges et visibles, une série d’approches avancées et efficaces sont proposés. Une méthode standard de fusion à base de NSCT multi-canal est présente comme référence pour les approches de fusion proposées suivantes. Une approche conjointe de fusion, impliquant l’Adaptive-Gaussian NSCT et la transformée en ondelettes (Wavelet Transform, WT) est propose, ce qui conduit à des résultats de fusion qui sont meilleurs que ceux obtenus avec les méthodes non-adaptatives générales. Une approche de fusion basée sur le NSCT employant la détection comprime (CS, compressed sensing) et de la variation totale (TV) à des coefficients d’échantillons clairsemés et effectuant la reconstruction de coefficients fusionnés de façon précise est proposée, qui obtient de bien meilleurs résultats de fusion par le biais d’une pré-amélioration de l’image infrarouge et en diminuant les informations redondantes des coefficients de fusion. Une procédure de fusion basée sur le NSCT utilisant une technique de détection rapide de rétrécissement itératif comprimé (fast iterative-shrinking compressed sensing, FISCS) est proposée pour compresser les coefficients décomposés et reconstruire les coefficients fusionnés dans le processus de fusion, qui conduit à de meilleurs résultats plus rapidement et d’une manière efficace.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
Résumé: Les pratiques du Sensible sont des pratiques d’accompagnement formatives et soignantes. Elles permettent d’apprendre comment l’expérience du corps et de son mouvement interne conduit au développement de la conscience et de la présence à soi ainsi qu’à l’autre, des qualités enviables pour des professionnelles et professionnels de la relation d’aide du secteur de la santé. Dans ces pratiques, le corps joue un rôle central à travers quatre types d’intervention : la thérapie manuelle, la gymnastique sensorielle, l’introspection sensorielle et l’entretien verbal à propos de l’expérience corporelle. Selon Large (2009), une qualité de présence particulière se construit chez les participantes et le participant aux pratiques du Sensible. Selon lui, ceux-ci se rapprochent de leur intériorité, parviennent à verbaliser à autrui ce qu’ils ressentent et en arrivent à poser des actions qui expriment ce qu’ils deviennent. Large (2009) constate qu’ils acquièrent de la stabilité, de l’adaptabilité et de l’autonomie. À la fois plus affirmatifs, ils gagnent aussi en proximité à l’autre. Bois (2007) note un changement de représentation lié aux idées, aux valeurs, à l’image de soi et au rapport perceptif à soi. Une chercheure et des chercheurs constatent un changement de conception de la santé (Duval, 2010; Laemmlin-Cencig et Humpich, 2009). À notre connaissance, il n’y a pas eu d’étude antérieure concernant l’influence des pratiques du Sensible auprès de médecins. Nous avons exploré, le cas échéant, comment une formation aux pratiques du Sensible, suivie par des médecins, a modifié leur rapport à leur corps, à leur propre santé, à leur conception de la santé, à la qualité de leur présence à eux-mêmes, aux autres professionnelles et professionnels et aux patientes et patients. Des entretiens semi-structurés d’une durée de 90 à 105 minutes ont été effectués auprès de six médecins français (cinq femmes et un homme) ayant été formés aux pratiques du Sensible entre 2005 et 2012. Deux types d’entretiens à visée compréhensive (Kaufmann, 2011) et d’explicitation (Vermersch, 2010; 2012) ont été réalisés. Des informations ont aussi été recueillies sur la formation et les activités professionnelles des participantes et du participant. Deux démarches d’analyse ont été utilisées, entre autres pour vérifier la cohérence des résultats et augmenter la rigueur de notre projet. Notre première démarche d’analyse a été conçue à partir de deux méthodes : au départ avec l’analyse thématique et par la suite une analyse avec les catégories conceptualisantes afin de déboucher sur une théorisation ancrée. La deuxième démarche d’analyse a consisté à créer une liste de vingt-six phénomènes présents pour la majorité des entretiens suite à des discussions tenues avec notre équipe de direction. Selon nos résultats, suite à la formation aux pratiques du Sensible, les cinq participantes témoignent d’une plus grande proximité et attention à leur corps et d’une meilleure écoute de celui-ci. Cet ancrage corporel de leur présence les informe davantage sur leur mode de vie et d’existence. Il en ressort ainsi des prises de conscience importantes grâce auxquelles les participantes font des choix nouveaux pour une vie plus cohérente et recentrée sur leur intériorité. Par le fait même, elles récupèrent leur pouvoir sur leur vie comme sur leur santé. En outre, parmi les six médecins, quatre ont modifié leur conception de la santé. Celle-ci s’est en effet élargie pour inclure de nouveaux éléments, dont la qualité du rapport à soi et l’accordage entre le corps et la pensée. Le corps semble être une voie souterraine à partir de laquelle se sont réalisées des transformations dans la personne, comme si le corps devenait une interface ayant des effets sur plusieurs facettes de la personne. Ces transformations semblent avoir une influence sur la manière dont celle-ci exerce sa profession, comme si un savoir-être renouvelé de la personne transformait son savoir-faire au sein de sa pratique. Au plan de la qualité de la présence aux autres, il est rapporté que les relations professionnelles se sont améliorées pour la majorité des participantes. Des transformations personnelles semblent avoir eu des effets sur leurs relations professionnelles. Par exemple, tous témoignent d’une meilleure qualité de présence et de disponibilité aux patientes et patients. La plupart signalent l’apprentissage d’une juste distance thérapeutique et, en même temps, d’une relation plus singulière avec chaque patiente et patient. Nous constatons par notre analyse que la relation aux patientes et patients est modifiée aux plans de la communication, du toucher et de l’écoute. Le parcours de formation des étudiantes et étudiants en médecine semble créer des conditions favorisant l’épuisement (Brazeau, Schroeder, Rovi et Boyd, 2010; Colombat, Altmeyer, Barruel, Bauchetet, Blanchard, Colombat et al., 2011; Ishak, Nikravesh, Lederer, Perry, Ogunyemi et Bernstein, 2013; Llera et Durante, 2014; Rodrigues, Albiges et Blanchard, 2012). Certaines interventions de type corps / esprit semblent pouvoir minimiser cet impact (Elder, Rakel, Heitkemper, Hustedde, Harazduk, Gerik et al., 2007; Hewson, Copeland, Mascha, Arrigain, Topol et Fox, 2006; Irving, Park-Saltzman, Fitzpatrick, Dobkin, Chen et Hutchinson, 2014; Maclaughlin, Wang, Noone, Liu, Harazduk, Lumpkin et al., 2011; Motz, Graves, Gross, Saunders, Amri, Harazduk et al., 2012; Rosenzweig, Reibel, Greeson, Brainard et Hojat, 2003; Saunders, Tractenberg, Chaterji, Amri, Harazduk, Gordon et al., 2007). Notre recherche démontre chez nos participantes et notre participant que la formation aux pratiques du Sensible leur a permis de faire plusieurs gains pour leur propre santé. Il semble qu’en amont des apprentissages liés à la profession médicale, une qualité de savoir-être puisse solidifier la personne, ses apprentissages et sa future pratique médicale. Les étudiantes et étudiants en médecine seraient ainsi mieux outillés pour traverser ce cursus de formation exigent et épuisant. Il serait intéressant de reprendre la recherche auprès d’un plus grand nombre de médecins ou d’étudiantes et étudiants en médecine afin d’y observer les éléments de théorisation répétitifs inclus dans la théorisation ancrée de notre étude exploratoire. Ainsi, selon les résultats, il serait alors plus aisé de promouvoir l’apprentissage expérientiel d’approches de type corps / esprit (dont les PS) dans les cursus universitaires en médecine.