925 resultados para Human visual processing
Resumo:
Abstract Background Despite new brain imaging techniques that have improved the study of the underlying processes of human decision-making, to the best of our knowledge, there have been very few studies that have attempted to investigate brain activity during medical diagnostic processing. We investigated brain electroencephalography (EEG) activity associated with diagnostic decision-making in the realm of veterinary medicine using X-rays as a fundamental auxiliary test. EEG signals were analysed using Principal Components (PCA) and Logistic Regression Analysis Results The principal component analysis revealed three patterns that accounted for 85% of the total variance in the EEG activity recorded while veterinary doctors read a clinical history, examined an X-ray image pertinent to a medical case, and selected among alternative diagnostic hypotheses. Two of these patterns are proposed to be associated with visual processing and the executive control of the task. The other two patterns are proposed to be related to the reasoning process that occurs during diagnostic decision-making. Conclusions PCA analysis was successful in disclosing the different patterns of brain activity associated with hypothesis triggering and handling (pattern P1); identification uncertainty and prevalence assessment (pattern P3), and hypothesis plausibility calculation (pattern P2); Logistic regression analysis was successful in disclosing the brain activity associated with clinical reasoning success, and together with regression analysis showed that clinical practice reorganizes the neural circuits supporting clinical reasoning.
Resumo:
Human brain is provided with a flexible audio-visual system, which interprets and guides responses to external events according to spatial alignment, temporal synchronization and effectiveness of unimodal signals. The aim of the present thesis was to explore the possibility that such a system might represent the neural correlate of sensory compensation after a damage to one sensory pathway. To this purpose, three experimental studies have been conducted, which addressed the immediate, short-term and long-term effects of audio-visual integration on patients with Visual Field Defect (VFD). Experiment 1 investigated whether the integration of stimuli from different modalities (cross-modal) and from the same modality (within-modal) have a different, immediate effect on localization behaviour. Patients had to localize modality-specific stimuli (visual or auditory), cross-modal stimulus pairs (visual-auditory) and within-modal stimulus pairs (visual-visual). Results showed that cross-modal stimuli evoked a greater improvement than within modal stimuli, consistent with a Bayesian explanation. Moreover, even when visual processing was impaired, cross-modal stimuli improved performance in an optimal fashion. These findings support the hypothesis that the improvement derived from multisensory integration is not attributable to simple target redundancy, and prove that optimal integration of cross-modal signals occurs in processing stage which are not consciously accessible. Experiment 2 examined the possibility to induce a short term improvement of localization performance without an explicit knowledge of visual stimulus. Patients with VFD and patients with neglect had to localize weak sounds before and after a brief exposure to a passive cross-modal stimulation, which comprised spatially disparate or spatially coincident audio-visual stimuli. After exposure to spatially disparate stimuli in the affected field, only patients with neglect exhibited a shifts of auditory localization toward the visual attractor (the so called Ventriloquism After-Effect). In contrast, after adaptation to spatially coincident stimuli, both neglect and hemianopic patients exhibited a significant improvement of auditory localization, proving the occurrence of After Effect for multisensory enhancement. These results suggest the presence of two distinct recalibration mechanisms, each mediated by a different neural route: a geniculo-striate circuit and a colliculus-extrastriate circuit respectively. Finally, Experiment 3 verified whether a systematic audio-visual stimulation could exert a long-lasting effect on patients’ oculomotor behaviour. Eye movements responses during a visual search task and a reading task were studied before and after visual (control) or audio-visual (experimental) training, in a group of twelve patients with VFD and twelve controls subjects. Results showed that prior to treatment, patients’ performance was significantly different from that of controls in relation to fixations and saccade parameters; after audiovisual training, all patients reported an improvement in ocular exploration characterized by fewer fixations and refixations, quicker and larger saccades, and reduced scanpath length. Similarly, reading parameters were significantly affected by the training, with respect to specific impairments observed in left and right hemisphere–damaged patients. The present findings provide evidence that a systematic audio-visual stimulation may encourage a more organized pattern of visual exploration with long lasting effects. In conclusion, results from these studies clearly demonstrate that the beneficial effects of audio-visual integration can be retained in absence of explicit processing of visual stimulus. Surprisingly, an improvement of spatial orienting can be obtained not only when a on-line response is required, but also after either a brief or a long adaptation to audio-visual stimulus pairs, so suggesting the maintenance of mechanisms subserving cross-modal perceptual learning after a damage to geniculo-striate pathway. The colliculus-extrastriate pathway, which is spared in patients with VFD, seems to play a pivotal role in this sensory compensation.
Resumo:
Generic object recognition is an important function of the human visual system and everybody finds it highly useful in their everyday life. For an artificial vision system it is a really hard, complex and challenging task because instances of the same object category can generate very different images, depending of different variables such as illumination conditions, the pose of an object, the viewpoint of the camera, partial occlusions, and unrelated background clutter. The purpose of this thesis is to develop a system that is able to classify objects in 2D images based on the context, and identify to which category the object belongs to. Given an image, the system can classify it and decide the correct categorie of the object. Furthermore the objective of this thesis is also to test the performance and the precision of different supervised Machine Learning algorithms in this specific task of object image categorization. Through different experiments the implemented application reveals good categorization performances despite the difficulty of the problem. However this project is open to future improvement; it is possible to implement new algorithms that has not been invented yet or using other techniques to extract features to make the system more reliable. This application can be installed inside an embedded system and after trained (performed outside the system), so it can become able to classify objects in a real-time. The information given from a 3D stereocamera, developed inside the department of Computer Engineering of the University of Bologna, can be used to improve the accuracy of the classification task. The idea is to segment a single object in a scene using the depth given from a stereocamera and in this way make the classification more accurate.
Resumo:
As a more complete picture of the clinical phenotype of Parkinson's disease emerges, non-motor symptoms have become increasingly studied. Prominent among these non-motor phenomena are mood disturbance, cognitive decline and dementia, sleep disorders, hyposmia and autonomic failure. In addition, visual symptoms are common, ranging from complaints of dry eyes and reading difficulties, through to perceptual disturbances (feelings of presence and passage) and complex visual hallucinations. Such visual symptoms are a considerable cause of morbidity in Parkinson's disease and, with respect to visual hallucinations, are an important predictor of cognitive decline as well as institutional care and mortality. Evidence exists of visual dysfunction at several levels of the visual pathway in Parkinson's disease. This includes psychophysical, electrophysiological and morphological evidence of disruption of retinal structure and function, in addition to disorders of ‘higher’ (cortical) visual processing. In this review, we will draw together work from animal and human studies in an attempt to provide an insight into how Parkinson's disease affects the retina and how these changes might contribute to the visual symptoms experienced by patients.
Resumo:
OBJECTIVE To quantify visual discrimination, space-motion, and object-form perception in patients with Parkinson disease dementia (PDD), dementia with Lewy bodies (DLB), and Alzheimer disease (AD). METHODS The authors used a cross-sectional study to compare three demented groups matched for overall dementia severity (PDD: n = 24; DLB: n = 20; AD: n = 23) and two age-, sex-, and education-matched control groups (PD: n = 24, normal controls [NC]: n = 25). RESULTS Visual perception was globally more impaired in PDD than in nondemented controls (NC, PD), but was not different from DLB. Compared to AD, PDD patients tended to perform worse in all perceptual scores. Visual perception of patients with PDD/DLB and visual hallucinations was significantly worse than in patients without hallucinations. CONCLUSIONS Parkinson disease dementia (PDD) is associated with profound visuoperceptual impairments similar to dementia with Lewy bodies (DLB) but different from Alzheimer disease. These findings are consistent with previous neuroimaging studies reporting hypoactivity in cortical areas involved in visual processing in PDD and DLB.
Resumo:
Much of the research on visual hallucinations (VHs) has been conducted in the context of eye disease and neurodegenerative conditions, but little is known about these phenomena in psychiatric and nonclinical populations. The purpose of this article is to bring together current knowledge regarding VHs in the psychosis phenotype and contrast this data with the literature drawn from neurodegenerative disorders and eye disease. The evidence challenges the traditional views that VHs are atypical or uncommon in psychosis. The weighted mean for VHs is 27% in schizophrenia, 15% in affective psychosis, and 7.3% in the general community. VHs are linked to a more severe psychopathological profile and less favorable outcome in psychosis and neurodegenerative conditions. VHs typically co-occur with auditory hallucinations, suggesting a common etiological cause. VHs in psychosis are also remarkably complex, negative in content, and are interpreted to have personal relevance. The cognitive mechanisms of VHs in psychosis have rarely been investigated, but existing studies point to source-monitoring deficits and distortions in top-down mechanisms, although evidence for visual processing deficits, which feature strongly in the organic literature, is lacking. Brain imaging studies point to the activation of visual cortex during hallucinations on a background of structural and connectivity changes within wider brain networks. The relationship between VHs in psychosis, eye disease, and neurodegeneration remains unclear, although the pattern of similarities and differences described in this review suggests that comparative studies may have potentially important clinical and theoretical implications.
Resumo:
Virtual reality (VR) is a powerful tool for simulating aspects of the real world. The success of VR is thought to depend on its ability to evoke a sense of "being there", that is, the feeling of "Presence". In view of the rapid progress in the development of increasingly more sophisticated virtual environments (VE), the importance of understanding the neural underpinnings of presence is growing. To date however, the neural correlates of this phenomenon have received very scant attention. An fMRI-based study with 52 adults and 25 children was therefore conducted using a highly immersive VE. The experience of presence in adult subjects was found to be modulated by two major strategies involving two homologous prefrontal brain structures. Whereas the right DLPFC controlled the sense of presence by down-regulating the activation in the egocentric dorsal visual processing stream, the left DLPFC up-regulated widespread areas of the medial prefrontal cortex known to be involved in self-reflective and stimulus-independent thoughts. In contrast, there was no evidence of these two strategies in children. In fact, anatomical analyses showed that these two prefrontal areas have not yet reached full maturity in children. Taken together, this study presents the first findings that show activation of a highly specific neural network orchestrating the experience of presence in adult subjects, and that the absence of activity in this neural network might contribute to the generally increased susceptibility of children for the experience of presence in VEs.
Resumo:
La medida de calidad de vídeo sigue siendo necesaria para definir los criterios que caracterizan una señal que cumpla los requisitos de visionado impuestos por el usuario. Las nuevas tecnologías, como el vídeo 3D estereoscópico o formatos más allá de la alta definición, imponen nuevos criterios que deben ser analizadas para obtener la mayor satisfacción posible del usuario. Entre los problemas detectados durante el desarrollo de esta tesis doctoral se han determinado fenómenos que afectan a distintas fases de la cadena de producción audiovisual y tipo de contenido variado. En primer lugar, el proceso de generación de contenidos debe encontrarse controlado mediante parámetros que eviten que se produzca el disconfort visual y, consecuentemente, fatiga visual, especialmente en lo relativo a contenidos de 3D estereoscópico, tanto de animación como de acción real. Por otro lado, la medida de calidad relativa a la fase de compresión de vídeo emplea métricas que en ocasiones no se encuentran adaptadas a la percepción del usuario. El empleo de modelos psicovisuales y diagramas de atención visual permitirían ponderar las áreas de la imagen de manera que se preste mayor importancia a los píxeles que el usuario enfocará con mayor probabilidad. Estos dos bloques se relacionan a través de la definición del término saliencia. Saliencia es la capacidad del sistema visual para caracterizar una imagen visualizada ponderando las áreas que más atractivas resultan al ojo humano. La saliencia en generación de contenidos estereoscópicos se refiere principalmente a la profundidad simulada mediante la ilusión óptica, medida en términos de distancia del objeto virtual al ojo humano. Sin embargo, en vídeo bidimensional, la saliencia no se basa en la profundidad, sino en otros elementos adicionales, como el movimiento, el nivel de detalle, la posición de los píxeles o la aparición de caras, que serán los factores básicos que compondrán el modelo de atención visual desarrollado. Con el objetivo de detectar las características de una secuencia de vídeo estereoscópico que, con mayor probabilidad, pueden generar disconfort visual, se consultó la extensa literatura relativa a este tema y se realizaron unas pruebas subjetivas preliminares con usuarios. De esta forma, se llegó a la conclusión de que se producía disconfort en los casos en que se producía un cambio abrupto en la distribución de profundidades simuladas de la imagen, aparte de otras degradaciones como la denominada “violación de ventana”. A través de nuevas pruebas subjetivas centradas en analizar estos efectos con diferentes distribuciones de profundidades, se trataron de concretar los parámetros que definían esta imagen. Los resultados de las pruebas demuestran que los cambios abruptos en imágenes se producen en entornos con movimientos y disparidades negativas elevadas que producen interferencias en los procesos de acomodación y vergencia del ojo humano, así como una necesidad en el aumento de los tiempos de enfoque del cristalino. En la mejora de las métricas de calidad a través de modelos que se adaptan al sistema visual humano, se realizaron también pruebas subjetivas que ayudaron a determinar la importancia de cada uno de los factores a la hora de enmascarar una determinada degradación. Los resultados demuestran una ligera mejora en los resultados obtenidos al aplicar máscaras de ponderación y atención visual, los cuales aproximan los parámetros de calidad objetiva a la respuesta del ojo humano. ABSTRACT Video quality assessment is still a necessary tool for defining the criteria to characterize a signal with the viewing requirements imposed by the final user. New technologies, such as 3D stereoscopic video and formats of HD and beyond HD oblige to develop new analysis of video features for obtaining the highest user’s satisfaction. Among the problems detected during the process of this doctoral thesis, it has been determined that some phenomena affect to different phases in the audiovisual production chain, apart from the type of content. On first instance, the generation of contents process should be enough controlled through parameters that avoid the occurrence of visual discomfort in observer’s eye, and consequently, visual fatigue. It is especially necessary controlling sequences of stereoscopic 3D, with both animation and live-action contents. On the other hand, video quality assessment, related to compression processes, should be improved because some objective metrics are adapted to user’s perception. The use of psychovisual models and visual attention diagrams allow the weighting of image regions of interest, giving more importance to the areas which the user will focus most probably. These two work fields are related together through the definition of the term saliency. Saliency is the capacity of human visual system for characterizing an image, highlighting the areas which result more attractive to the human eye. Saliency in generation of 3DTV contents refers mainly to the simulated depth of the optic illusion, i.e. the distance from the virtual object to the human eye. On the other hand, saliency is not based on virtual depth, but on other features, such as motion, level of detail, position of pixels in the frame or face detection, which are the basic features that are part of the developed visual attention model, as demonstrated with tests. Extensive literature involving visual comfort assessment was looked up, and the development of new preliminary subjective assessment with users was performed, in order to detect the features that increase the probability of discomfort to occur. With this methodology, the conclusions drawn confirmed that one common source of visual discomfort was when an abrupt change of disparity happened in video transitions, apart from other degradations, such as window violation. New quality assessment was performed to quantify the distribution of disparities over different sequences. The results confirmed that abrupt changes in negative parallax environment produce accommodation-vergence mismatches derived from the increasing time for human crystalline to focus the virtual objects. On the other side, for developing metrics that adapt to human visual system, additional subjective tests were developed to determine the importance of each factor, which masks a concrete distortion. Results demonstrated slight improvement after applying visual attention to objective metrics. This process of weighing pixels approximates the quality results to human eye’s response.
Resumo:
La gran cantidad de datos que se registran diariamente en los sistemas de base de datos de las organizaciones ha generado la necesidad de analizarla. Sin embargo, se enfrentan a la complejidad de procesar enormes volúmenes de datos a través de métodos tradicionales de análisis. Además, dentro de un contexto globalizado y competitivo las organizaciones se mantienen en la búsqueda constante de mejorar sus procesos, para lo cual requieren herramientas que les permitan tomar mejores decisiones. Esto implica estar mejor informado y conocer su historia digital para describir sus procesos y poder anticipar (predecir) eventos no previstos. Estos nuevos requerimientos de análisis de datos ha motivado el desarrollo creciente de proyectos de minería de datos. El proceso de minería de datos busca obtener desde un conjunto masivo de datos, modelos que permitan describir los datos o predecir nuevas instancias en el conjunto. Implica etapas de: preparación de los datos, procesamiento parcial o totalmente automatizado para identificar modelos en los datos, para luego obtener como salida patrones, relaciones o reglas. Esta salida debe significar un nuevo conocimiento para la organización, útil y comprensible para los usuarios finales, y que pueda ser integrado a los procesos para apoyar la toma de decisiones. Sin embargo, la mayor dificultad es justamente lograr que el analista de datos, que interviene en todo este proceso, pueda identificar modelos lo cual es una tarea compleja y muchas veces requiere de la experiencia, no sólo del analista de datos, sino que también del experto en el dominio del problema. Una forma de apoyar el análisis de datos, modelos y patrones es a través de su representación visual, utilizando las capacidades de percepción visual del ser humano, la cual puede detectar patrones con mayor facilidad. Bajo este enfoque, la visualización ha sido utilizada en minería datos, mayormente en el análisis descriptivo de los datos (entrada) y en la presentación de los patrones (salida), dejando limitado este paradigma para el análisis de modelos. El presente documento describe el desarrollo de la Tesis Doctoral denominada “Nuevos Esquemas de Visualizaciones para Mejorar la Comprensibilidad de Modelos de Data Mining”. Esta investigación busca aportar con un enfoque de visualización para apoyar la comprensión de modelos minería de datos, para esto propone la metáfora de modelos visualmente aumentados. ABSTRACT The large amount of data to be recorded daily in the systems database of organizations has generated the need to analyze it. However, faced with the complexity of processing huge volumes of data over traditional methods of analysis. Moreover, in a globalized and competitive environment organizations are kept constantly looking to improve their processes, which require tools that allow them to make better decisions. This involves being bettered informed and knows your digital story to describe its processes and to anticipate (predict) unanticipated events. These new requirements of data analysis, has led to the increasing development of data-mining projects. The data-mining process seeks to obtain from a massive data set, models to describe the data or predict new instances in the set. It involves steps of data preparation, partially or fully automated processing to identify patterns in the data, and then get output patterns, relationships or rules. This output must mean new knowledge for the organization, useful and understandable for end users, and can be integrated into the process to support decision-making. However, the biggest challenge is just getting the data analyst involved in this process, which can identify models is complex and often requires experience not only of the data analyst, but also the expert in the problem domain. One way to support the analysis of the data, models and patterns, is through its visual representation, i.e., using the capabilities of human visual perception, which can detect patterns easily in any context. Under this approach, the visualization has been used in data mining, mostly in exploratory data analysis (input) and the presentation of the patterns (output), leaving limited this paradigm for analyzing models. This document describes the development of the doctoral thesis entitled "New Visualizations Schemes to Improve Understandability of Data-Mining Models". This research aims to provide a visualization approach to support understanding of data mining models for this proposed metaphor visually enhanced models.
Resumo:
The computations involved in the processing of a visual scene invariably involve the interactions among neurons throughout all of visual cortex. One hypothesis is that the timing of neuronal activity, as well as the amplitude of activity, provides a means to encode features of objects. The experimental data from studies on cat [Gray, C. M., Konig, P., Engel, A. K. & Singer, W. (1989) Nature (London) 338, 334–337] support a view in which only synchronous (no phase lags) activity carries information about the visual scene. In contrast, theoretical studies suggest, on the one hand, the utility of multiple phases within a population of neurons as a means to encode independent visual features and, on the other hand, the likely existence of timing differences solely on the basis of network dynamics. Here we use widefield imaging in conjunction with voltage-sensitive dyes to record electrical activity from the virtually intact, unanesthetized turtle brain. Our data consist of single-trial measurements. We analyze our data in the frequency domain to isolate coherent events that lie in different frequency bands. Low frequency oscillations (<5 Hz) are seen in both ongoing activity and activity induced by visual stimuli. These oscillations propagate parallel to the afferent input. Higher frequency activity, with spectral peaks near 10 and 20 Hz, is seen solely in response to stimulation. This activity consists of plane waves and spiral-like waves, as well as more complex patterns. The plane waves have an average phase gradient of ≈π/2 radians/mm and propagate orthogonally to the low frequency waves. Our results show that large-scale differences in neuronal timing are present and persistent during visual processing.
Resumo:
In subjects suffering from early onset strabismus, signals conveyed by the two eyes are not perceived simultaneously but in alternation. We exploited this phenomenon of interocular suppression to investigate the neuronal correlate of binocular rivalry in primary visual cortex of awake strabismic cats. Monocularly presented stimuli that were readily perceived by the animal evoked synchronized discharges with an oscillatory patterning in the γ-frequency range. Upon dichoptic stimulation, neurons responding to the stimulus that continued to be perceived increased the synchronicity and the regularity of their oscillatory patterning while the reverse was true for neurons responding to the stimulus that was no longer perceived. These differential changes were not associated with modifications of discharge rate, suggesting that at early stages of visual processing the degree of synchronicity rather than the amplitude of responses determines which signals are perceived and control behavioral responses.
Resumo:
Expression of G protein-regulated phospholipase C (PLC) β4 in the retina, lateral geniculate nucleus, and superior colliculus implies that PLC β4 may play a role in the mammalian visual process. A mouse line that lacks PLC β4 was generated and the physiological significance of PLC β4 in murine visual function was investigated. Behavioral tests using a shuttle box demonstrated that the mice lacking PLC β4 were impaired in their visual processing abilities, whereas they showed no deficit in their auditory abilities. In addition, the PLC β4-null mice showed 4-fold reduction in the maximal amplitude of the rod a- and b-wave components of their electroretinograms relative to their littermate controls. However, recording from single rod photoreceptors did not reveal any significant differences between the PLC β4-null and wild-type littermates, nor were there any apparent differences in retinas examined with light microscopy. While the behavioral and electroretinographic results indicate that PLC β4 plays a significant role in mammalian visual signal processing, isolated rod recording shows little or no apparent deficit, suggesting that the effect of PLC β4 deficiency on the rod signaling pathway occurs at some stage after the initial phototransduction cascade and may require cell–cell interactions between rods and other retinal cells.
Resumo:
Syntax denotes a rule system that allows one to predict the sequencing of communication signals. Despite its significance for both human speech processing and animal acoustic communication, the representation of syntactic structure in the mammalian brain has not been studied electrophysiologically at the single-unit level. In the search for a neuronal correlate for syntax, we used playback of natural and temporally destructured complex species-specific communication calls—so-called composites—while recording extracellularly from neurons in a physiologically well defined area (the FM–FM area) of the mustached bat’s auditory cortex. Even though this area is known to be involved in the processing of target distance information for echolocation, we found that units in the FM–FM area were highly responsive to composites. The finding that neuronal responses were strongly affected by manipulation in the time domain of the natural composite structure lends support to the hypothesis that syntax processing in mammals occurs at least at the level of the nonprimary auditory cortex.
Resumo:
The human cerebral cortex is notorious for the depth and irregularity of its convolutions and for its variability from one individual to the next. These complexities of cortical geography have been a chronic impediment to studies of functional specialization in the cortex. In this report, we discuss ways to compensate for the convolutions by using a combination of strategies whose common denominator involves explicit reconstructions of the cortical surface. Surface-based visualization involves reconstructing cortical surfaces and displaying them, along with associated experimental data, in various complementary formats (including three-dimensional native configurations, two-dimensional slices, extensively smoothed surfaces, ellipsoidal representations, and cortical flat maps). Generating these representations for the cortex of the Visible Man leads to a surface-based atlas that has important advantages over conventional stereotaxic atlases as a substrate for displaying and analyzing large amounts of experimental data. We illustrate this by showing the relationship between functionally specialized regions and topographically organized areas in human visual cortex. Surface-based warping allows data to be mapped from individual hemispheres to a surface-based atlas while respecting surface topology, improving registration of identifiable landmarks, and minimizing unwanted distortions. Surface-based warping also can aid in comparisons between species, which we illustrate by warping a macaque flat map to match the shape of a human flat map. Collectively, these approaches will allow more refined analyses of commonalities as well as individual differences in the functional organization of primate cerebral cortex.
Resumo:
Functional anatomical and single-unit recording studies indicate that a set of neural signals in parietal and frontal cortex mediates the covert allocation of attention to visual locations, as originally proposed by psychological studies. This frontoparietal network is the source of a location bias that interacts with extrastriate regions of the ventral visual system during object analysis to enhance visual processing. The frontoparietal network is not exclusively related to visual attention, but may coincide or overlap with regions involved in oculomotor processing. The relationship between attention and eye movement processes is discussed at the psychological, functional anatomical, and cellular level of analysis.