985 resultados para visual objects
Resumo:
Human observers exhibit large systematic distance-dependent biases when estimating the three-dimensional (3D) shape of objects defined by binocular image disparities. This has led some to question the utility of disparity as a cue to 3D shape and whether accurate estimation of 3D shape is at all possible. Others have argued that accurate perception is possible, but only with large continuous perspective transformations of an object. Using a stimulus that is known to elicit large distance-dependent perceptual bias (random dot stereograms of elliptical cylinders) we show that contrary to these findings the simple adoption of a more naturalistic viewing angle completely eliminates this bias. Using behavioural psychophysics, coupled with a novel surface-based reverse correlation methodology, we show that it is binocular edge and contour information that allows for accurate and precise perception and that observers actively exploit and sample this information when it is available.
Resumo:
Observers generally fail to recover three-dimensional shape accurately from binocular disparity. Typically, depth is overestimated at near distances and underestimated at far distances [Johnston, E. B. (1991). Systematic distortions of shape from stereopsis. Vision Research, 31, 1351–1360]. A simple prediction from this is that disparity-defined objects should appear to expand in depth when moving towards the observer, and compress in depth when moving away. However, additional information is provided when an object moves from which 3D Euclidean shape can be recovered, be this through the addition of structure from motion information [Richards, W. (1985). Structure from stereo and motion. Journal of the Optical Society of America A, 2, 343–349], or the use of non-generic strategies [Todd, J. T., & Norman, J. F. (2003). The visual perception of 3-D shape from multiple cues: Are observers capable of perceiving metric structure? Perception and Psychophysics, 65, 31–47]. Here, we investigated shape constancy for objects moving in depth. We found that to be perceived as constant in shape, objects needed to contract in depth when moving toward the observer, and expand in depth when moving away, countering the effects of incorrect distance scaling (Johnston, 1991). This is a striking example of the failure of shape con- stancy, but one that is predicted if observers neither accurately estimate object distance in order to recover Euclidean shape, nor are able to base their responses on a simpler processing strategy.
Resumo:
Biological systems have facility to capture salient object(s) in a given scene, but it is still a difficult task to be accomplished by artificial vision systems. In this paper a visual selection mechanism based on the integrate and fire neural network is proposed. The model not only can discriminate objects in a given visual scene, but also can deliver focus of attention to the salient object. Moreover, it processes a combination of relevant features of an input scene, such as intensity, color, orientation, and the contrast of them. In comparison to other visual selection approaches, this model presents several interesting features. It is able to capture attention of objects in complex forms, including those linearly nonseparable. Moreover, computer simulations show that the model produces results similar to those observed in natural vision systems.
Resumo:
The discrimination learning is assessed through instrumental tasks in which the individual is rewarded for choosing one item over another. Thus, in concurrent visual discrimination of objects the animal must learn that only one of the objects will be rewarded. The concurrent visual discrimination is relatively simple, and already been observed Callithrix jacchus is able to accomplish this task. As yet wasn't seen the influence of the qualitative aspects of the rewards, in the performance of concurrent visual discrimination of objects in nonhuman primates, and as in most tests are used isolated animals, the present study had two stages: at first we had as objective to analyze the influence of the caloric value of the reward on the performance in concurrent visual discrimination of objects in isolated animals; in the second, we had the intention analyze performance of C. jacchus in realization of discrimination task in different social contexts, as well as, analyze the influence of previous experience in task performance. In the first stage (Study 1), the animals were not able to discriminate foods that presented small caloric differences . This incapacity in discriminates the rewards was responsible by generating randomness in task of concurrent visual discrimination of objects. In the second stage (Study 2), observed that, independent of social context in which the task was presented, the performance both of the experienced animals as the inexperienced animals tended to randomness. In the first case, is likely that the pattern of responses of the experienced animals is a reflection of their own performance when they were observed in isolation. In the second case, in turn, the randomness was probably due to the small number of sessions. Although present a pattern of performance similar to inexperienced individuals, we verify that the experienced animals monopolize the food consumption when they were in the presence of inexperienced individuals. This was a consequence of the experienced animals have presented lower latency the approximation of apparatus and, consequently, obtain more food. In turn, the inexperienced animals, when were in the presence of experienced, had to adopt alternative strategies to obtain food. Thus, C. jacchus is able to use the previous information he had about the task of solving their own benefit.
Resumo:
Synchronous generators are essential components of electric power systems. They are present both in hydro and thermal power plants, performing the function of converting mechanical into electrical energy. This paper presents a visual approach to manipulate parameters that affect operation limits of synchronous generators, using a specifically designed software. The operating characteristics of synchronous generators, for all possible modes of operation, are revised in order to link the concepts to the graphic objects. The approach matches the distance learning tool requirements and also enriches the learning process by developing student trust and understanding of the concepts involved in building synchronous machine capability curves. © 2012 IEEE.
Resumo:
Pós-graduação em Educação - FCT
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Generic object recognition is an important function of the human visual system and everybody finds it highly useful in their everyday life. For an artificial vision system it is a really hard, complex and challenging task because instances of the same object category can generate very different images, depending of different variables such as illumination conditions, the pose of an object, the viewpoint of the camera, partial occlusions, and unrelated background clutter. The purpose of this thesis is to develop a system that is able to classify objects in 2D images based on the context, and identify to which category the object belongs to. Given an image, the system can classify it and decide the correct categorie of the object. Furthermore the objective of this thesis is also to test the performance and the precision of different supervised Machine Learning algorithms in this specific task of object image categorization. Through different experiments the implemented application reveals good categorization performances despite the difficulty of the problem. However this project is open to future improvement; it is possible to implement new algorithms that has not been invented yet or using other techniques to extract features to make the system more reliable. This application can be installed inside an embedded system and after trained (performed outside the system), so it can become able to classify objects in a real-time. The information given from a 3D stereocamera, developed inside the department of Computer Engineering of the University of Bologna, can be used to improve the accuracy of the classification task. The idea is to segment a single object in a scene using the depth given from a stereocamera and in this way make the classification more accurate.
Resumo:
Edges are crucial for the formation of coherent objects from sequential sensory inputs within a single modality. Moreover, temporally coincident boundaries of perceptual objects across different sensory modalities facilitate crossmodal integration. Here, we used functional magnetic resonance imaging in order to examine the neural basis of temporal edge detection across modalities. Onsets of sensory inputs are not only related to the detection of an edge but also to the processing of novel sensory inputs. Thus, we used transitions from input to rest (offsets) as convenient stimuli for studying the neural underpinnings of visual and acoustic edge detection per se. We found, besides modality-specific patterns, shared visual and auditory offset-related activity in the superior temporal sulcus and insula of the right hemisphere. Our data suggest that right hemispheric regions known to be involved in multisensory processing are crucial for detection of edges in the temporal domain across both visual and auditory modalities. This operation is likely to facilitate cross-modal object feature binding based on temporal coincidence. Hum Brain Mapp, 2008. (c) 2008 Wiley-Liss, Inc.
Resumo:
The aging population has become a burning issue for all modern societies around the world recently. There are two important issues existing now to be solved. One is how to continuously monitor the movements of those people having suffered a stroke in natural living environment for providing more valuable feedback to guide clinical interventions. The other one is how to guide those old people effectively when they are at home or inside other buildings and to make their life easier and convenient. Therefore, human motion tracking and navigation have been active research fields with the increasing number of elderly people. However, motion capture has been extremely challenging to go beyond laboratory environments and obtain accurate measurements of human physical activity especially in free-living environments, and navigation in free-living environments also poses some problems such as the denied GPS signal and the moving objects commonly presented in free-living environments. This thesis seeks to develop new technologies to enable accurate motion tracking and positioning in free-living environments. This thesis comprises three specific goals using our developed IMU board and the camera from the imaging source company: (1) to develop a robust and real-time orientation algorithm using only the measurements from IMU; (2) to develop a robust distance estimation in static free-living environments to estimate people’s position and navigate people in static free-living environments and simultaneously the scale ambiguity problem, usually appearing in the monocular camera tracking, is solved by integrating the data from the visual and inertial sensors; (3) in case of moving objects viewed by the camera existing in free-living environments, to firstly design a robust scene segmentation algorithm and then respectively estimate the motion of the vIMU system and moving objects. To achieve real-time orientation tracking, an Adaptive-Gain Orientation Filter (AGOF) is proposed in this thesis based on the basic theory of deterministic approach and frequency-based approach using only measurements from the newly developed MARG (Magnet, Angular Rate, and Gravity) sensors. To further obtain robust positioning, an adaptive frame-rate vision-aided IMU system is proposed to develop and implement fast vIMU ego-motion estimation algorithms, where the orientation is estimated in real time from MARG sensors in the first step and then used to estimate the position based on the data from visual and inertial sensors. In case of the moving objects viewed by the camera existing in free-living environments, a robust scene segmentation algorithm is firstly proposed to obtain position estimation and simultaneously the 3D motion of moving objects. Finally, corresponding simulations and experiments have been carried out.
Resumo:
Visual fixation is employed by humans and some animals to keep a specific 3D location at the center of the visual gaze. Inspired by this phenomenon in nature, this paper explores the idea to transfer this mechanism to the context of video stabilization for a handheld video camera. A novel approach is presented that stabilizes a video by fixating on automatically extracted 3D target points. This approach is different from existing automatic solutions that stabilize the video by smoothing. To determine the 3D target points, the recorded scene is analyzed with a stateof- the-art structure-from-motion algorithm, which estimates camera motion and reconstructs a 3D point cloud of the static scene objects. Special algorithms are presented that search either virtual or real 3D target points, which back-project close to the center of the image for as long a period of time as possible. The stabilization algorithm then transforms the original images of the sequence so that these 3D target points are kept exactly in the center of the image, which, in case of real 3D target points, produces a perfectly stable result at the image center. Furthermore, different methods of additional user interaction are investigated. It is shown that the stabilization process can easily be controlled and that it can be combined with state-of-theart tracking techniques in order to obtain a powerful image stabilization tool. The approach is evaluated on a variety of videos taken with a hand-held camera in natural scenes.
Resumo:
Background Visual symptoms are common in Parkinson's disease (PD) and are frequently under-diagnosed. The detection of visual symptoms is important for differential diagnosis and patient management. Aim To establish the prevalence of recurrent visual complaints (RVC) and recurrent visual hallucinations (RVH) and to investigate their interaction in PD patients and controls. Methods This cross-sectional study included 88 PD patients and 90 controls. RVC and RVH were assessed with a visual symptom questionnaire and the North-East-Visual-Hallucinations-Interview (NEVHI). Results Double vision (PD vs. Controls: 18.2% vs. 1.3%; p < 0.001), misjudging objects when walking (PD vs. Controls: 12.5% vs. 1.3%; p < 0.01), words moving whilst reading (PD vs. Controls: 17.0% vs. 1.3%; p < 0.001) and freezing in narrow spaces (PD vs. Controls: 30.7% vs. 0%; p < 0.001) were almost exclusively found in PD patients. The same was true for recurrent complex visual hallucinations and illusions (PD vs. Controls: both 17.0% vs. 0%; p < 0.001). Multiple RVC (43.2% vs. 15.8%) and multiple RVH (29.5% vs. 5.6%) were also more common in PD patients (both p < 0.001). RVC did not predict recurrent complex visual hallucinations; but double vision (p = 0.018, R2 = 0.302) and misjudging objects (p = 0.002, R2 = 0.302) predicted passage hallucinations. Misjudging objects also predicted the feeling of presence (p = 0.010, R2 = 0.321). Conclusions Multiple and recurrent visual symptoms are common in PD. RVC emerged as risk factors predictive of the minor forms of hallucinations, but not recurrent complex visual hallucinations.
Resumo:
The macaque cortical visual system is hierarchically organized into two streams, the ventral stream for recognizing objects and the dorsal stream for analyzing spatial relationships. The ventral stream extends from striate cortex or area V1 to inferior temporal cortex (IT) through extra-striate areas V2 and V4. Between V1 and V2, the ventral stream consists of two roughly parallel sub-streams, one extending from the cytochrome oxidase (CO) rich blobs in V1 to the CO rich thin stripes in V2, the other extending from the interblobs in V1 to interstripes, in V2. The blob-dominated sub-stream is thought to analyze the surface features such as color, whereas the interblob-dominated one is thought to analyze the contour features such as shape. ^ In the current study, the organization of cortical pathways linking V2 thin stripe and interstripe compartments with area V4 was investigated using a combination of physiological and anatomical techniques. Different compartments of V2 were first characterized, in vivo, using optical recording of intrinsic cortical signals. These functionally derived maps of V2 stripe compartments were then used to guide iontophoretic injections of multiple, distinguishable, anterograde tracers into specific V2 compartments. The distribution of labeled axons was analyzed either in horizontal sections through the prelunate gyrus, or in tangentially sectioned portions of physically unfolded cortex containing the lunate sulcus, prelunate gyrus and superior temporal sulcus. When a V2 thin stripe and adjacent interstripe were injected with distinguishable tracers, a large primary and several secondary foci were observed in V4. The primary focus from the thin stripe injection was spatially segregated from the primary focus from the V2 interstripe injection, suggesting a retention of the pattern of compartmentation. ^ We examined the distribution of retrogradely labeled cells in V1 following the injections of tracers into V2 different compartments, in order to quantitate just how parallel the two sub-streams are from V1 to V2. Our results suggest that both blobs and interblobs project to thin stripes in V2, whereas only interblobs project to interstripes. This asymmetrical segregation argues against the original proposal of strict parallelism. (Abstract shortened by UMI.) ^