946 resultados para VISUAL INFORMATION
Resumo:
We investigated whether attention shifts and eye movement preparation are mediated by shared control mechanisms, as claimed by the premotor theory of attention. ERPs were recorded in three tasks where directional cues presented at the beginning of each trial instructed participants to direct their attention to the cued side without eye movements (Covert task), to prepare an eye movement in the cued direction without attention shifts (Saccade task) or both (Combined task). A peripheral visual Go/Nogo stimulus that was presented 800 ms after cue onset signalled whether responses had to be executed or withheld. Lateralised ERP components triggered during the cue–target interval, which are assumed to reflect preparatory control mechanisms that mediate attentional orienting, were very similar across tasks. They were also present in the Saccade task, which was designed to discourage any concomitant covert attention shifts. These results support the hypothesis that saccade preparation and attentional orienting are implemented by common control structures. There were however systematic differences in the impact of eye movement programming and covert attention on ERPs triggered in response to visual stimuli at cued versus uncued locations. It is concluded that, although the preparatory processes underlying saccade programming and covert attentional orienting may be based on common mechanisms, they nevertheless differ in their spatially specific effects on visual information processing.
Resumo:
The Virtual Lightbox for Museums and Archives (VLMA) is a tool for collecting and reusing, in a structured fashion, the online contents of museums and archive datasets. It is not restricted to datasets with visual components although VLMA includes a lightbox service that enables comparison and manipulation of visual information. With VLMA, one can browse and search collections, construct personal collections, annotate them, export these collections to XML or Impress (Open Office) presentation format, and share collections with other VLMA users. VLMA was piloted as an e-Learning tool as part of JISC’s e-Learning focus in its first phase (2004-2005) and in its second phase (2005-2006) it has incorporated new partner collections while improving and expanding interfaces and services. This paper concerns its development as a research and teaching tool, especially to teachers using museum collections, and discusses the recent development of VLMA.
Resumo:
The existence of hand-centred visual processing has long been established in the macaque premotor cortex. These hand-centred mechanisms have been thought to play some general role in the sensory guidance of movements towards objects, or, more recently, in the sensory guidance of object avoidance movements. We suggest that these hand-centred mechanisms play a specific and prominent role in the rapid selection and control of manual actions following sudden changes in the properties of the objects relevant for hand-object interactions. We discuss recent anatomical and physiological evidence from human and non-human primates, which indicates the existence of rapid processing of visual information for hand-object interactions. This new evidence demonstrates how several stages of the hierarchical visual processing system may be bypassed, feeding the motor system with hand-related visual inputs within just 70 ms following a sudden event. This time window is early enough, and this processing rapid enough, to allow the generation and control of rapid hand-centred avoidance and acquisitive actions, for aversive and desired objects, respectively
Resumo:
Analysis of human behaviour through visual information has been a highly active research topic in the computer vision community. This was previously achieved via images from a conventional camera, but recently depth sensors have made a new type of data available. This survey starts by explaining the advantages of depth imagery, then describes the new sensors that are available to obtain it. In particular, the Microsoft Kinect has made high-resolution real-time depth cheaply available. The main published research on the use of depth imagery for analysing human activity is reviewed. Much of the existing work focuses on body part detection and pose estimation. A growing research area addresses the recognition of human actions. The publicly available datasets that include depth imagery are listed, as are the software libraries that can acquire it from a sensor. This survey concludes by summarising the current state of work on this topic, and pointing out promising future research directions.
Resumo:
In the present study, to shed light on a role of positional error correction mechanism and prediction mechanism in the proactive control discovered earlier, we carried out a visual tracking experiment, in which the region where target was shown, was regulated in a circular orbit. Main results found in this research were following. Recognition of a time step, obtained from the environmental stimuli, is required for the predictive function. The period of the rhythm in the brain obtained from environmental stimuli is shortened about 10%, when the visual information is cut-off. The shortening of the period of the rhythm in the brain accelerates the motion as soon as the visual information is cut-off, and lets the hand motion precedes the target motion. Although the precedence of the hand in the blind region is reset by the environmental information when the target enters the visible region, the hand precedes in average the target when the predictive mechanism dominates the error-corrective mechanism.
Resumo:
For many tasks, such as retrieving a previously viewed object, an observer must form a representation of the world at one location and use it at another. A world-based 3D reconstruction of the scene built up from visual information would fulfil this requirement, something computer vision now achieves with great speed and accuracy. However, I argue that it is neither easy nor necessary for the brain to do this. I discuss biologically plausible alternatives, including the possibility of avoiding 3D coordinate frames such as ego-centric and world-based representations. For example, the distance, slant and local shape of surfaces dictate the propensity of visual features to move in the image with respect to one another as the observer’s perspective changes (through movement or binocular viewing). Such propensities can be stored without the need for 3D reference frames. The problem of representing a stable scene in the face of continual head and eye movements is an appropriate starting place for understanding the goal of 3D vision, more so, I argue, than the case of a static binocular observer.
Resumo:
Navigation is a broad topic that has been receiving considerable attention from the mobile robotic community over the years. In order to execute autonomous driving in outdoor urban environments it is necessary to identify parts of the terrain that can be traversed and parts that should be avoided. This paper describes an analyses of terrain identification based on different visual information using a MLP artificial neural network and combining responses of many classifiers. Experimental tests using a vehicle and a video camera have been conducted in real scenarios to evaluate the proposed approach.
Resumo:
A forum is a valuable tool to foster reflection in an in-depth discussion; however, it forces the course mediator to continually pay close attention in order to coordinate learners` activities. Moreover, monitoring a forum is time consuming given that it is impossible to know in advance when new messages are going to be posted. Additionally, a forum may be inactive for a long period and suddenly receive a burst of messages forcing forum mediators to frequently log on in order to know how the discussion is unfolding to intervene whenever it is necessary. Mediators also need to deal with a large amount of messages to identify off-pattern situations. This work presents a piece of action research that investigates how to improve coordination support in a forum using mobile devices for mitigating mediator`s difficulties in following the status of a forum. Based on summarized information extracted from message meta-data, mediators consult visual information summaries on PDAs and receive textual notifications in their mobile phone. This investigation revealed that mediators used the mobile-based coordination support to keep informed on what is taking place within the forum without the need to log on their desktop computer. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
Automated virtual camera control has been widely used in animation and interactive virtual environments. We have developed a multiple sparse camera based free view video system prototype that allows users to control the position and orientation of a virtual camera, enabling the observation of a real scene in three dimensions (3D) from any desired viewpoint. Automatic camera control can be activated to follow selected objects by the user. Our method combines a simple geometric model of the scene composed of planes (virtual environment), augmented with visual information from the cameras and pre-computed tracking information of moving targets to generate novel perspective corrected 3D views of the virtual camera and moving objects. To achieve real-time rendering performance, view-dependent textured mapped billboards are used to render the moving objects at their correct locations and foreground masks are used to remove the moving objects from the projected video streams. The current prototype runs on a PC with a common graphics card and can generate virtual 2D views from three cameras of resolution 768 x 576 with several moving objects at about 11 fps. (C)2011 Elsevier Ltd. All rights reserved.
Resumo:
This work introduces a new method for environment mapping with three-dimensional information from visual information for robotic accurate navigation. Many approaches of 3D mapping using occupancy grid typically requires high computacional effort to both build and store the map. We introduce an 2.5-D occupancy-elevation grid mapping, which is a discrete mapping approach, where each cell stores the occupancy probability, the height of the terrain at current place in the environment and the variance of this height. This 2.5-dimensional representation allows that a mobile robot to know whether a place in the environment is occupied by an obstacle and the height of this obstacle, thus, it can decide if is possible to traverse the obstacle. Sensorial informations necessary to construct the map is provided by a stereo vision system, which has been modeled with a robust probabilistic approach, considering the noise present in the stereo processing. The resulting maps favors the execution of tasks like decision making in the autonomous navigation, exploration, localization and path planning. Experiments carried out with a real mobile robots demonstrates that this proposed approach yields useful maps for robot autonomous navigation
Resumo:
We propose a new approach to reduction and abstraction of visual information for robotics vision applications. Basically, we propose to use a multi-resolution representation in combination with a moving fovea for reducing the amount of information from an image. We introduce the mathematical formalization of the moving fovea approach and mapping functions that help to use this model. Two indexes (resolution and cost) are proposed that can be useful to choose the proposed model variables. With this new theoretical approach, it is possible to apply several filters, to calculate disparity and to obtain motion analysis in real time (less than 33ms to process an image pair at a notebook AMD Turion Dual Core 2GHz). As the main result, most of time, the moving fovea allows the robot not to perform physical motion of its robotics devices to keep a possible region of interest visible in both images. We validate the proposed model with experimental results
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
Processing in the visual system starts in the retina. Its complex network of cells with different properties enables for parallel encoding and transmission of visual information to the lateral geniculate nucleus (LGN) and to the cortex. In the retina, it has been shown that responses are often accompanied by fast synchronous oscillations (30 - 90 Hz) in a stimulus-dependent manner. Studies in the frog, rabbit, cat and monkey, have shown strong oscillatory responses to large stimuli which probably encode global stimulus properties, such as size and continuity (Neuenschwander and Singer, 1996; Ishikane et al., 2005). Moreover, simultaneous recordings from different levels in the visual system have demonstrated that the oscillatory patterning of retinal ganglion cell responses are transmitted to the cortex via the LGN (Castelo-Branco et al., 1998). Overall these results suggest that feedforward synchronous oscillations contribute to visual encoding. In the present study on the LGN of the anesthetized cat, we further investigate the role of retinal oscillations in visual processing by applying complex stimuli, such as natural visual scenes, light spots of varying size and contrast, and flickering checkerboards. This is a necessary step for understanding encoding mechanisms in more naturalistic conditions, as currently most data on retinal oscillations have been limited to simple, flashed and stationary stimuli. Correlation analysis of spiking responses confirmed previous results showing that oscillatory responses in the retina (observed here from the LGN responses) largely depend on the size and stationarity of the stimulus. For natural scenes (gray-level and binary movies) oscillations appeared only for brief moments probably when receptive fields were dominated by large continuous, flat-contrast surfaces. Moreover, oscillatory responses to a circle stimulus could be broken with an annular mask indicating that synchronization arises from relatively local interactions among populations of activated cells in the retina. A surprising finding in this study was that retinal oscillations are highly dependent on halothane anesthesia levels. In the absence of halothane, oscillatory activity vanished independent of the characteristics of the stimuli. The same results were obtained for isoflurane, which has similar pharmacological properties. These new and unexpected findings question whether feedfoward oscillations in the early visual system are simply due to an imbalance between excitation and inhibition in the retinal networks generated by the halogenated anesthetics. Further studies in awake behaving animals are necessary to extend these conclusions
Resumo:
The pregeniculate nucleus (PGN) of the primate s thalamus is an agglomerate neuronal having a cap shaped located dorsomedially to the main relay visual information to the cerebral cortex, the dorsal lateral geniculate nucleus (GLD). Several cytoarchitectonic, neurochemical and retinal projections studies have pointed PGN as a structure homologous to intergeniculate leaflet (IGL) of rodents. The IGL receives retinal terminals and appears to be involved in the integration of photic and non-photic information relaying them, through geniculo-hypothalamic tract (TGH), to the main circadian oscillator in mammals, the suprachiasmatic nucleus (SCN) of the hypothalamus. Thus, the IGL participates in the control of the biological rhythm by modulating the activity of the SCN. Pharmacological and IGL injury studies conclude that it is critical in the processing of non-photic information which is transmitted to the SCN. Other studies have found that especially neurons immunoreactive to neuropeptide Y (NPY) respond to this type of stimulation, determined by its colocation with the FOS protein. Has not been determined if the PGN responds, expressing the FOS protein, to the non-photic stimulus nor the neurochemical nature of these cells. Thus, we apply a dark pulse in the specifics circadian phases and analyze the pattern of expression of FOS protein in PGN of the marmoset (Callithrix jacchus). We found that in all animals analyzed the FOS expression was higher in the experimental than in the control group. There was a higher expression of FOS when the dark pulse was applied during the subjective day between the groups. Still, a subregion of the PGN, known by immunoreactive to NPY, had a greater number of FOS-positive cells in relation to his other just close dorsal region. Our data corroborate the theory that the PGN and IGL are homologous structures that were anatomically modified during the evolutionary process, but kept its main neurochemical and functional characteristics. However, injury and hodological studies are still needed for a more accurate conclusion
Resumo:
Esse estudo teve como objetivo examinar possíveis alterações na dinâmica intrínseca de crianças e adultos decorrentes de informações externas na realização de uma tarefa de manutenção da postura ereta. Participaram do estudo dez crianças de 8 anos de idade e dez adultos jovens de ambos os gêneros. Eles permaneceram na posição ereta dentro de uma sala móvel que foi movimentada continuamente para frente e para trás. Os participantes recebiam informação sobre o movimento da sala e eram solicitados a não oscilar ou a oscilar junto com o movimento da mesma. Os resultados mostraram que a manipulação da informação visual induziu oscilação corporal correspondente (dinâmica intrínseca) em crianças e adultos. Informação sobre o movimento da sala e solicitação de uma ação (informação comportamental) alteraram o relacionamento entre informação visual e oscilação corporal. Crianças apresentaram mais dificuldades em alterar a dinâmica intrínseca do que adultos, indicando que elas são mais dependentes da dinâmica intrínseca do que adultos. Esses resultados trazem implicações importantes para a situação de ensino-aprendizagem, pois indica que aprendizagem envolvendo crianças deve ser estruturada propiciando condições mais favoráveis para alterações na dinâmica intrínseca para que os objetivos da mesma sejam alcançados.