118 resultados para Visual Recognition
em University of Queensland eSpace - Australia
Resumo:
Probabilistic robotics most often applied to the problem of simultaneous localisation and mapping (SLAM), requires measures of uncertainty to accompany observations of the environment. This paper describes how uncertainty can be characterised for a vision system that locates coloured landmarks in a typical laboratory environment. The paper describes a model of the uncertainty in segmentation, the internal cameral model and the mounting of the camera on the robot. It explains the implementation of the system on a laboratory robot, and provides experimental results that show the coherence of the uncertainty model.
Resumo:
Previous work examining context effects in children has been limited to semantic context. The current research examined the effects of grammatical priming of word-naming in fourth-grade children. In Experiment 1, children named both inflected and uninflected noun and verb target words faster when they were preceded by grammatically constraining primes than when they were preceded by neutral primes. Experiment 1 used a long stimulus onset asynchrony (SOA) interval of 750 msec. Experiment 2 replicated the grammatical priming effect at two SOA intervals (400 msec and 700 msec), suggesting that the grammatical priming effect does not reflect the operation of any gross strategic effects directly attributable to the long SOA interval employed in Experiment 1. Grammatical context appears to facilitate target word naming by constraining target word class. Further work is required to elucidate the loci of this effect.
Resumo:
Recovering position from sensor information is an important problem in mobile robotics, known as localisation. Localisation requires a map or some other description of the environment to provide the robot with a context to interpret sensor data. The mobile robot system under discussion is using an artificial neural representation of position. Building a geometrical map of the environment with a single camera and artificial neural networks is difficult. Instead it would be simpler to learn position as a function of the visual input. Usually when learning images, an intermediate representation is employed. An appropriate starting point for biologically plausible image representation is the complex cells of the visual cortex, which have invariance properties that appear useful for localisation. The effectiveness for localisation of two different complex cell models are evaluated. Finally the ability of a simple neural network with single shot learning to recognise these representations and localise a robot is examined.
Resumo:
In studies of mirror-self-recognition subjects are usually surreptitiously marked on their head, and then presented with a mirror. Scores of studies have established that by 18 to 24 months, children investigate their own head upon seeing the mark in the mirror. Scores of papers have debated what this means. Suggestions range from rich interpretations (e.g., the development of self-awareness) to lean accounts (e.g., the development of proprioceptivevisual matching), and include numerous more moderate proposals (e.g., the development of a concept of one's face). In Study 1, 18-24-monthold toddlers were given the standard test and a novel task in which they were marked on their legs rather than on their face. Toddlers performed equivalently on both tasks, suggesting that passing the test does not rely on information specific to facial features. In Study 2, toddlers were surreptitiously slipped into trouser legs that were prefixed to a highchair. Toddlers failed to retrieve the sticker now that their legs looked different from expectations. This finding, together with the findings from a third study which showed that self-recognition in live video feedback develops later than mirror selfrecognition, suggests that performance is not solely the result of proprioceptive-visual matching.
Resumo:
The branching structure of neurones is thought to influence patterns of connectivity and how inputs are integrated within the arbor. Recent studies have revealed a remarkable degree of variation in the branching structure of pyramidal cells in the cerebral cortex of diurnal primates, suggesting regional specialization in neuronal function. Such specialization in pyramidal cell structure may be important for various aspects of visual function, such as object recognition and color processing. To better understand the functional role of regional variation in the pyramidal cell phenotype in visual processing, we determined the complexity of the dendritic branching pattern of pyramidal cells in visual cortex of the nocturnal New World owl monkey. We used the fractal dilation method to quantify the branching structure of pyramidal cells in the primary visual area (V1), the second visual area (V2) and the caudal and rostral subdivisions of inferotemporal cortex (ITc and ITr, respectively), which are often associated with color processing. We found that, as in diurnal monkeys, there was a trend for cells of increasing fractal dimension with progression through these cortical areas. The increasing complexity paralleled a trend for increasing symmetry. That we found a similar trend in both diurnal and nocturnal monkeys suggests that it was a feature of a common anthropoid ancestor.
Resumo:
The McGurk effect, in which auditory [ba] dubbed onto [go] lip movements is perceived as da or tha, was employed in a real-time task to investigate auditory-visual speech perception in prelingual infants. Experiments 1A and 1B established the validity of real-time dubbing for producing the effect. In Experiment 2, 4(1)/(2)-month-olds were tested in a habituation-test paradigm, in which 2 an auditory-visual stimulus was presented contingent upon visual fixation of a live face. The experimental group was habituated to a McGurk stimulus (auditory [ba] visual [ga]), and the control group to matching auditory-visual [ba]. Each group was then presented with three auditory-only test trials, [ba], [da], and [deltaa] (as in then). Visual-fixation durations in test trials showed that the experimental group treated the emergent percept in the McGurk effect, [da] or [deltaa], as familiar (even though they had not heard these sounds previously) and [ba] as novel. For control group infants [da] and [deltaa] were no more familiar than [ba]. These results are consistent with infants'perception of the McGurk effect, and support the conclusion that prelinguistic infants integrate auditory and visual speech information. (C) 2004 Wiley Periodicals, Inc.
Resumo:
Children with autistic spectrum disorder (ASD) may have poor audio-visual integration, possibly reflecting dysfunctional 'mirror neuron' systems which have been hypothesised to be at the core of the condition. In the present study, a computer program, utilizing speech synthesizer software and a 'virtual' head (Baldi), delivered speech stimuli for identification in auditory, visual or bimodal conditions. Children with ASD were poorer than controls at recognizing stimuli in the unimodal conditions, but once performance on this measure was controlled for, no group difference was found in the bimodal condition. A group of participants with ASD were also trained to develop their speech-reading ability. Training improved visual accuracy and this also improved the children's ability to utilize visual information in their processing of speech. Overall results were compared to predictions from mathematical models based on integration and non-integration, and were most consistent with the integration model. We conclude that, whilst they are less accurate in recognizing stimuli in the unimodal condition, children with ASD show normal integration of visual and auditory speech stimuli. Given that training in recognition of visual speech was effective, children with ASD may benefit from multi-modal approaches in imitative therapy and language training. (C) 2004 Elsevier Ltd. All rights reserved.
Resumo:
Previously it has been shown that the branching pattern of pyramidal cells varies markedly between different cortical areas in simian primates. These differences are thought to influence the functional complexity of the cells. In particular, there is a progressive increase in the fractal dimension of pyramidal cells with anterior progression through cortical areas in the occipitotemporal (OT) visual stream, including the primary visual area (V1), the second visual area (V2), the dorsolateral area (DL, corresponding to the fourth visual area) and inferotemporal cortex (IT). However, there are as yet no data on the fractal dimension of these neurons in prosimian primates. Here we focused on the nocturnal prosimian galago (Otolemur garnetti). The fractal dimension (D), and aspect ratio (a measure of branching symmetry), was determined for I I I layer III pyramidal cells in V1, V2, DL and IT. We found, as in simian primates, that the fractal dimension of neurons increased with anterior progression from V1 through V2, DL, and IT. Two important conclusions can be drawn from these results: (1) the trend for increasing branching complexity with anterior progression through OT areas was likely to be present in a common primate ancestor, and (2) specialization in neuron structure more likely facilitates object recognition than spectral processing.
A longitudinal investigation of imitation, pretend play and mirror self-recognition in human infants
Resumo:
By 24-months of age most children show mirror self-recognition. When surreptitiously marked on their forehead and then presented with a mirror, they explore their own head for the unexpected mark. Here we demonstrate that self-recognition in mirrors does not generalize to other visual feedback. We tested 80 children on mirror and live video versions of the task. Whereas 90% of 24-month olds passed the mirror version, only 35% passed the video version. Seventy percent of 30-month olds showed video selfrecognition and only by age 36-months did the pass rate on the video version reach 90%. It remains to be y 24-months of age most children show mirror self-recognition. When surreptitiously marked on their forehead and then presented with a mirror, they explore their own head for the unexpected mark. Here we demonstrate that self-recognition in mirrors does not generalize to other visual feedback. We tested 80 children on mirror and live video versions of the task. Whereas 90% of 24-month olds passed the mirror version, only 35% passed the video version. Seventy percent of 30-month olds showed video selfrecognition and only by age 36-months did the pass rate on the video version reach 90%. It remains to be
Resumo:
The influence of temporal association on the representation and recognition of objects was investigated. Observers were shown sequences of novel faces in which the identity of the face changed as the head rotated. As a result, observers showed a tendency to treat the views as if they were of the same person. Additional experiments revealed that this was only true if the training sequences depicted head rotations rather than jumbled views; in other words, the sequence had to be spatially as well as temporally smooth. Results suggest that we are continuously associating views of objects to support later recognition, and that we do so not only on the basis of the physical similarity, but also the correlated appearance in time of the objects.
Resumo:
Some motor tasks can be completed, quite literally, with our eyes shut. Most people can touch their nose without looking or reach for an object after only a brief glance at its location. This distinction leads to one of the defining questions of movement control: is information gleaned prior to starting the movement sufficient to complete the task (open loop), or is feedback about the progress of the movement required (closed loop)? One task that has commanded considerable interest in the literature over the years is that of steering a vehicle, in particular lane-correction and lane-changing tasks. Recent work has suggested that this type of task can proceed in a fundamentally open loop manner [1 and 2], with feedback mainly serving to correct minor, accumulating errors. This paper reevaluates the conclusions of these studies by conducting a new set of experiments in a driving simulator. We demonstrate that, in fact, drivers rely on regular visual feedback, even during the well-practiced steering task of lane changing. Without feedback, drivers fail to initiate the return phase of the maneuver, resulting in systematic errors in final heading. The results provide new insight into the control of vehicle heading, suggesting that drivers employ a simple policy of “turn and see,” with only limited understanding of the relationship between steering angle and vehicle heading.
Resumo:
We examined the influence of backrest inclination and vergence demand on the posture and gaze angle that-workers adopt to view visual targets placed in different vertical locations. In the study 12 participants viewed a small video monitor placed in 7 locations around a 0.65-m radius arc (from 650 below to 300 above horizontal eye height). Trunk posture was manipulated by changing the backrest inclination of an adjustable chair. Vergence demand was manipulated by using ophthalmic lenses and prisms to mimic the visual consequences of varying target distance. Changes in vertical target location caused large changes in atlantooccipital posture and gaze angle. Cervical posture was altered to a lesser extent by changes in vertical target location. Participants compensated for changes in backrest inclination by changing cervical posture, though they did not significantly alter atlanto-occipital posture and gaze angle. The posture adopted to view any target represents a compromise between visual and musculoskeletal demands. These results provide support for the argument that the optimal location of visual targets is at least 15 below horizontal eye level. Actual or potential applications of this work include the layout of computer workstations and the viewing of displays from a seated posture.
Resumo:
Spectral peak resolution was investigated in normal hearing (NH), hearing impaired (HI), and cochlear implant (CI) listeners. The task involved discriminating between two rippled noise stimuli in which the frequency positions of the log-spaced peaks and valleys were interchanged. The ripple spacing was varied adaptively from 0.13 to 11.31 ripples/octave, and the minimum ripple spacing at which a reversal in peak and trough positions could be detected was determined as the spectral peak resolution threshold for each listener. Spectral peak resolution was best, on average, in NH listeners, poorest in CI listeners, and intermediate for HI listeners. There was a significant relationship between spectral peak resolution and both vowel and consonant recognition in quiet across the three listener groups. The results indicate that the degree of spectral peak resolution required for accurate vowel and consonant recognition in quiet backgrounds is around 4 ripples/octave, and that spectral peak resolution poorer than around 1–2 ripples/octave may result in highly degraded speech recognition. These results suggest that efforts to improve spectral peak resolution for HI and CI users may lead to improved speech recognition
Resumo:
The purpose of this study was to explore the potential advantages, both theoretical and applied, of preserving low-frequency acoustic hearing in cochlear implant patients. Several hypotheses are presented that predict that residual low-frequency acoustic hearing along with electric stimulation for high frequencies will provide an advantage over traditional long-electrode cochlear implants for the recognition of speech in competing backgrounds. A simulation experiment in normal-hearing subjects demonstrated a clear advantage for preserving low-frequency residual acoustic hearing for speech recognition in a background of other talkers, but not in steady noise. Three subjects with an implanted "short-electrode" cochlear implant and preserved low-frequency acoustic hearing were also tested on speech recognition in the same competing backgrounds and compared to a larger group of traditional cochlear implant users. Each of the three short-electrode subjects performed better than any of the traditional long-electrode implant subjects for speech recognition in a background of other talkers, but not in steady noise, in general agreement with the simulation studies. When compared to a subgroup of traditional implant users matched according to speech recognition ability in quiet, the short-electrode patients showed a 9-dB advantage in the multitalker background. These experiments provide strong preliminary support for retaining residual low-frequency acoustic hearing in cochlear implant patients. The results are consistent with the idea that better perception of voice pitch, which can aid in separating voices in a background of other talkers, was responsible for this advantage.