946 resultados para Estimation of gaze direction
Resumo:
Purpose. The purpose of the study was to investigate the changes in axial length occurring with shifts in gaze direction. Methods. Axial length measurements were obtained from the left eye of 30 young adults (10 emmetropes, 10 low myopes, and 10 moderate myopes) through a rotating prism with 15° deviation, along the foveal axis, using a noncontact optical biometer in each of the nine different cardinal directions of gaze over 5 minutes. The subject's fellow eye fixated on an external distance (6 m) target to control accommodation, also with 15° deviation. Axial length measurements were also performed in 15° and 25° downward gaze with the biometer inclined on a tilting table, allowing gaze shifts to be achieved with either full head turn but no eye turn, or full eye turn with no head turn. Results. There was a significant influence of gaze angle and time on axial length (both P < 0.001), with the greatest axial elongation (+18 ± 8 μm) occurring with inferonasal gaze (P < 0.001) and a slight decrease in axial length in superior gaze (−12 ± 17 μm) compared with primary gaze (P < 0.001). In downward gaze, a significant axial elongation occurred when eye turn was used (P < 0.001), but not when head turn was used to shift gaze (P > 0.05). Conclusions. The angle of gaze has a small but significant short-term effect on axial length, with greatest elongation occurring in inferonasal gaze. The elongation of the eye appears to be due to the influence of the extraocular muscles, in particular the oblique muscles.
Resumo:
Control of the neck muscles is coordinated with the sensory organs of vision, hearing and balance. For instance, activity of splenius capitis (SC) is modified with gaze shift. This interaction between eye movement and neck muscle activity is likely to influence the control of neck movement. The aim of this study was to investigate the effect of eye position on neck muscle activity during cervical rotation. In eleven subjects we recorded electromyographic activity (EMG) of muscles that rotate the neck to the right [right obliquus capitis inferior (OI), multifides (MF), and SC, and left sternocleidomastoid (SCM)] with intramuscular or surface electrodes. In sitting, subjects rotated the neck in each direction to specific points in range that were held statically with gaze either fixed to a guide (at three different positions) that moved with the head to maintain a constant intra-orbit eye position or to a panel in front of the subject. Although right SC and left SCM EMG increased with rotation to the right, contrary to anatomical texts, OI EMG increased with both directions and MF EMG did not change from the activity recorded at rest. During neck rotation SCM and MF EMG was less when the eyes were maintained with a constant intra-orbit position that was opposite to the direction of rotation compared to trials in which the eyes were maintained in the same direction as the head movement. The inter-relationship between eye position and neck muscle activity may affect the control of neck posture and movement.
Resumo:
The current study assessed the importance of infant detection of contingency and head and eye gaze direction in the emergence of social referencing. Five- to six-month-old infants' detection of affect-object relations and subsequent manual preferences for objects paired with positive expressions were assessed. In particular, the role of contingency between toys' movements and an actress's emotional expressions as well as the role of gaze direction toward the toys' location were examined. Infants were habituated to alternating films of two toys each paired with an actress's affective expression (happy and fearful) under contingent or noncontingent and gaze congruent or gaze incongruent conditions. Results indicated that gaze congruence and contingency between toys' movements and a person's affective expressions were important for infant perception of affect-object relations. Furthermore, infant perception of the relation between affective expressions and toys translated to their manual preferences for the 3-dimensional toys. Infants who received contingent affective responses to the movements of the toys spent more time touching the toy that was previously paired with the positive expression. These findings demonstrate the role of contingency and gaze direction in the emergence of social referencing in the first half year of life.^
Resumo:
Ce mémoire présente deux algorithmes qui ont pour but d’améliorer la précision de l’estimation de la direction d’arrivée de sources sonores et de leurs échos. Le premier algorithme, qui s’appelle la méthode par élimination des sources, permet d’améliorer l’estimation de la direction d’arrivée d’échos qui sont noyés dans le bruit. Le second, qui s’appelle Multiple Signal Classification à focalisation de phase, utilise l’information dans la phase à chaque fréquence pour déterminer la direction d’arrivée de sources à large bande. La combinaison de ces deux algorithmes permet de localiser des échos dont la puissance est de -17 dB par rapport à la source principale, jusqu’à un rapport échoà- bruit de -15 dB. Ce mémoire présente aussi des mesures expérimentales qui viennent confirmer les résultats obtenus lors de simulations.
Resumo:
The fundamental senses of the human body are: vision, hearing, touch, taste and smell. These senses are the functions that provide our relationship with the environment. The vision serves as a sensory receptor responsible for obtaining information from the outside world that will be sent to the brain. The gaze reflects its attention, intention and interest. Therefore, the estimation of gaze direction, using computer tools, provides a promising alternative to improve the capacity of human-computer interaction, mainly with respect to those people who suffer from motor deficiencies. Thus, the objective of this work is to present a non-intrusive system that basically uses a personal computer and a low cost webcam, combined with the use of digital image processing techniques, Wavelets transforms and pattern recognition, such as artificial neural network models, resulting in a complete system that performs since the image acquisition (including face detection and eye tracking) to the estimation of gaze direction. The obtained results show the feasibility of the proposed system, as well as several feature advantages.
Resumo:
In this paper we present a convolutional neuralnetwork (CNN)-based model for human head pose estimation inlow-resolution multi-modal RGB-D data. We pose the problemas one of classification of human gazing direction. We furtherfine-tune a regressor based on the learned deep classifier. Next wecombine the two models (classification and regression) to estimateapproximate regression confidence. We present state-of-the-artresults in datasets that span the range of high-resolution humanrobot interaction (close up faces plus depth information) data tochallenging low resolution outdoor surveillance data. We buildupon our robust head-pose estimation and further introduce anew visual attention model to recover interaction with theenvironment. Using this probabilistic model, we show thatmany higher level scene understanding like human-human/sceneinteraction detection can be achieved. Our solution runs inreal-time on commercial hardware
Resumo:
In this letter, a new wind-vector algorithm is presented that uses radar backscatter sigma(0) measurements at two adjacent subscenes of RADARSAT-1 synthetic aperture radar (SAR) images, with each subscene having slightly different geometry. Resultant wind vectors are validated using in situ buoy measurements and compared with wind vectors determined from a hybrid wind-retrieval model using wind directions determined by spectral analysis of wind-induced image streaks and observed by colocated QuikSCAT measurements. The hybrid wind-retrieval model consists of CMOD-IFR2 [applicable to C-band vertical-vertical (W) polarization] and a C-band copolarization ratio according to Kirchhoff scattering. The new algorithm displays improved skill in wind-vector estimation for RADARSAT-1 SAR data when compared to conventional wind-retrieval methodology. In addition, unlike conventional methods, the present method is applicable to RADARSAT-1 images both with and without visible streaks. However, this method requires ancillary data such as buoy measurements to resolve the ambiguity in retrieved wind direction.
Resumo:
In this paper three problems related to the analysis of facial images are addressed: the illuminant direction, the compensation of illumination effects and, finally, the recovery of the pose of the face, restricted to in-depth rotations. The solutions proposed for these problems rely on the use of computer graphics techniques to provide images of faces under different illumination and pose, starting from a database of frontal views under frontal illumination.
Resumo:
The capability of estimating the walking direction of people would be useful in many applications such as those involving autonomous cars and robots. We introduce an approach for estimating the walking direction of people from images, based on learning the correct classification of a still image by using SVMs. We find that the performance of the system can be improved by classifying each image of a walking sequence and combining the outputs of the classifier. Experiments were performed to evaluate our system and estimate the trade-off between number of images in walking sequences and performance.
Resumo:
The over represented number of novice drivers involved in crashes is alarming. Driver training is one of the interventions aimed at mitigating the number of crashes that involve young drivers. Experienced drivers have better hazard perception ability compared to inexperienced drivers. Eye gaze patterns have been found to be an indicator of the driver's competency level. The aim of this paper is to develop an in-vehicle system which correlates information about the driver's gaze and vehicle dynamics, which is then used to assist driver trainers in assessing driving competency. This system allows visualization of the complete driving manoeuvre data on interactive maps. It uses an eye tracker and perspective projection algorithms to compute the depth of gaze and plots it on Google maps. This interactive map also features the trajectory of the vehicle and turn indicator usage. This system allows efficient and user friendly analysis of the driving task. It can be used by driver trainers and trainees to understand objectively the risks encountered during driving manoeuvres. This paper presents a prototype that plots the driver's eye gaze depth and direction on an interactive map along with the vehicle dynamics information. This prototype will be used in future to study the difference in gaze patterns in novice and experienced drivers prior to a certain manoeuvre.
Resumo:
A numerical solution for the transient temperature distribution in a cylindrical disc heated on its top surface by a circular source is presented. A finite difference form of the governing equations is solved by the Alternating Direction Implicit (ADI) time marching scheme. This solution has direct applications in analyzing transient electron beam heating of target materials as encountered in the prebreakdown current enhancement and consequent breakdown in high voltage vacuum gaps stressed by alternating and pulsed voltages. The solution provides an estimate of the temperature for pulsed electron beam heating and the size of thermally activated microparticles originating from anode hot spots. The calculated results for a typical 45kV (a.c.) electron beam of radius 2.5 micron indicate that the temperature of such spots can reach melting point and could give rise to microparticles which could initiate breakdown.
Resumo:
A Monte Carlo model of ultrasound modulation of multiply scattered coherent light in a highly scattering media has been carried out for estimating the phase shift experienced by a photon beam on its transit through US insonified region. The phase shift is related to the tissue stiffness, thereby opening an avenue for possible breast tumor detection. When the scattering centers in the tissue medium is exposed to a deterministic forcing with the help of a focused ultrasound (US) beam, due to the fact that US-induced oscillation is almost along particular direction, the direction defined by the transducer axis, the scattering events increase, thereby increasing the phase shift experienced by light that traverses through the medium. The phase shift is found to increase with increase in anisotropy g of the medium. However, as the size of the focused region which is the region of interest (ROI) increases, a large number of scattering events take place within the ROI, the ensemble average of the phase shift (Delta phi) becomes very close to zero. The phase of the individual photon is randomly distributed over 2 pi when the scattered photon path crosses a large number of ultrasound wavelengths in the focused region. This is true at high ultrasound frequency (1 MHz) when mean free path length of photon l(s) is comparable to wavelength of US beam. However, at much lower US frequencies (100 Hz), the wavelength of sound is orders of magnitude larger than l(s), and with a high value of g (g 0.9), there is a distinct measurable phase difference for the photon that traverses through the insonified region. Experiments are carried out for validation of simulation results.
Resumo:
Regions in video streams attracting human interest contribute significantly to human understanding of the video. Being able to predict salient and informative Regions of Interest (ROIs) through a sequence of eye movements is a challenging problem. Applications such as content-aware retargeting of videos to different aspect ratios while preserving informative regions and smart insertion of dialog (closed-caption text) into the video stream can significantly be improved using the predicted ROIs. We propose an interactive human-in-the-loop framework to model eye movements and predict visual saliency into yet-unseen frames. Eye tracking and video content are used to model visual attention in a manner that accounts for important eye-gaze characteristics such as temporal discontinuities due to sudden eye movements, noise, and behavioral artifacts. A novel statistical-and algorithm-based method gaze buffering is proposed for eye-gaze analysis and its fusion with content-based features. Our robust saliency prediction is instantiated for two challenging and exciting applications. The first application alters video aspect ratios on-the-fly using content-aware video retargeting, thus making them suitable for a variety of display sizes. The second application dynamically localizes active speakers and places dialog captions on-the-fly in the video stream. Our method ensures that dialogs are faithful to active speaker locations and do not interfere with salient content in the video stream. Our framework naturally accommodates personalisation of the application to suit biases and preferences of individual users.