951 resultados para Binocular Stereo


Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we present a novel structure from motion (SfM) approach able to infer 3D deformable models from uncalibrated stereo images. Using a stereo setup dramatically improves the 3D model estimation when the observed 3D shape is mostly deforming without undergoing strong rigid motion. Our approach first calibrates the stereo system automatically and then computes a single metric rigid structure for each frame. Afterwards, these 3D shapes are aligned to a reference view using a RANSAC method in order to compute the mean shape of the object and to select the subset of points on the object which have remained rigid throughout the sequence without deforming. The selected rigid points are then used to compute frame-wise shape registration and to extract the motion parameters robustly from frame to frame. Finally, all this information is used in a global optimization stage with bundle adjustment which allows to refine the frame-wise initial solution and also to recover the non-rigid 3D model. We show results on synthetic and real data that prove the performance of the proposed method even when there is no rigid motion in the original sequence

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Catadioptric sensors are combinations of mirrors and lenses made in order to obtain a wide field of view. In this paper we propose a new sensor that has omnidirectional viewing ability and it also provides depth information about the nearby surrounding. The sensor is based on a conventional camera coupled with a laser emitter and two hyperbolic mirrors. Mathematical formulation and precise specifications of the intrinsic and extrinsic parameters of the sensor are discussed. Our approach overcomes limitations of the existing omni-directional sensors and eventually leads to reduced costs of production

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Resúmen basado en el del autor. Resúmen en castellano y en inglés

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The human visual ability to perceive depth looks like a puzzle. We perceive three-dimensional spatial information quickly and efficiently by using the binocular stereopsis of our eyes and, what is mote important the learning of the most common objects which we achieved through living. Nowadays, modelling the behaviour of our brain is a fiction, that is why the huge problem of 3D perception and further, interpretation is split into a sequence of easier problems. A lot of research is involved in robot vision in order to obtain 3D information of the surrounded scene. Most of this research is based on modelling the stereopsis of humans by using two cameras as if they were two eyes. This method is known as stereo vision and has been widely studied in the past and is being studied at present, and a lot of work will be surely done in the future. This fact allows us to affirm that this topic is one of the most interesting ones in computer vision. The stereo vision principle is based on obtaining the three dimensional position of an object point from the position of its projective points in both camera image planes. However, before inferring 3D information, the mathematical models of both cameras have to be known. This step is known as camera calibration and is broadly describes in the thesis. Perhaps the most important problem in stereo vision is the determination of the pair of homologue points in the two images, known as the correspondence problem, and it is also one of the most difficult problems to be solved which is currently investigated by a lot of researchers. The epipolar geometry allows us to reduce the correspondence problem. An approach to the epipolar geometry is describes in the thesis. Nevertheless, it does not solve it at all as a lot of considerations have to be taken into account. As an example we have to consider points without correspondence due to a surface occlusion or simply due to a projection out of the camera scope. The interest of the thesis is focused on structured light which has been considered as one of the most frequently used techniques in order to reduce the problems related lo stereo vision. Structured light is based on the relationship between a projected light pattern its projection and an image sensor. The deformations between the pattern projected into the scene and the one captured by the camera, permits to obtain three dimensional information of the illuminated scene. This technique has been widely used in such applications as: 3D object reconstruction, robot navigation, quality control, and so on. Although the projection of regular patterns solve the problem of points without match, it does not solve the problem of multiple matching, which leads us to use hard computing algorithms in order to search the correct matches. In recent years, another structured light technique has increased in importance. This technique is based on the codification of the light projected on the scene in order to be used as a tool to obtain an unique match. Each token of light is imaged by the camera, we have to read the label (decode the pattern) in order to solve the correspondence problem. The advantages and disadvantages of stereo vision against structured light and a survey on coded structured light are related and discussed. The work carried out in the frame of this thesis has permitted to present a new coded structured light pattern which solves the correspondence problem uniquely and robust. Unique, as each token of light is coded by a different word which removes the problem of multiple matching. Robust, since the pattern has been coded using the position of each token of light with respect to both co-ordinate axis. Algorithms and experimental results are included in the thesis. The reader can see examples 3D measurement of static objects, and the more complicated measurement of moving objects. The technique can be used in both cases as the pattern is coded by a single projection shot. Then it can be used in several applications of robot vision. Our interest is focused on the mathematical study of the camera and pattern projector models. We are also interested in how these models can be obtained by calibration, and how they can be used to obtained three dimensional information from two correspondence points. Furthermore, we have studied structured light and coded structured light, and we have presented a new coded structured light pattern. However, in this thesis we started from the assumption that the correspondence points could be well-segmented from the captured image. Computer vision constitutes a huge problem and a lot of work is being done at all levels of human vision modelling, starting from a)image acquisition; b) further image enhancement, filtering and processing, c) image segmentation which involves thresholding, thinning, contour detection, texture and colour analysis, and so on. The interest of this thesis starts in the next step, usually known as depth perception or 3D measurement.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La percepció per visió es millorada quan es pot gaudir d'un camp de visió ampli. Aquesta tesi es concentra en la percepció visual de la profunditat amb l'ajuda de càmeres omnidireccionals. La percepció 3D s'obté generalment en la visió per computadora utilitzant configuracions estèreo amb el desavantatge del cost computacional elevat a l'hora de buscar els elements visuals comuns entre les imatges. La solució que ofereix aquesta tesi és l'ús de la llum estructurada per resoldre el problema de relacionar les correspondències. S'ha realitzat un estudi sobre els sistemes de visió omnidireccional. S'han avaluat vàries configuracions estèreo i s'ha escollit la millor. Els paràmetres del model són difícils de mesurar directament i, en conseqüència, s'ha desenvolupat una sèrie de mètodes de calibració. Els resultats obtinguts són prometedors i demostren que el sensor pot ésser utilitzat en aplicacions per a la percepció de la profunditat com serien el modelatge de l'escena, la inspecció de canonades, navegació de robots, etc.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

On 15-17 February 2008, a CME with an approximately circular cross section was tracked through successive images obtained by the Heliospheric Imager (HI) instrument onboard the STEREO-A spacecraft. Reasoning that an idealised flux rope is cylindrical in shape with a circular cross-section, best fit circles are used to determine the radial width of the CME. As part of the process the radial velocity and longitude of propagation are determined by fits to elongation-time maps as 252±5 km/s and 70±5° respectively. With the longitude known, the radial size is calculated from the images, taking projection effects into account. The radial width of the CME, S (AU), obeys a power law with heliocentric distance, R, as the CME travels between 0.1 and 0.4 AU, such that S=0.26 R0.6±0.1. The exponent value obtained is compared to published studies based on statistical surveys of in situ spacecraft observations of ICMEs between 0.3 and 1.0 AU, and general agreement is found. This paper demonstrates the new opportunities provided by HI to track the radial width of CMEs through the previously unobservable zone between the LASCO field of view and Helios in situ measurements.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The suite of SECCHI optical imaging instruments on the STEREO-A spacecraft is used to track a solar storm, consisting of several coronal mass ejections (CMEs) and other coronal loops, as it propagates from the Sun into the heliosphere during May 2007. The 3-D propagation path of the largest interplanetary CME (ICME) is determined from the observations made by the SECCHI Heliospheric Imager (HI) on STEREO-A (HI-1/2A). Two parts of the CME are tracked through the SECCHI images, a bright loop and a V-shaped feature located at the rear of the event. We show that these two structures could be the result of line-of-sight integration of the light scattered by electrons located on a single flux rope. In addition to being imaged by HI, the CME is observed simultaneously by the plasma and magnetic field experiments on the Venus Express and MESSENGER spacecraft. The imaged loop and V-shaped structure bound, as expected, the flux rope observed in situ. The SECCHI images reveal that the leading loop-like structure propagated faster than the V-shaped structure, and a decrease in in situ CME speed occurred during the passage of the flux rope.We interpret this as the result of the continuous radial expansion of the flux rope as it progressed outward through the interplanetary medium. An expansion speed in the radial direction of ~30 km s-1 is obtained directly from the SECCHI-HI images and is in agreement with the difference in speed of the two structures observed in situ. This paper shows that the flux rope location can be determined from white light images, which could have important space weather applications.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Plasma parcels are observed propagating from the Sun out to the large coronal heights monitored by the Heliospheric Imagers (HI) instruments onboard the NASA STEREO spacecraft during September 2007. The source region of these out-flowing parcels is found to corotate with the Sun and to be rooted near the western boundary of an equatorial coronal hole. These plasma enhancements evolve during their propagation through the HI cameras’ fields of view and only becoming fully developed in the outer camera field of view. We provide evidence that HI is observing the formation of a Corotating Interaction Region(CIR) where fast solar wind from the equatorial coronal hole is interacting with the slow solar wind of the streamer belt located on the western edge of that coronal hole. A dense plasma parcel is also observed near the footpoint of the observed CIR at a distance less than 0.1AU from the Sun where fast wind would have not had time to catch up slow wind. We suggest that this low-lying plasma enhancement is a plasma parcel which has been disconnected from a helmet streamer and subsequently becomes embedded inside the corotating interaction region.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Saccadic eye movements and fixations are the behavioral means by which we visually sample text during reading. Human oculomotor control is governed by a complex neurophysiological system involving the brain stem, superior colliculus, and several cortical areas [1, 2]. A very widely held belief among researchers investigating primate vision is that the oculomotor system serves to orient the visual axes of both eyes to fixate the same target point in space. It is argued that such precise positioning of the eyes is necessary to place images on corresponding retinal locations, such that on each fixation a single, nondiplopic, visual representation is perceived [3]. Vision works actively through a continual sampling process involving saccades and fixations [4]. Here we report that during normal reading, the eyes do not always fixate the same letter within a word. We also demonstrate that saccadic targeting is yoked and based on a unified cyclopean percept of a whole word since it is unaffected if different word parts are delivered exclusively to each eye via a dichoptic presentation technique. These two findings together suggest that the visual signal from each eye is fused at a very early stage in the visual pathway, even when the fixation disparity is greater than one character (0.29 deg), and that saccade metrics for each eye are computed on the basis of that fused signal.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Typically, algorithms for generating stereo disparity maps have been developed to minimise the energy equation of a single image. This paper proposes a method for implementing cross validation in a belief propagation optimisation. When tested using the Middlebury online stereo evaluation, the cross validation improves upon the results of standard belief propagation. Furthermore, it has been shown that regions of homogeneous colour within the images can be used for enforcing the so-called "Segment Constraint". Developing from this, Segment Support is introduced to boost belief between pixels of the same image region and improve propagation into textureless regions.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In collaborative situations, eye gaze is a critical element of behavior which supports and fulfills many activities and roles. In current computer-supported collaboration systems, eye gaze is poorly supported. Even in a state-of-the-art video conferencing system such as the access grid, although one can see the face of the user, much of the communicative power of eye gaze is lost. This article gives an overview of some preliminary work that looks towards integrating eye gaze into an immersive collaborative virtual environment and assessing the impact that this would have on interaction between the users of such a system. Three experiments were conducted to assess the efficacy of eye gaze within immersive virtual environments. In each experiment, subjects observed on a large screen the eye-gaze behavior of an avatar. The eye-gaze behavior of that avatar had previously been recorded from a user with the use of a head-mounted eye tracker. The first experiment was conducted to assess the difference between users' abilities to judge what objects an avatar is looking at with only head gaze being viewed and also with eye- and head-gaze data being displayed. The results from the experiment show that eye gaze is of vital importance to the subjects, correctly identifying what a person is looking at in an immersive virtual environment. The second experiment examined whether a monocular or binocular eye-tracker would be required. This was examined by testing subjects' ability to identify where an avatar was looking from their eye direction alone, or by eye direction combined with convergence. This experiment showed that convergence had a significant impact on the subjects' ability to identify where the avatar was looking. The final experiment looked at the effects of stereo and mono-viewing of the scene, with the subjects being asked to identify where the avatar was looking. This experiment showed that there was no difference in the subjects' ability to detect where the avatar was gazing. This is followed by a description of how the eye-tracking system has been integrated into an immersive collaborative virtual environment and some preliminary results from the use of such a system.