978 resultados para IMAGE SEQUENCES


Relevância:

30.00% 30.00%

Publicador:

Resumo:

[EN] [EN] In this paper we present a new method for image primitives tracking based on a CART (Classification and Regression Tree). Primitives tracking procedure uses lines and circles as primitives. We have applied the proposed method to sport event scenarios, specifically, soccer matches. We estimate CART parameters using a learning procedure based on RGB image channels. In order to illustrate its performance, it has been applied to real HD (High Definition) video sequences and some numerical experiments are shown. The quality of the primitives tracking with the decision tree is validated by the percentage error rates obtained and the comparison with other techniques as a morphological method. We also present applications of the proposed method to camera calibration and graphic object insertion in real video sequences.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Facial expression recognition is one of the most challenging research areas in the image recognition ¯eld and has been actively studied since the 70's. For instance, smile recognition has been studied due to the fact that it is considered an important facial expression in human communication, it is therefore likely useful for human–machine interaction. Moreover, if a smile can be detected and also its intensity estimated, it will raise the possibility of new applications in the future

Relevância:

30.00% 30.00%

Publicador:

Resumo:

An algorithm for the real-time registration of a retinal video sequence captured with a scanning digital ophthalmoscope (SDO) to a retinal composite image is presented. This method is designed for a computer-assisted retinal laser photocoagulation system to compensate for retinal motion and hence enhance the accuracy, speed, and patient safety of retinal laser treatments. The procedure combines intensity and feature-based registration techniques. For the registration of an individual frame, the translational frame-to-frame motion between preceding and current frame is detected by normalized cross correlation. Next, vessel points on the current video frame are identified and an initial transformation estimate is constructed from the calculated translation vector and the quadratic registration matrix of the previous frame. The vessel points are then iteratively matched to the segmented vessel centerline of the composite image to refine the initial transformation and register the video frame to the composite image. Criteria for image quality and algorithm convergence are introduced, which assess the exclusion of single frames from the registration process and enable a loss of tracking signal if necessary. The algorithm was successfully applied to ten different video sequences recorded from patients. It revealed an average accuracy of 2.47 ± 2.0 pixels (∼23.2 ± 18.8 μm) for 2764 evaluated video frames and demonstrated that it meets the clinical requirements.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Music consists of sound sequences that require integration over time. As we become familiar with music, associations between notes, melodies, and entire symphonic movements become stronger and more complex. These associations can become so tight that, for example, hearing the end of one album track can elicit a robust image of the upcoming track while anticipating it in total silence. Here, we study this predictive “anticipatory imagery” at various stages throughout learning and investigate activity changes in corresponding neural structures using functional magnetic resonance imaging. Anticipatory imagery (in silence) for highly familiar naturalistic music was accompanied by pronounced activity in rostral prefrontal cortex (PFC) and premotor areas. Examining changes in the neural bases of anticipatory imagery during two stages of learning conditional associations between simple melodies, however, demonstrates the importance of fronto-striatal connections, consistent with a role of the basal ganglia in “training” frontal cortex (Pasupathy and Miller, 2005). Another striking change in neural resources during learning was a shift between caudal PFC earlier to rostral PFC later in learning. Our findings regarding musical anticipation and sound sequence learning are highly compatible with studies of motor sequence learning, suggesting common predictive mechanisms in both domains.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVES: To assess magnetic resonance (MR)-colonography (MRC) for detection of colorectal lesions using two different T1w three-dimensional (3D)-gradient-recalled echo (GRE)-sequences and integrated parallel data acquisition (iPAT) at a 3.0 Tesla MR-unit. MATERIALS AND METHODS: In this prospective study, 34 symptomatic patients underwent dark lumen MRC at a 3.0 Tesla unit before conventional colonoscopy (CC). After colon distension with tap water, 2 high-resolution T1w 3D-GRE [3-dimensional fast low angle shot (3D-FLASH), iPAT factor 2 and 3D-volumetric interpolated breathhold examination (VIBE), iPAT 3] sequences were acquired without and after bolus injection of gadolinium. Prospective evaluation of MRC was performed. Image quality of the different sequences was assessed qualitatively and quantitatively. The findings of the same day CC served as standard of reference. RESULTS: MRC identified all polyps >5 mm (16 of 16) in size and all carcinomas (4 of 4) correctly. Fifty percent of the small polyps image quality was ranked lower in the VIBE, the image quality score of both sequences showed no statistical significant difference (chi > 0.6). CONCLUSIONS: MRC using 3D-GRE-sequences and iPAT is feasible at 3.0 T-systems. The high-resolution 3D-FLASH was slightly preferred over the 3D-VIBE because of better image quality, although both used sequences showed no statistical significant difference.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis deals with the problem of efficiently tracking 3D objects in sequences of images. We tackle the efficient 3D tracking problem by using direct image registration. This problem is posed as an iterative optimization procedure that minimizes a brightness error norm. We review the most popular iterative methods for image registration in the literature, turning our attention to those algorithms that use efficient optimization techniques. Two forms of efficient registration algorithms are investigated. The first type comprises the additive registration algorithms: these algorithms incrementally compute the motion parameters by linearly approximating the brightness error function. We centre our attention on Hager and Belhumeur’s factorization-based algorithm for image registration. We propose a fundamental requirement that factorization-based algorithms must satisfy to guarantee good convergence, and introduce a systematic procedure that automatically computes the factorization. Finally, we also bring out two warp functions to register rigid and nonrigid 3D targets that satisfy the requirement. The second type comprises the compositional registration algorithms, where the brightness function error is written by using function composition. We study the current approaches to compositional image alignment, and we emphasize the importance of the Inverse Compositional method, which is known to be the most efficient image registration algorithm. We introduce a new algorithm, the Efficient Forward Compositional image registration: this algorithm avoids the necessity of inverting the warping function, and provides a new interpretation of the working mechanisms of the inverse compositional alignment. By using this information, we propose two fundamental requirements that guarantee the convergence of compositional image registration methods. Finally, we support our claims by using extensive experimental testing with synthetic and real-world data. We propose a distinction between image registration and tracking when using efficient algorithms. We show that, depending whether the fundamental requirements are hold, some efficient algorithms are eligible for image registration but not for tracking.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a mapping method for wide row crop fields. The resulting map shows the crop rows and weeds present in the inter-row spacing. Because field videos are acquired with a camera mounted on top of an agricultural vehicle, a method for image sequence stabilization was needed and consequently designed and developed. The proposed stabilization method uses the centers of some crop rows in the image sequence as features to be tracked, which compensates for the lateral movement (sway) of the camera and leaves the pitch unchanged. A region of interest is selected using the tracked features, and an inverse perspective technique transforms the selected region into a bird’s-eye view that is centered on the image and that enables map generation. The algorithm developed has been tested on several video sequences of different fields recorded at different times and under different lighting conditions, with good initial results. Indeed, lateral displacements of up to 66% of the inter-row spacing were suppressed through the stabilization process, and crop rows in the resulting maps appear straight

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Although a vast amount of life sciences data is generated in the form of images, most scientists still store images on extremely diverse and often incompatible storage media, without any type of metadata structure, and thus with no standard facility with which to conduct searches or analyses. Here we present a solution to unlock the value of scientific images. The Global Image Database (GID) is a web-based (http://www.g wer.ch/qv/gid/gid.htm) structured central repository for scientific annotated images. The GID was designed to manage images from a wide spectrum of imaging domains ranging from microscopy to automated screening. The annotations in the GID define the source experiment of the images by describing who the authors of the experiment are, when the images were created, the biological origin of the experimental sample and how the sample was processed for visualization. A collection of experimental imaging protocols provides details of the sample preparation, and labeling, or visualization procedures. In addition, the entries in the GID reference these imaging protocols with the probe sequences or antibody names used in labeling experiments. The GID annotations are searchable by field or globally. The query results are first shown as image thumbnail previews, enabling quick browsing prior to original-sized annotated image retrieval. The development of the GID continues, aiming at facilitating the management and exchange of image data in the scientific community, and at creating new query tools for mining image data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Comunicación presentada en el VII Symposium Nacional de Reconocimiento de Formas y Análisis de Imágenes, SNRFAI, Barcelona, abril 1997.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recognition of the object contours in the image as sequences of digital straight segments and/or digital curve arcs is considered in this article. The definitions of digital straight segments and of digital curve arcs are proposed. The methods and programs to recognize the object contours are represented. The algorithm to recognize the digital straight segments is formulated in terms of the growing pyramidal networks taking into account the conceptual model of memory and identification (Rabinovich [4]).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Congenital nystagmus (CN) is an ocular-motor disorder that appears at birth or during the first few months of life; it is characterised by involuntary, conjugated, bilateral to and fro ocular oscillations. Pathogenesis of congenital nystagmus is still unknown. Eye movement recording allow to extract and analyse nystagmus main features such as shape, amplitude and frequency; depending on the morphology of the oscillations nystagmus can be classified in different categories (pendular, jerk, horizontal unidirectional, bidirectional). In general, CN patient show a considerable decrease of the visual acuity: image fixation on the retina is disturbed by nystagmus continuous oscillations; however, image stabilisation is still achieved during the short foveation periods in which eye velocity slows down while the target image is placed onto the fovea. Visual acuity was found to be mainly dependent on foveation periods duration, but cycle-to-cycle foveation repeatability and reduction of retinal image velocities also contribute in increasing visual acuity. This study concentrate on cycle-to-cycle image position variation onto fovea, trying to characterise the sequences of foveation positions. Eye-movement (infrared oculographic or electro oculographic) recordings, relative to different gaze positions and belonging to more than 30 CN patients, were analysed. Preliminary results suggest that sequences of foveations show a cyclic pattern with a dominant frequency (around 0.3 Hz on average) much lower than that of the nystagmus (about 3.3 Hz on average). Sequences of foveations reveals an horizontal ocular swing of more than 2 degree on average, which can explain the low visual acuity of the CN patient. Current CN therapies, pharmacological treatment or surgery of the ocular muscles, mainly aim to increase the patient's visual acuity. Hence, it is fundamental to have an objective parameter (expected visual acuity) for therapy planning. The information about sequences of foveations can improve estimation of patient visual acuity. © 2008 Springer-Verlag.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Congenital nystagmus (CN) is an ocular-motor disorder characterised by involuntary, conjugated ocular oscillations, that can arise since the first months of life. Pathogenesis of congenital nystagmus is still under investigation. In general, CN patients show a considerable decrease of their visual acuity: image fixation on the retina is disturbed by nystagmus continuous oscillations, mainly horizontal. However, image stabilisation is still achieved during the short periods in which eye velocity slows down while the target image is placed onto the fovea (called foveation intervals). To quantify the extent of nystagmus, eye movement recording are routinely employed, allowing physicians to extract and analyse nystagmus main features such as shape, amplitude and frequency. Using eye movement recording, it is also possible to compute estimated visual acuity predictors: analytical functions which estimates expected visual acuity using signal features such as foveation time and foveation position variability. Use of those functions add information to typical visual acuity measurement (e.g. Landolt C test) and could be a support for therapy planning or monitoring. This study focus on robust detection of CN patients' foveations. Specifically, it proposes a method to recognize the exact signal tracts in which a subject foveates, This paper also analyses foveation sequences. About 50 eyemovement recordings, either infrared-oculographic or electrooculographic, from different CN subjects were acquired. Results suggest that an exponential interpolation for the slow phases of nystagmus could improve foveation time computing and reduce influence of breaking saccades and data noise. Moreover a concise description of foveation sequence variability can be achieved using non-fitting splines. © 2009 Springer Berlin Heidelberg.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Image processing offers unparalleled potential for traffic monitoring and control. For many years engineers have attempted to perfect the art of automatic data abstraction from sequences of video images. This paper outlines a research project undertaken at Napier University by the authors in the field of image processing for automatic traffic analysis. A software based system implementing TRIP algorithms to count cars and measure vehicle speed has been developed by members of the Transport Engineering Research Unit (TERU) at the University. The TRIP algorithm has been ported and evaluated on an IBM PC platform with a view to hardware implementation of the pre-processing routines required for vehicle detection. Results show that a software based traffic counting system is realisable for single window processing. Due to the high volume of data required to be processed for full frames or multiple lanes, system operations in real time are limited. Therefore specific hardware is required to be designed. The paper outlines a hardware design for implementation of inter-frame and background differencing, background updating and shadow removal techniques. Preliminary results showing the processing time and counting accuracy for the routines implemented in software are presented and a real time hardware pre-processing architecture is described.