993 resultados para video sequences


Relevância:

60.00% 60.00%

Publicador:

Resumo:

We propose a scheme for indoor place identification based on the recognition of global scene views. Scene views are encoded using a holistic representation that provides low-resolution spatial and spectral information. The holistic nature of the representation dispenses with the need to rely on specific objects or local landmarks and also renders it robust against variations in object configurations. We demonstrate the scheme on the problem of recognizing scenes in video sequences captured while walking through an office environment. We develop a method for distinguishing between 'diagnostic' and 'generic' views and also evaluate changes in system performances as a function of the amount of training data available and the complexity of the representation.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A common problem in video surveys in very shallow waters is the presence of strong light fluctuations, due to sun light refraction. Refracted sunlight casts fast moving patterns, which can significantly degrade the quality of the acquired data. Motivated by the growing need to improve the quality of shallow water imagery, we propose a method to remove sunlight patterns in video sequences. The method exploits the fact that video sequences allow several observations of the same area of the sea floor, over time. It is based on computing the image difference between a given reference frame and the temporal median of a registered set of neighboring images. A key observation is that this difference will have two components with separable spectral content. One is related to the illumination field (lower spatial frequencies) and the other to the registration error (higher frequencies). The illumination field, recovered by lowpass filtering, is used to correct the reference image. In addition to removing the sunflickering patterns, an important advantage of the approach is the ability to preserve the sharpness in corrected image, even in the presence of registration inaccuracies. The effectiveness of the method is illustrated in image sets acquired under strong camera motion containing non-rigid benthic structures. The results testify the good performance and generality of the approach

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents a complete solution for creating accurate 3D textured models from monocular video sequences. The methods are developed within the framework of sequential structure from motion, where a 3D model of the environment is maintained and updated as new visual information becomes available. The camera position is recovered by directly associating the 3D scene model with local image observations. Compared to standard structure from motion techniques, this approach decreases the error accumulation while increasing the robustness to scene occlusions and feature association failures. The obtained 3D information is used to generate high quality, composite visual maps of the scene (mosaics). The visual maps are used to create texture-mapped, realistic views of the scene

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents a paralleled Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA., Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). We introduced hashtable into video processing and completed parallel implementation. We propose and evaluate parallel implementations of the LHMEA of TPA on clusters of workstations for real time video compression. It discusses how parallel video coding on load balanced multiprocessor systems can help, especially on motion estimation. The effect of load balancing for improved performance is discussed. The performance or the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents a novel two-pass algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS). compensation. for block base motion On the basis of research from previous algorithms, especially an on-the-edge motion estimation algorithm called hexagonal search (HEXBS), we propose the LHMEA and the Two-Pass Algorithm (TPA). We introduce hashtable into video compression. In this paper we employ LHMEA for the first-pass search in all the Macroblocks (MB) in the picture. Motion Vectors (MV) are then generated from the first-pass and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of MBs. The evaluation of the algorithm considers the three important metrics being time, compression rate and PSNR. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms. Experimental results show that the proposed algorithm can offer the same compression rate as the Full Search. LHMEA with TPA has significant improvement on HEXBS and shows a direction for improving other fast motion estimation algorithms, for example Diamond Search.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents a paralleled Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA, Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). We introduced hashtable into video processing and completed parallel implementation. We propose and evaluate parallel implementations of the LHMEA of TPA on clusters of workstations for real time video compression. It discusses how parallel video coding on load balanced multiprocessor systems can help, especially on motion estimation. The effect of load balancing for improved performance is discussed. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents an improved Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA, Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). The hashtable structure of LHMEA is improved compared to the original TPA and LHMEA. The evaluation of the algorithm considers the three important metrics being processing time, compression rate and PSNR. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A novel Linear Hashtable Method Predicted Hexagonal Search (LHMPHS) method for block based motion compensation is proposed. Fast block matching algorithms use the origin as the initial search center, which often does not track motion very well. To improve the accuracy of the fast BMA's, we employ a predicted starting search point, which reflects the motion trend of the current block. The predicted search centre is found closer to the global minimum. Thus the center-biased BMA's can be used to find the motion vector more efficiently. The performance of the algorithm is evaluated by using standard video sequences, considers the three important metrics: The results show that the proposed algorithm enhances the accuracy of current hexagonal algorithms and is better than Full Search, Logarithmic Search etc.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents a novel two-pass algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for block base motion compensation. On the basis of research from previous algorithms, especially an on-the-edge motion estimation algorithm called hexagonal search (HEXBS), we propose the LHMEA and the Two-Pass Algorithm (TPA). We introduced hashtable into video compression. In this paper we employ LHMEA for the first-pass search in all the Macroblocks (MB) in the picture. Motion Vectors (MV) are then generated from the first-pass and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of MBs. The evaluation of the algorithm considers the three important metrics being time, compression rate and PSNR. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms, Experimental results show that the proposed algorithm can offer the same compression rate as the Full Search. LHMEA with TPA has significant improvement on HEXBS and shows a direction for improving other fast motion estimation algorithms, for example Diamond Search.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Automatically extracting interesting objects from videos is a very challenging task and is applicable to many research areas such robotics, medical imaging, content based indexing and visual surveillance. Automated visual surveillance is a major research area in computational vision and a commonly applied technique in an attempt to extract objects of interest is that of motion segmentation. Motion segmentation relies on the temporal changes that occur in video sequences to detect objects, but as a technique it presents many challenges that researchers have yet to surmount. Changes in real-time video sequences not only include interesting objects, environmental conditions such as wind, cloud cover, rain and snow may be present, in addition to rapid lighting changes, poor footage quality, moving shadows and reflections. The list provides only a sample of the challenges present. This thesis explores the use of motion segmentation as part of a computational vision system and provides solutions for a practical, generic approach with robust performance, using current neuro-biological, physiological and psychological research in primate vision as inspiration.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The researcher worked closely with two biology-trained teachers to plan three teaching sequences in the topics of forces, substances and astronomy that were subsequently taught to Year 7 students. The sequences sought to develop a model of classroom practice that foregrounds students’ negotiation of conceptual representations.

The difficulties encountered by individuals in learning science point to the need for a very strong emphasis of the role of representations in learning. There is a need for learners to use their own representational, cultural and cognitive resources to engage with the subject-specific representational practices of science. Researchers who have undertaken classroom studies whereby students have constructed and used their own representations have pointed to several principles in the planning, execution and assessment of student learning (diSessa, 2004; Greeno & Hall, 1997). A key principle is that teachers need to identify big ideas, key concepts, of the topic at the planning stage in order to guide refinement of representational work. These researchers also point out the need for students to engage with multiple representations in different modes that are both teacher and student generated. A representation can only partially explain a particular phenomenon or process and has both positive and negative attributes to the target that it represents. The issue of the partial nature of representations needs to be a component of classroom practice (Greeno & Hall, 1997) in terms of students critiquing representations for their limitations and affordances and explicitly linking multiple representations to construct a fuller understanding of the phenomenon or process under study. The classroom practice should also provide opportunities for students to manipulate representations as reasoning tools (Cox, 1999) in constructing the scientifically acceptable ideas and communicating them.

Research question: What impact was there on the participating teacher’s practice through the adoption of a representational focus to teaching science?

Data collection included video sequences of classroom practice and student responses, student work, field notes, tape records of meetings and discussions, and student and teacher interviews based in some cases on video stimulated recall. Video analysis software was used to capture the variety of representations used, and sequences of representational negotiation.

The teachers in this study reported substantial shifts in their classroom practices, and in the quality of classroom discussions, arising from adopting a representational focus. The shifts were reported by them as a three-fold challenge. First, there was an epistemological challenge as they came to terms with the culturally produced nature of representations in the topics of force, substance and astronomy and their flexibility and power as tools for analysis and communication, as opposed to their previous assumption that this was given knowledge to be learnt as an end point. The second challenge was pedagogical, in that this approach was acknowledged to place much greater agency in the hands of students, and this brought a need to learn to run longer and more structured discussions around conceptual problems. The third challenge related to content coverage. The teachers sacrificed coverage for the greater depth offered by this approach, and were unanimous in their judgment that this had been a change that had paid dividends in terms of student learning.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Composing a multimedia presentation may require creation or generation of suitable images and video segments, as well as animation, sound, or special effects. Obtaining images or video sequences can be prohibitively expensive when costs of travel to location, equipment, staff, etc, are considered. Those problems can be alleviated with the use of pictorial and video digital libraries, such libraries require methods for comprehensive indexing and annotation of stored items and efficient retrieval tools.

We propose a system based on user oriented perceptions as they influence query formation in image and video retrieval. We present a method based on user dependent conceptual structures for creating and maintaining indexes to images and video sequences.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The aim of this work is to devise an effective method for static summarization of home video sequences. Based on the premise that the user watching a summary is interested in people related (how many, who, emotional state) or activity related aspects, we formulate a novel approach to video summarization that works to specifically expose relevant video frames that make the content spotting tasks possible. Unlike existing approaches, which work on low-level features which often produce the summary not appealing to the viewer due to the semantic gap between low-level features and high-level concepts, our approach is driven by various utility functions (identity count, identity recognition, emotion recognition, activity recognition, sense of space) that use the results of face detection, face clustering, shot clustering and within cluster frame alignment. The summarization problem is then treated as the problem of extracting the set of key frames that have the maximum combined utility.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Movie directors and producers worldwide, in their quest to narrate a good story that warrants repeated audience viewing, use many cinematic elements to intensify and clarify the viewing experience. One such element that directors manipulate is lighting. In this paper we examine one aspect of lighting, namely flashing lights, and its role as an intensifier of dramatic effects in film. We present an algorithm for robust extraction of flashing lights and a simple mechanism to group detected flashing lights into flashing light scenes and analyze the role of these segments in story narration. In addition, we demonstrate how flashing lights detection can improve the performance of shot-based video segmentation. Experiments on a number of video sequences extracted from real movies yields good results. Our technique detects 90.4% of flashing lights. The detected flashing lights correctly eliminates 92.7% of false cuts in these sequences. In addition, data support is compiled to demonstrate the association between flashing light scenes and certain dramatic intensification events such as supernatural power, crisis or excitement.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The Point Distribution Model (PDM) has proven effective in modelling variations in shape in sets of images, including those in which motion is involved such as body and hand tracking. This paper proposes an extension to the PDM through a re-parameterisation of the model which uses factors such as the angular velocity and distance travelled for sets of points on a moving shape. This then enables non-linear quantities such as acceleration and the average velocity of the body to be expressed in a linear model by the PDM. Results are shown for objects with known acceleration and deceleration components, these being a simulated pendulum modelled using simple harmonic motion and video sequences of a real pendulum in motion.