864 resultados para camera shake
Resumo:
When depicting both virtual and physical worlds, the viewer's impression of presence in these worlds is strongly linked to camera motion. Plausible and artist-controlled camera movement can substantially increase scene immersion. While physical camera motion exhibits subtle details of position, rotation, and acceleration, these details are often missing for virtual camera motion. In this work, we analyze camera movement using signal theory. Our system allows us to stylize a smooth user-defined virtual base camera motion by enriching it with plausible details. A key component of our system is a database of videos filmed by physical cameras. These videos are analyzed with a camera-motion estimation algorithm (structure-from-motion) and labeled manually with a specific style. By considering spectral properties of location, orientation and acceleration, our solution learns camera motion details. Consequently, an arbitrary virtual base motion, defined in any conventional animation package, can be automatically modified according to a user-selected style. In an animation package the camera motion base path is typically defined by the user via function curves. Another possibility is to obtain the camera path by using a mixed reality camera in motion capturing studio. As shown in our experiments, the resulting shots are still fully artist-controlled, but appear richer and more physically plausible.
Resumo:
Bilayer segmentation of live video in uncontrolled environments is an essential task for home applications in which the original background of the scene must be replaced, as in videochats or traditional videoconference. The main challenge in such conditions is overcome all difficulties in problem-situations (e. g., illumination change, distract events such as element moving in the background and camera shake) that may occur while the video is being captured. This paper presents a survey of segmentation methods for background substitution applications, describes the main concepts and identifies events that may cause errors. Our analysis shows that although robust methods rely on specific devices (multiple cameras or sensors to generate depth maps) which aid the process. In order to achieve the same results using conventional devices (monocular video cameras), most current research relies on energy minimization frameworks, in which temporal and spacial information are probabilistically combined with those of color and contrast.
Resumo:
Visual fixation is employed by humans and some animals to keep a specific 3D location at the center of the visual gaze. Inspired by this phenomenon in nature, this paper explores the idea to transfer this mechanism to the context of video stabilization for a handheld video camera. A novel approach is presented that stabilizes a video by fixating on automatically extracted 3D target points. This approach is different from existing automatic solutions that stabilize the video by smoothing. To determine the 3D target points, the recorded scene is analyzed with a stateof- the-art structure-from-motion algorithm, which estimates camera motion and reconstructs a 3D point cloud of the static scene objects. Special algorithms are presented that search either virtual or real 3D target points, which back-project close to the center of the image for as long a period of time as possible. The stabilization algorithm then transforms the original images of the sequence so that these 3D target points are kept exactly in the center of the image, which, in case of real 3D target points, produces a perfectly stable result at the image center. Furthermore, different methods of additional user interaction are investigated. It is shown that the stabilization process can easily be controlled and that it can be combined with state-of-theart tracking techniques in order to obtain a powerful image stabilization tool. The approach is evaluated on a variety of videos taken with a hand-held camera in natural scenes.
Resumo:
Dynamic foam films have been investigated using an improved experimental set-up with a CCD high-speed linescan camera in conjunction with the Scheludko micro-interferometric cell for studying the drainage and rupture of liquid foam films. The improved experimental set-up increased the sensibility of detection of the local thickness heterogeneities and domains during the film evolution. The evolution of the foam films up to the formation of black spots was recorded in the time intervals of 50ms. The wavelengths of the propagating surface waves and their frequencies were determined experimentally. The experimental results show that the current quasi-static hydrodynamic theory does not properly describe the wave dynamics with inter-domain channels. However, the thermodynamic condition for formation of black spots in the foam films was met by the experimental results. (c) 2005 Elsevier B.V. All rights reserved.
Resumo:
In this paper we present a method for real-time detection and tracking of people in video captured by a depth camera. For each object to be assessed, an ordered sequence of values that represents the distances between its center of mass to the boundary points is calculated. The recognition is based on the analysis of the total distance value between the above sequence and some pre-defined human poses, after apply the Dynamic Time Warping. This similarity approach showed robust results in people detection.
Resumo:
We discuss the operation of a new type of optical sensor (MISCam) based on a metal-insulator-semiconductor (MIS) structure. The operation principle relies on light-induced changes of the band bending and barrier height at the interface between semiconductor and insulator. An image is obtained from the quenching of the ac signal in analogy to the principle of the laser-scanned photodiode (LSP). Lateral resolution depends on the semiconductor material chosen. We have characterised the MIS structures by C-V, I-V, and spectral response measurements testing different types of insulators like a-Si3N4, SiO2, and AlN. The presence of slow interface charges allows for image memory. Colour sensors can be realised by controlling sign and magnitude of the electric fields in the base and the interface region.
Resumo:
Objective - To describe and validate the simulation of the basic features of GE Millennium MG gamma camera using the GATE Monte Carlo platform. Material and methods - Crystal size and thickness, parallel-hole collimation and a realistic energy acquisition window were simulated in the GATE platform. GATE results were compared to experimental data in the following imaging conditions: a point source of 99mTc at different positions during static imaging and tomographic acquisitions using two different energy windows. The accuracy between the events expected and detected by simulation was obtained with the Mann–Whitney–Wilcoxon test. Comparisons were made regarding the measurement of sensitivity and spatial resolution, static and tomographic. Simulated and experimental spatial resolutions for tomographic data were compared with the Kruskal–Wallis test to assess simulation accuracy for this parameter. Results - There was good agreement between simulated and experimental data. The number of decays expected when compared with the number of decays registered, showed small deviation (≤0.007%). The sensitivity comparisons between static acquisitions for different distances from source to collimator (1, 5, 10, 20, 30cm) with energy windows of 126–154 keV and 130–158 keV showed differences of 4.4%, 5.5%, 4.2%, 5.5%, 4.5% and 5.4%, 6.3%, 6.3%, 5.8%, 5.3%, respectively. For the tomographic acquisitions, the mean differences were 7.5% and 9.8% for the energy window 126–154 keV and 130–158 keV. Comparison of simulated and experimental spatial resolutions for tomographic data showed no statistically significant differences with 95% confidence interval. Conclusions - Adequate simulation of the system basic features using GATE Monte Carlo simulation platform was achieved and validated.
Resumo:
The Casa da Música Foundation, responsible for the management of Casa da Música do Porto building, has the need to obtain statistical data related to the number of building’s visitors. This information is a valuable tool for the elaboration of periodical reports concerning the success of this cultural institution. For this reason it was necessary to develop a system capable of returning the number of visitors for a requested period of time. This represents a complex task due to the building’s unique architectural design, characterized by very large doors and halls, and the sudden large number of people that pass through them in moments preceding and proceeding the different activities occurring in the building. To achieve the technical solution for this challenge, several image processing methods, for people detection with still cameras, were first studied. The next step was the development of a real time algorithm, using OpenCV libraries and computer vision concepts,to count individuals with the desired accuracy. This algorithm includes the scientific and technical knowledge acquired in the study of the previous methods. The themes developed in this thesis comprise the fields of background maintenance, shadow and highlight detection, and blob detection and tracking. A graphical interface was also built, to help on the development, test and tunning of the proposed system, as a complement to the work. Furthermore, tests to the system were also performed, to certify the proposed techniques against a set of limited circumstances. The results obtained revealed that the algorithm was successfully applied to count the number of people in complex environments with reliable accuracy.
Resumo:
Oceans - San Diego, 2013
Resumo:
Dissertação para obtenção do Grau de Mestre em Engenharia Biomédica
Resumo:
The Stiles-Crawford effect (SCE) is the well-known phenomenon in which the brightness of light perceived by the human eye depends upon its entrance point in the pupil. This physiological characteristic is due to the directional sensitivity of the cone photoreceptors in the retina and it displays an approximately Gaussian dependency which is altered in a number of pathologies. Retinal imaging, a widely spread clinical practice, may be used to evaluate the SCE and thus serve as diagnostic tool. Nonetheless, its use for such a purpose is still underdeveloped and far from the clinical reality. In this project a fundus camera was built and used to assess the cone photoreceptor directionality by reflective imaging of the retina in healthy individuals. The physical and physiological implications of its development are addressed in detail in the text: the optical properties of the human eye, illumination issues, acquiring a retinal image formed by the eye, among others. A full description of the developmental process that led to the final measuring method and results is also given. The developed setup was successfully used to obtain high quality images of the eye fundus and in particular the parafoveal cone photoreceptors. The SCE was successfully observed and characterized. Even though considerable improvements could be done to the measurement method, the project showed the feasibility of using retinal imaging to evaluate the SCE thus motivating its usage in a clinical environment.
Resumo:
This research addresses the problem of creating interactive experiences to encourage people to explore spaces. Besides the obvious spaces to visit, such as museums or art galleries, spaces that people visit can be, for example, a supermarket or a restaurant. As technology evolves, people become more demanding in the way they use it and expect better forms of interaction with the space that surrounds them. Interaction with the space allows information to be transmitted to the visitors in a friendly way, leading visitors to explore it and gain knowledge. Systems to provide better experiences while exploring spaces demand hardware and software that is not in the reach of every space owner either because of the cost or inconvenience of the installation, that can damage artefacts or the space environment. We propose a system adaptable to the spaces, that uses a video camera network and a wi-fi network present at the space (or that can be installed) to provide means to support interactive experiences using the visitor’s mobile device. The system is composed of an infrastructure (called vuSpot), a language grammar used to describe interactions at a space (called XploreDescription), a visual tool used to design interactive experiences (called XploreBuilder) and a tool used to create interactive experiences (called urSpace). By using XploreBuilder, a tool built of top of vuSpot, a user with little or no experience in programming can define a space and design interactive experiences. This tool generates a description of the space and of the interactions at that space (that complies with the XploreDescription grammar). These descriptions can be given to urSpace, another tool built of top of vuSpot, that creates the interactive experience application. With this system we explore new forms of interaction and use mobile devices and pico projectors to deliver additional information to the users leading to the creation of interactive experiences. The several components are presented as well as the results of the respective user tests, which were positive. The design and implementation becomes cheaper, faster, more flexible and, since it does not depend on the knowledge of a programming language, accessible for the general public.
Resumo:
In order to evaluate the efficiency of different mammalian survey methods, we compared traditional sampling techniques (use of camera-traps on roads and artificial trails, track censuses, and direct field visualization) with an alternative sampling design (camera-traps positioned in natural areas such as natural trails and shelters). We conducted the study in a deciduous Atlantic-Forest park in southern Brazil, and additionally compared our results with a previous intensive study carried out in the same area. Our considerably smaller sampling effort (example: 336 trap.day for our camera-traps versus 2,154 trap.day for the earlier study) registered the presence of 85% of the local known species, with camera-traps being 68% efficient. Moreover, shelter camera-traps revealed a different species composition regarding most of other sampling methods. This sampling strategy involving natural forest sites was therefore able to effectively optimize the chances of evaluating species composition in a shorter period, especially with respect to lower-density and cryptic species, as well as to detect species that avoid open, disturbed sites such as roads and man-made forest trails.