26 resultados para Augmented-Reality
Resumo:
In this paper, we propose the use of specific system architecture, based on mobile device, for navigation in urban environments. The aim of this work is to assess how virtual and augmented reality interface paradigms can provide enhanced location based services using real-time techniques in the context of these two different technologies. The virtual reality interface is based on faithful graphical representation of the localities of interest, coupled with sensory information on the location and orientation of the user, while the augmented reality interface uses computer vision techniques to capture patterns from the real environment and overlay additional way-finding information, aligned with real imagery, in real-time. The knowledge obtained from the evaluation of the virtual reality navigational experience has been used to inform the design of the augmented reality interface. Initial results of the user testing of the experimental augmented reality system for navigation are presented.
Resumo:
Television and movie images have been altered ever since it was technically possible. Nowadays embedding advertisements, or incorporating text and graphics in TV scenes, are common practice, but they can not be considered as integrated part of the scene. The introduction of new services for interactive augmented television is discussed in this paper. We analyse the main aspects related with the whole chain of augmented reality production. Interactivity is one of the most important added values of the digital television: This paper aims to break the model where all TV viewers receive the same final image. Thus, we introduce and discuss the new concept of interactive augmented television, i. e. real time composition of video and computer graphics - e.g. a real scene and freely selectable images or spatial rendered objects - edited and customized by the end user within the context of the user's set top box and TV receiver.
Resumo:
This article deals with embodied user interfaces for handheld augmented reality games, which consist of both physical and virtual components. We have developed a number of spatial interaction techniques that optically capture the device's movement and orientation relative to a visual marker. Such physical interactions in 3-D space enable manipulative control of mobile games. In addition to acting as a physical controller that recognizes multiple game-dependent gestures, the mobile device augments the camera view with graphical overlays. We describe three game prototypes that use ubiquitous product packaging and other passive media as backgrounds for handheld augmentation. The prototypes can be realized on widely available off-the-shelf hardware and require only minimal setup and infrastructure support.
Resumo:
Projects in the area of architectural design and urban planning typically engage several architects as well as experts from other professions. While the design and review meetings thus often involve a large number of cooperating participants, the actual design is still done by the individuals in the time in between those meetings using desktop PCs and CAD applications. A real collaborative approach to architectural design and urban planning is often limited to early paper-based sketches.In order to overcome these limitations, we designed and realized the ARTHUR system, an Augmented Reality (AR) enhanced round table to support complex design and planning decisions for architects. WhileAR has been applied to this area earlier, our approach does not try to replace the use of CAD systems but rather integrates them seamlessly into the collaborative AR environment. The approach is enhanced by intuitiveinteraction mechanisms that can be easily con-figured for different application scenarios.
Resumo:
Virtual studio technology plays an important role for modern television productions. Blue-screen matting is a common technique for integrating real actors or moderators into computer generated sceneries. Augmented reality offers the possibility to mix real and virtual in a more general context. This article proposes a new technological approach for combining real studio content with computergenerated information. Digital light projection allows a controlled spatial, temporal, chrominance and luminance modulation of illumination – opening new possibilities for TV studios.
Resumo:
In this paper the software architecture of a framework which simplifies the development of applications in the area of Virtual and Augmented Reality is presented. It is based on VRML/X3D to enable rendering of audio-visual information. We extended our VRML rendering system by a device management system that is based on the concept of a data-flow graph. The aim of the system is to create Mixed Reality (MR) applications simply by plugging together small prefabricated software components, instead of compiling monolithic C++ applications. The flexibility and the advantages of the presented framework are explained on the basis of an exemplary implementation of a classic Augmented Realityapplication and its extension to a collaborative remote expert scenario.
Resumo:
Spatial tracking is one of the most challenging and important parts of Mixed Reality environments. Many applications, especially in the domain of Augmented Reality, rely on the fusion of several tracking systems in order to optimize the overall performance. While the topic of spatial tracking sensor fusion has already seen considerable interest, most results only deal with the integration of carefully arranged setups as opposed to dynamic sensor fusion setups. A crucial prerequisite for correct sensor fusion is the temporal alignment of the tracking data from several sensors. Tracking sensors are typically encountered in Mixed Reality applications, are generally not synchronized. We present a general method to calibrate the temporal offset between different sensors by the Time Delay Estimation method which can be used to perform on-line temporal calibration. By applying Time Delay Estimation on the tracking data, we show that the temporal offset between generic Mixed Reality spatial tracking sensors can be calibrated. To show the correctness and the feasibility of this approach, we have examined different variations of our method and evaluated various combinations of tracking sensors. We furthermore integrated this time synchronization method into our UBITRACK Mixed Reality tracking framework to provide facilities for calibration and real-time data alignment.
Resumo:
Interactive TV technology has been addressed in many previous works, but there is sparse research on the topic of interactive content broadcasting and how to support the production process. In this article, the interactive broadcasting process is broadly defined to include studio technology and digital TV applications at consumer set-top boxes. In particular, augmented reality studio technology employs smart-projectors as light sources and blends real scenes with interactive computer graphics that are controlled at end-user terminals. Moreover, TV producer-friendly multimedia authoring tools empower the development of novel TV formats. Finally, the support for user-contributed content raises the potential to revolutionize the hierarchical TV production process, by introducing the viewer as part of content delivery chain.
Resumo:
In this paper we present a model-based approach for real-time camera pose estimation in industrial scenarios. The line model which is used for tracking is generated by rendering a polygonal model and extracting contours out of the rendered scene. By un-projecting a point on the contour with the depth value stored in the z-buffer, the 3D coordinates of the contour can be calculated. For establishing 2D/3D correspondences the 3D control points on the contour are projected into the image and a perpendicular search for gradient maxima for every point on the contour is performed. Multiple hypotheses of 2D image points corresponding to a 3D control point make the pose estimation robust against ambiguous edges in the image.
Resumo:
Die Ergebnisse der Konstruktion können so aufbereitet werden, dass sie nach entsprechenden Berechnungen und Simulationen als virtuelle Prototypen zur Verfügen gestellt werden können. Die Möglichkeiten des Einsatzes virtueller Prototypen werden aufgezeigt. Der Unterschied zwischen virtuellen und realen Prototypen in Bezug auf die individuelle Wahrnehmung aufgrund der Sinnesmodalitäten wird erläutert. Die gegenwärtigen Grenzen der virtuellen Prototypen werden aufgezeigt.
Resumo:
Adding virtual objects to real environments plays an important role in todays computer graphics: Typical examples are virtual furniture in a real room and virtual characters in real movies. For a believable appearance, consistent lighting of the virtual objects is required. We present an augmented reality system that displays virtual objects with consistent illumination and shadows in the image of a simple webcam. We use two high dynamic range video cameras with fisheye lenses permanently recording the environment illumination. A sampling algorithm selects a few bright parts in one of the wide angle images and the corresponding points in the second camera image. The 3D position can then be calculated using epipolar geometry. Finally, the selected point lights are used in a multi pass algorithm to draw the virtual object with shadows. To validate our approach, we compare the appearance and shadows of the synthetic objects with real objects.
Resumo:
Image-based Relighting (IBRL) has recently attracted a lot of research interest for its ability to relight real objects or scenes, from novel illuminations captured in natural/synthetic environments. Complex lighting effects such as subsurface scattering, interreflection, shadowing, mesostructural self-occlusion, refraction and other relevant phenomena can be generated using IBRL. The main advantage of image-based graphics is that the rendering time is independent of scene complexity as the rendering is actually a process of manipulating image pixels, instead of simulating light transport. The goal of this paper is to provide a complete and systematic overview of the research in Imagebased Relighting. We observe that essentially all IBRL techniques can be broadly classified into three categories (Fig. 9), based on how the scene/illumination information is captured: Reflectance function-based, Basis function-based and Plenoptic function-based. We discuss the characteristics of each of these categories and their representative methods. We also discuss about the sampling density and types of light source(s), relevant issues of IBRL.
Resumo:
Manuelle Kommissioniersysteme sind aufgrund ihrer Flexibilität weit verbreitet, jedoch treten bei diesen auch vermehrt Kommissionierfehler auf, die weitreichende Folgen haben können. Deshalb werden zu Beginn des Beitrags die Auswirkungen und Ursachen von Kommissionierfehlern betrachtet. Der Fokus des Beitrags liegt auf der Darlegung von Möglichkeiten zur Fehlervermeidung. Dabei wird ein ganzheitlicher Ansatz verfolgt, der Maßnahmen aus den Handlungsfeldern Technik, Prozess, Organisation und Motivation umfasst. Als Ausblick auf zukünftige technische Hilfsmittel zur Fehlervermeidung wird das Augmented-Reality-unterstützte System Pick-by-Vision aus der Forschung vorgestellt, das wesentliche Vorteile existierender Techniken der Informationsbereitstellung vereinigt.
Resumo:
Der Beitrag fokussiert die Entwicklung, den Einsatz und die Nutzung von innovativen Technologien zur Unterstützung von Bildungsszenarien in Schule, Hochschule und Weiterbildung. Ausgehend von den verschiedenen Phasen des Corporate Learning, Social Learning, Mobile Learning und Intelligent Learning wird in einem ersten Abschnitt das Nutzungsverhalten von Technologien durch Kinder, Jugendliche und (junge) Erwachsene in Schule, Studium und Lehre betrachtet. Es folgt die Darstellung technologischer Entwicklungen auf Basis des Technology Life Cycle und die Konsequenzen von unterschiedlichen Entwicklungszuständen und Reifegraden von Technologien wie Content Learning Management, sozialen Netzwerken, mobilen Endgeräten, multidimensionalen und -modalen Räumen bis hin zu Anwendungen augmentierter Realität und des Internets der Dinge, Dienste und Daten für den Einsatz und die Nutzung in Bildungsszenarien. Nach der Darstellung von Anforderungen an digitale Technologien hinsichtlich Inhalte, Didaktik und Methodik wie etwa hinsichtlich der Erstellung von Inhalten, deren Wiederverwendung, Digitalisierung und Auffindbarkeit sowie Standards werden methodische Hinweise zur Nutzung digitaler Technologien zur Interaktion von Lernenden, von Lehrenden, sozialer Interaktion, kollaborativem Autorieren, Kommentierung, Evaluation und Begutachtung gegeben. Abschließend werden - differenziert für Schule und Hochschule - Erkenntnisse zu Rahmenbedingungen, Einflussgrößen, hemmenden und fördernden Faktoren sowie Herausförderungen bei der Einführung und nachhaltigen Implementation digitaler Technologien im schulischen Unterricht, in Lehre, Studium und Weiterbildung im Überblick zusammengefasst.