956 resultados para SLAM RGB-D SlamDunk Android 3D mobile
Resumo:
En este trabajo se presenta Capaware, una plataforma de software libre para el desarrollo de aplicaciones geográficas 3D multicapa, que surge a partir de la iniciativa del Instituto Tecnológico de Canarias en colaboración con la Universidad de Las Palmas de Gran Canaria. Este entorno simplifica la creación de aplicaciones 3D sobre territorios geográficos extensos, disponiendo de una herramienta muy visual que aporta un nuevo punto de vista muy importante para una toma de decisiones eficaz. Capaware proporciona una interfaz fácil de usar y muy flexible que simplifica el desarrollo de nuevas aplicaciones, permitiéndonos crear rápidamente entornos virtuales con múltiples capas de información sobre el terreno. Con las capacidades clásicas de un Sistema de Información Geográfica (SIG), Capaware permite actualmente la carga de capas WMS sobre entornos 3D, añadir objetos 3D sobre el terreno, y visualizar elementos dinámicos, ofreciendo una nueva perspectiva de la información analizada. Así mismo, podemos administrar las capas de recursos y elementos que se pueden representar sobre la zona geográfica en cuestión. (...)
Resumo:
Inicialmente integrada en el piloto de gvSIG Mobile, la librería libLocation tiene como objetivo dotar a los proyectos gvSIG Desktop y gvSIG Mobile un acceso transparente a fuentes de localización. La librería se fundamenta en las especificaciones JSR-179 -API de localización para J2ME- y JSR-293 -API de localización para J2ME v2.0-, proporcionando una interfaz uniforme a diferentes fuentes de localización, mediante funciones de alto nivel. Asimismo, se extiende la funcionalidad de estas APIs para permitir la gestión de datos específicos del tipo de fuente de localización y el ajuste de parámetros de bajo nivel, además de incorporar métodos de localización adicionales, como la aplicación de correcciones vía protocolo NTRIP. La librería libLocation está actualmente en proceso de desarrollo y será publicada y liberada junto con la versión definitiva de gvSIG Mobile. Junto con libLocation se están desarrollando extensiones que permiten el acceso a esta librería desde gvSIG Desktop y gvSIG Mobile
Resumo:
gvSIG Mobile, la versión de gvSIG para dispositivos móviles presenta su nueva versión que incluye las esperadas funcionalidades de creación de nuevas entidades geográficas y utilización de formularios personalizados para edición de datos, además de nuevos formatos de datos vectoriales (GML, KML, GPX) y sistemas de referencia. Funcionalidades que se suman a las capacidades de visor de cartografía (ECW, SHP, WMS) y sistema de localización mediante GPS que ya posee. gvSIG Mobile está siendo desarrollado por Prodevelop, la Universitat de València e Iver para la Conselleria d’Infraestructures i Transport de la Generalitat Valenciana y se distribuye con una licencia GPL
Resumo:
Desde la aparición de Google Earth se ha desencadenado una proliferación de aplicaciones de “visores 3D” o globos virtuales facilitando a los usuarios la navegación y exploración de cualquier localización de la Tierra en formato 3D. A estas alturas, nadie puede negar la repercusión mediática de este tipo de aplicaciones y, sobretodo, el impacto diseminador de conceptos geográficos (capas, mapas, etc.) que ha tenido sobre la gran mayoría de usuarios, normalmente no especializados en el dominio geográfico. En el ámbito del software libre, a los originales World Wind (.NET) y Ossimplanet se han venido uniendo iniciativas en otras plataformas como Marble KDE, nuevas versiones de WordWind (Java SDK), a las que últimamente se están añadiendo iniciativas españolas dignas de consideración, como el Capaware canario y el módulo 3D de gvSIG. En este artículo se analiza el estado del desarrollo de estos proyectos, así como los estándares y formatos específicos que se están elaborando en el OGC para el ámbito específico de la información geográfica tridimensional (modelado de escenarios, objetos, terreno o cachés)
Resumo:
The estimation of camera egomotion is a well established problem in computer vision. Many approaches have been proposed based on both the discrete and the differential epipolar constraint. The discrete case is mainly used in self-calibrated stereoscopic systems, whereas the differential case deals with a unique moving camera. The article surveys several methods for mobile robot egomotion estimation covering more than 0.5 million samples using synthetic data. Results from real data are also given
Resumo:
A technique for simultaneous localisation and mapping (SLAM) for large scale scenarios is presented. This solution is based on the use of independent submaps of a limited size to map large areas. In addition, a global stochastic map, containing the links between adjacent submaps, is built. The information in both levels is corrected every time a loop is closed: local maps are updated with the information from overlapping maps, and the global stochastic map is optimised by means of constrained minimisation
Resumo:
This paper proposes a hybrid coordination method for behavior-based control architectures. The hybrid method takes advantages of the robustness and modularity in competitive approaches as well as optimized trajectories in cooperative ones. This paper shows the feasibility of applying this hybrid method with a 3D-navigation to an autonomous underwater vehicle (AUV). The behaviors are learnt online by means of reinforcement learning. A continuous Q-learning implemented with a feed-forward neural network is employed. Realistic simulations were carried out. The results obtained show the good performance of the hybrid method on behavior coordination as well as the convergence of the behaviors
Resumo:
The accuracy of a 3D reconstruction using laser scanners is significantly determined by the detection of the laser stripe. Since the energy pattern of such a stripe corresponds to a Gaussian profile, it makes sense to detect the point of maximum light intensity (or peak) by computing the zero-crossing point of the first derivative of such Gaussian profile. However, because noise is present in every physical process, such as electronic image formation, it is not sensitive to perform the derivative of the image of the stripe in almost any situation, unless a previous filtering stage is done. Considering that stripe scanning is an inherently row-parallel process, every row of a given image must be processed independently in order to compute its corresponding peak position in the row. This paper reports on the use of digital filtering techniques in order to cope with the scanning of different surfaces with different optical properties and different noise levels, leading to the proposal of a more accurate numerical peak detector, even at very low signal-to-noise ratios
Resumo:
This paper presents a complete solution for creating accurate 3D textured models from monocular video sequences. The methods are developed within the framework of sequential structure from motion, where a 3D model of the environment is maintained and updated as new visual information becomes available. The camera position is recovered by directly associating the 3D scene model with local image observations. Compared to standard structure from motion techniques, this approach decreases the error accumulation while increasing the robustness to scene occlusions and feature association failures. The obtained 3D information is used to generate high quality, composite visual maps of the scene (mosaics). The visual maps are used to create texture-mapped, realistic views of the scene
Resumo:
We present a computer vision system that associates omnidirectional vision with structured light with the aim of obtaining depth information for a 360 degrees field of view. The approach proposed in this article combines an omnidirectional camera with a panoramic laser projector. The article shows how the sensor is modelled and its accuracy is proved by means of experimental results. The proposed sensor provides useful information for robot navigation applications, pipe inspection, 3D scene modelling etc
Resumo:
This research work deals with the problem of modeling and design of low level speed controller for the mobile robot PRIM. The main objective is to develop an effective educational tool. On one hand, the interests in using the open mobile platform PRIM consist in integrating several highly related subjects to the automatic control theory in an educational context, by embracing the subjects of communications, signal processing, sensor fusion and hardware design, amongst others. On the other hand, the idea is to implement useful navigation strategies such that the robot can be served as a mobile multimedia information point. It is in this context, when navigation strategies are oriented to goal achievement, that a local model predictive control is attained. Hence, such studies are presented as a very interesting control strategy in order to develop the future capabilities of the system
Resumo:
This paper is focused on the robot mobile platform PRIM (platform robot information multimedia). This robot has been made in order to cover two main needs of our group, on one hand the need for a full open mobile robotic platform that is very useful in fulfilling the teaching and research activity of our school community, and on the other hand with the idea of introducing an ethical product which would be useful as mobile multimedia information point as a service tool. This paper introduces exactly how the system is made up and explains just what the philosophy is behind this work. The navigation strategies and sensor fusion, where machine vision system is the most important one, are oriented towards goal achievement and are the key to the behaviour of the robot
Resumo:
This paper presents the use of a mobile robot platform as an innovative educational tool in order to promote and integrate different curriculum knowledge. Hence, it is presented the acquired experience within a summer course named ldquoapplied mobile roboticsrdquo. The main aim of the course is to integrate different subjects as electronics, programming, architecture, perception systems, communications, control and trajectory planning by using the educational open mobile robot platform PRIM. The summer course is addressed to a wide range of student profiles. However, it is of special interests to the students of electrical and computer engineering around their final academic year. The summer course consists of the theoretical and laboratory sessions, related to the following topics: design & programming of electronic devices, modelling and control systems, trajectory planning and control, and computer vision systems. Therefore, the clues for achieving a renewed path of progress in robotics are the integration of several knowledgeable fields, such as computing, communications, and control sciences, in order to perform a higher level reasoning and use decision tools with strong theoretical base
Resumo:
This work extends a previously developed research concerning about the use of local model predictive control in differential driven mobile robots. Hence, experimental results are presented as a way to improve the methodology by considering aspects as trajectory accuracy and time performance. In this sense, the cost function and the prediction horizon are important aspects to be considered. The aim of the present work is to test the control method by measuring trajectory tracking accuracy and time performance. Moreover, strategies for the integration with perception system and path planning are briefly introduced. In this sense, monocular image data can be used to plan safety trajectories by using goal attraction potential fields
Resumo:
This paper presents the implementation details of a coded structured light system for rapid shape acquisition of unknown surfaces. Such techniques are based on the projection of patterns onto a measuring surface and grabbing images of every projection with a camera. Analyzing the pattern deformations that appear in the images, 3D information of the surface can be calculated. The implemented technique projects a unique pattern so that it can be used to measure moving surfaces. The structure of the pattern is a grid where the color of the slits are selected using a De Bruijn sequence. Moreover, since both axis of the pattern are coded, the cross points of the grid have two codewords (which permits to reconstruct them very precisely), while pixels belonging to horizontal and vertical slits have also a codeword. Different sets of colors are used for horizontal and vertical slits, so the resulting pattern is invariant to rotation. Therefore, the alignment constraint between camera and projector considered by a lot of authors is not necessary