875 resultados para 3D programming
Resumo:
Image stitching is the process of joining several images to obtain a bigger view of a scene. It is used, for example, in tourism to transmit to the viewer the sensation of being in another place. I am presenting an inexpensive solution for automatic real time video and image stitching with two web cameras as the video/image sources. The proposed solution relies on the usage of several markers in the scene as reference points for the stitching algorithm. The implemented algorithm is divided in four main steps, the marker detection, camera pose determination (in reference to the markers), video/image size and 3d transformation, and image translation. Wii remote controllers are used to support several steps in the process. The built‐in IR camera provides clean marker detection, which facilitates the camera pose determination. The only restriction in the algorithm is that markers have to be in the field of view when capturing the scene. Several tests where made to evaluate the final algorithm. The algorithm is able to perform video stitching with a frame rate between 8 and 13 fps. The joining of the two videos/images is good with minor misalignments in objects at the same depth of the marker,misalignments in the background and foreground are bigger. The capture process is simple enough so anyone can perform a stitching with a very short explanation. Although real‐time video stitching can be achieved by this affordable approach, there are few shortcomings in current version. For example, contrast inconsistency along the stitching line could be reduced by applying a color correction algorithm to every source videos. In addition, the misalignments in stitched images due to camera lens distortion could be eased by optical correction algorithm. The work was developed in Apple’s Quartz Composer, a visual programming environment. A library of extended functions was developed using Xcode tools also from Apple.
Resumo:
Este trabalho foi realizado na Universidade da Madeira, no âmbito do Mestrado em Engenharia Informática, tendo como título “Implementação de um sistema autónomo para a geração de visualizações 3D de dados ambientais”. A visualização 3D tem vindo a ganhar terreno em áreas como o entretenimento, medicina, arquitetura e desenho de equipamentos, entre outros. Relativamente à visualização de dados ambientais (oceano e atmosfera) em 3D, esta é uma área pouco explorada. Como tal, foi proposto o estudo e criação de um sistema autónomo capaz de gerar imagens 3D de dados ambientais e disponibilizar as mesmas na web. O estudo iniciou-se com testes às ferramentas que permitem a visualização 3D de dados ambientais, de forma a escolher a que mais se adequa ao sistema a implementar. Após diversos testes às várias ferramentas, a que mais se destacou foi o IDV (Integrated Data Visualization), pelas suas funcionalidades e capacidade de ser executado de forma automática. A implementação do sistema foi efetuada ao longo de diversas etapas: pré-processamento dos dados; escolha das visualizações a disponibilizar na web; escrita do script responsável pelo processo; criação das páginas web para visualização das imagens e implementação de todo o sistema em ambiente Linux. Este sistema foi desenvolvido tendo em conta a adição de novas visualizações, assim como a alteração das visualizações existentes. A possibilidade de adicionar/alterar visualizações de forma simples também foi tida em conta, de forma a não ser necessário reestruturar todo o sistema.
Resumo:
Tabletop computers featuring multi-touch input and object tracking are a common platform for research on Tangible User Interfaces (also known as Tangible Interaction). However, such systems are confined to sensing activity on the tabletop surface, disregarding the rich and relatively unexplored interaction canvas above the tabletop. This dissertation contributes with tCAD, a 3D modeling tool combining fiducial marker tracking, finger tracking and depth sensing in a single system. This dissertation presents the technical details of how these features were integrated, attesting to its viability through the design, development and early evaluation of the tCAD application. A key aspect of this work is a description of the interaction techniques enabled by merging tracked objects with direct user input on and above a table surface.
Resumo:
A constraint satisfaction problem is a classical artificial intelligence paradigm characterized by a set of variables (each variable with an associated domain of possible values), and a set of constraints that specify relations among subsets of these variables. Solutions are assignments of values to all variables that satisfy all the constraints. Many real world problems may be modelled by means of constraints. The range of problems that can use this representation is very diverse and embraces areas like resource allocation, scheduling, timetabling or vehicle routing. Constraint programming is a form of declarative programming in the sense that instead of specifying a sequence of steps to execute, it relies on properties of the solutions to be found, which are explicitly defined by constraints. The idea of constraint programming is to solve problems by stating constraints which must be satisfied by the solutions. Constraint programming is based on specialized constraint solvers that take advantage of constraints to search for solutions. The success and popularity of complex problem solving tools can be greatly enhanced by the availability of friendly user interfaces. User interfaces cover two fundamental areas: receiving information from the user and communicating it to the system; and getting information from the system and deliver it to the user. Despite its potential impact, adequate user interfaces are uncommon in constraint programming in general. The main goal of this project is to develop a graphical user interface that allows to, intuitively, represent constraint satisfaction problems. The idea is to visually represent the variables of the problem, their domains and the problem constraints and enable the user to interact with an adequate constraint solver to process the constraints and compute the solutions. Moreover, the graphical interface should be capable of configure the solver’s parameters and present solutions in an appealing interactive way. As a proof of concept, the developed application – GraphicalConstraints – focus on continuous constraint programming, which deals with real valued variables and numerical constraints (equations and inequalities). RealPaver, a state-of-the-art solver in continuous domains, was used in the application. The graphical interface supports all stages of constraint processing, from the design of the constraint network to the presentation of the end feasible space solutions as 2D or 3D boxes.
Resumo:
Humans can perceive three dimension, our world is three dimensional and it is becoming increasingly digital too. We have the need to capture and preserve our existence in digital means perhaps due to our own mortality. We have also the need to reproduce objects or create small identical objects to prototype, test or study them. Some objects have been lost through time and are only accessible through old photographs. With robust model generation from photographs we can use one of the biggest human data sets and reproduce real world objects digitally and physically with printers. What is the current state of development in three dimensional reconstruction through photographs both in the commercial world and in the open source world? And what tools are available for a developer to build his own reconstruction software? To answer these questions several pieces of software were tested, from full commercial software packages to open source small projects, including libraries aimed at computer vision. To bring to the real world the 3D models a 3D printer was built, tested and analyzed, its problems and weaknesses evaluated. Lastly using a computer vision library a small software with limited capabilities was developed.
Resumo:
The seismic method is of extreme importance in geophysics. Mainly associated with oil exploration, this line of research focuses most of all investment in this area. The acquisition, processing and interpretation of seismic data are the parts that instantiate a seismic study. Seismic processing in particular is focused on the imaging that represents the geological structures in subsurface. Seismic processing has evolved significantly in recent decades due to the demands of the oil industry, and also due to the technological advances of hardware that achieved higher storage and digital information processing capabilities, which enabled the development of more sophisticated processing algorithms such as the ones that use of parallel architectures. One of the most important steps in seismic processing is imaging. Migration of seismic data is one of the techniques used for imaging, with the goal of obtaining a seismic section image that represents the geological structures the most accurately and faithfully as possible. The result of migration is a 2D or 3D image which it is possible to identify faults and salt domes among other structures of interest, such as potential hydrocarbon reservoirs. However, a migration fulfilled with quality and accuracy may be a long time consuming process, due to the mathematical algorithm heuristics and the extensive amount of data inputs and outputs involved in this process, which may take days, weeks and even months of uninterrupted execution on the supercomputers, representing large computational and financial costs, that could derail the implementation of these methods. Aiming at performance improvement, this work conducted the core parallelization of a Reverse Time Migration (RTM) algorithm, using the parallel programming model Open Multi-Processing (OpenMP), due to the large computational effort required by this migration technique. Furthermore, analyzes such as speedup, efficiency were performed, and ultimately, the identification of the algorithmic scalability degree with respect to the technological advancement expected by future processors
Resumo:
This work presents an optimization technique based on structural topology optimization methods, TOM, designed to solve problems of thermoelasticity 3D. The presented approach is based on the adjoint method of sensitivity analysis unified design and is intended to loosely coupled thermomechanical problems. The technique makes use of analytical expressions of sensitivities, enabling a reduction in the computational cost through the use of a coupled field adjoint equation, defined in terms the of temperature and displacement fields. The TOM used is based on the material aproach. Thus, to make the domain is composed of a continuous distribution of material, enabling the use of classical models in nonlinear programming optimization problem, the microstructure is considered as a porous medium and its constitutive equation is a function only of the homogenized relative density of the material. In this approach, the actual properties of materials with intermediate densities are penalized based on an artificial microstructure model based on the SIMP (Solid Isotropic Material with Penalty). To circumvent problems chessboard and reduce dependence on layout in relation to the final optimal initial mesh, caused by problems of numerical instability, restrictions on components of the gradient of relative densities were applied. The optimization problem is solved by applying the augmented Lagrangian method, the solution being obtained by applying the finite element method of Galerkin, the process of approximation using the finite element Tetra4. This element has the ability to interpolate both the relative density and the displacement components and temperature. As for the definition of the problem, the heat load is assumed in steady state, i.e., the effects of conduction and convection of heat does not vary with time. The mechanical load is assumed static and distributed
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
This article proposes a method for 3D road extraction from a stereopair of aerial images. The dynamic programming (DP) algorithm is used to carry out the optimization process in the object-space, instead of usually doing it in the image-space such as the DP traditional methodologies. This means that road centerlines are directly traced in the object-space, implying that a mathematical relationship is necessary to connect road points in object and image-space. This allows the integration of radiometric information from images into the associate mathematical road model. As the approach depends on an initial approximation of each road, it is necessary a few seed points to coarsely describe the road. Usually, the proposed method allows good results to be obtained, but large anomalies along the road can disturb its performance. Therefore, the method can be used for practical application, although it is expected some kind of local manual edition of the extracted road centerline.
Resumo:
fit the context of normalized variable formulation (NVF) of Leonard and total variation diminishing (TVD) constraints of Harten. this paper presents an extension of it previous work by the authors for solving unsteady incompressible flow problems. The main contributions of the paper are threefold. First, it presents the results of the development and implementation of a bounded high order upwind adaptative QUICKEST scheme in the 3D robust code (Freeflow), for the numerical solution of the full incompressible Navier-Stokes equations. Second, it reports numerical simulation results for 1D hock tube problem, 2D impinging jet and 2D/3D broken clam flows. Furthermore, these results are compared with existing analytical and experimental data. and third, it presents the application of the numerical method for solving 3D free surface flow problems. (C) 2007 IMACS. Published by Elsevier B.V. All rights reserved,
Resumo:
A contaminated site from a downstream municipal solid waste disposal site in Brazil was investigated by using a 3D resistivity and induced polarization (IP) imaging technique. This investigation purpose was to detect and delineate contamination plume produced by wastes. The area was selected based on previous geophysical investigations, and chemical analyses carried out in the site, indicating the presence of a contamination plume in the area. Resistivity model has successfully imaged waste presence (rho < 20 Omega m), water table depth, and groundwater flow direction. A conductive anomaly (rho < 20 Omega m) outside wastes placement was interpreted as a contamination plume. Chargeability model was also able to imaging waste presence (m > 31 mV/V), water table depth, and groundwater flow direction. A higher chargeability zone (m > 31 mV/V) outside wastes placement and following conductive anomaly was interpreted as a contamination plume. Normalized chargeability (MN = m/rho) confirmed polarizable zone, which could be an effect of a salinity increase (contamination plume), and the clay presence in the environment.
Resumo:
This paper describes the development and solution of binary integer formulations for production scheduling problems in market-driven foundries. This industrial sector is comprised of small and mid-sized companies with little or no automation, working with diversified production, involving several different metal alloy specifications in small tailor-made product lots. The characteristics and constraints involved in a typical production environment at these industries challenge the formulation of mathematical programming models that can be computationally solved when considering real applications. However, despite the interest on the part of these industries in counting on effective methods for production scheduling, there are few studies available on the subject. The computational tests prove the robustness and feasibility of proposed models in situations analogous to those found in production scheduling at the analyzed industrial sector. (C) 2010 Elsevier Ltd. All rights reserved.