887 resultados para 3D Computer Graphics
Resumo:
Monte Carlo integration is firmly established as the basis for most practical realistic image synthesis algorithms because of its flexibility and generality. However, the visual quality of rendered images often suffers from estimator variance, which appears as visually distracting noise. Adaptive sampling and reconstruction algorithms reduce variance by controlling the sampling density and aggregating samples in a reconstruction step, possibly over large image regions. In this paper we survey recent advances in this area. We distinguish between “a priori” methods that analyze the light transport equations and derive sampling rates and reconstruction filters from this analysis, and “a posteriori” methods that apply statistical techniques to sets of samples to drive the adaptive sampling and reconstruction process. They typically estimate the errors of several reconstruction filters, and select the best filter locally to minimize error. We discuss advantages and disadvantages of recent state-of-the-art techniques, and provide visual and quantitative comparisons. Some of these techniques are proving useful in real-world applications, and we aim to provide an overview for practitioners and researchers to assess these approaches. In addition, we discuss directions for potential further improvements.
Resumo:
We present a novel algorithm to reconstruct high-quality images from sampled pixels and gradients in gradient-domain rendering. Our approach extends screened Poisson reconstruction by adding additional regularization constraints. Our key idea is to exploit local patches in feature images, which contain per-pixels normals, textures, position, etc., to formulate these constraints. We describe a GPU implementation of our approach that runs on the order of seconds on megapixel images. We demonstrate a significant improvement in image quality over screened Poisson reconstruction under the L1 norm. Because we adapt the regularization constraints to the noise level in the input, our algorithm is consistent and converges to the ground truth.
Resumo:
With the ongoing shift in the computer graphics industry toward Monte Carlo rendering, there is a need for effective, practical noise-reduction techniques that are applicable to a wide range of rendering effects and easily integrated into existing production pipelines. This course surveys recent advances in image-space adaptive sampling and reconstruction algorithms for noise reduction, which have proven very effective at reducing the computational cost of Monte Carlo techniques in practice. These approaches leverage advanced image-filtering techniques with statistical methods for error estimation. They are attractive because they can be integrated easily into conventional Monte Carlo rendering frameworks, they are applicable to most rendering effects, and their computational overhead is modest.
Resumo:
Mass estimates for Late Miocene and Pliocene (8.6-3.25 Ma) Discoaster species and Sphenolithus are determined using samples of the equatorial Atlantic (Ceara Rise: ODP Site 927). Based on morphometric measurements, 3D computer models were created for 11 Discoaster species and their volumes calculated. From these, shape factors (ks) were derived to allow calculation of mass for different-sized discoasters and Sphenolithus abies. The mass estimates were then used to calculate the contribution of nannofossils to the total nannofossil carbonate. The discoaster contribution ranges from 10% to 40%, with a decreasing trend through the investigated interval. However, our estimates of total nannofossil carbonate from size-corrected abundance data are consistently 30-50% lower than estimates from grain-size measurement; this suggests that data based on mass estimates need to be interpreted with caution.
Resumo:
The analysis of time-dependent data is an important problem in many application domains, and interactive visualization of time-series data can help in understanding patterns in large time series data. Many effective approaches already exist for visual analysis of univariate time series supporting tasks such as assessment of data quality, detection of outliers, or identification of periodically or frequently occurring patterns. However, much fewer approaches exist which support multivariate time series. The existence of multiple values per time stamp makes the analysis task per se harder, and existing visualization techniques often do not scale well. We introduce an approach for visual analysis of large multivariate time-dependent data, based on the idea of projecting multivariate measurements to a 2D display, visualizing the time dimension by trajectories. We use visual data aggregation metaphors based on grouping of similar data elements to scale with multivariate time series. Aggregation procedures can either be based on statistical properties of the data or on data clustering routines. Appropriately defined user controls allow to navigate and explore the data and interactively steer the parameters of the data aggregation to enhance data analysis. We present an implementation of our approach and apply it on a comprehensive data set from the field of earth bservation, demonstrating the applicability and usefulness of our approach.
Resumo:
The analysis of research data plays a key role in data-driven areas of science. Varieties of mixed research data sets exist and scientists aim to derive or validate hypotheses to find undiscovered knowledge. Many analysis techniques identify relations of an entire dataset only. This may level the characteristic behavior of different subgroups in the data. Like automatic subspace clustering, we aim at identifying interesting subgroups and attribute sets. We present a visual-interactive system that supports scientists to explore interesting relations between aggregated bins of multivariate attributes in mixed data sets. The abstraction of data to bins enables the application of statistical dependency tests as the measure of interestingness. An overview matrix view shows all attributes, ranked with respect to the interestingness of bins. Complementary, a node-link view reveals multivariate bin relations by positioning dependent bins close to each other. The system supports information drill-down based on both expert knowledge and algorithmic support. Finally, visual-interactive subset clustering assigns multivariate bin relations to groups. A list-based cluster result representation enables the scientist to communicate multivariate findings at a glance. We demonstrate the applicability of the system with two case studies from the earth observation domain and the prostate cancer research domain. In both cases, the system enabled us to identify the most interesting multivariate bin relations, to validate already published results, and, moreover, to discover unexpected relations.
Resumo:
lnfográfica (en inglés Computer Graphics) es la parte de la Informática que se especializa en las técnicas de entrada y salida de datos en forma gráfica. Es sabido que la mente humana digiere difícilmente los datos presentados en forma numérica tabular, por lo que las técnicas infográficas vienen a añadir uno de los eslabones, más importantes en la cadena de acercamiento del ordenador al hombre, desde que se creó la programación automática.
Resumo:
Validating modern oceanographic theories using models produced through stereo computer vision principles has recently emerged. Space-time (4-D) models of the ocean surface may be generated by stacking a series of 3-D reconstructions independently generated for each time instant or, in a more robust manner, by simultaneously processing several snapshots coherently in a true ?4-D reconstruction.? However, the accuracy of these computer-vision-generated models is subject to the estimations of camera parameters, which may be corrupted under the influence of natural factors such as wind and vibrations. Therefore, removing the unpredictable errors of the camera parameters is necessary for an accurate reconstruction. In this paper, we propose a novel algorithm that can jointly perform a 4-D reconstruction as well as correct the camera parameter errors introduced by external factors. The technique is founded upon variational optimization methods to benefit from their numerous advantages: continuity of the estimated surface in space and time, robustness, and accuracy. The performance of the proposed algorithm is tested using synthetic data produced through computer graphics techniques, based on which the errors of the camera parameters arising from natural factors can be simulated.
Resumo:
El período de la Historia comprendido entre 1570 y 1620 nos ha dejado un importante conjunto de documentos relacionados con la construcción naval en la Península Ibérica. En una época convulsa en la que los reinos de España y Portugal se aglutinaron bajo una misma Corona, surgen una serie de manuscritos, libros y leyes que reflejan la creciente preocupación de la sociedad por el tema naval. Entre sus páginas encontramos las descripciones del proceso constructivo de los buques que sus autores consideraban más significativos para las demandas que se planteaban en ese momento. Este proceso que combinaba generación de formas y construcción del buque provenía de una secular tradición nacida en el Mediterráneo. Mediante reglas geométricas sencillas, el constructor naval trazaba las secciones centrales y el perfil de la nao, quedando los extremos de la misma (hasta más de la mitad de la eslora) a su buen hacer y experiencia. Las herramientas informáticas de generación de superficies mediante NURBs (Non- Uniform Rational B-spline) permiten reconstruir las formas de los navíos reproduciendo con fiabilidad las carenas de los mismos a partir de los documentos de la época. Mediante un estudio detallado de interpretación de los textos y transcribiendo los procesos, llegamos a obtener con un buen grado de precisión las carenas de los buques descritos en sus páginas. A partir de ahí y mediante el análisis cualitativo y cuantitativo de los parámetros obtenidos es posible valorar si las soluciones representadas por los barcos respondían a las preguntas planteadas por sus autores , la influencia de factores externos a la construcción naval tales como las regulaciones del Estado o identificar su relación con el germen y la expansión de la teoría que ha determinado los efectos de la Ciencia en la Arquitectura Naval. Comenzando por la nao veneciana de 1550, heredera de la secular tradición constructiva mediterránea, hasta llegar a las Reales Ordenanzas promulgadas en 1618, se reproducen hasta nueve carenas a partir de otros tantos documentos, se dibujan sus planos de formas y se exportan para su análisis hidrostático. El trabajo requiere la realización de otros estudios en paralelo necesarios para entender aquellos factores que formaron parte del desarrollo tecnológico naval como son, las unidades de medida en uso en los astilleros, los distintos sistemas de arqueo impuestos por la Corona y la representación de los diferentes instrumentos geométricos de modificación de los parámetros de diseño. A lo largo del trabajo se dan respuesta a interrogantes planteados por la arqueología en relación con el desarrollo de la arquitectura naval poniendo en evidencia que durante este período quedaron establecidos los fundamentos teórico-prácticos de lo que más adelante se convirtió en la ciencia de la ingeniería naval y se plantean nuevos retos para aquellos que deseen continuar la apasionante tarea de la investigación científica de nuestra historia. ABSTRACT The period of the History comprised between 1570 and 1620 has left an important set of shipbuilding documents in the Iberian Peninsula. In a turbulent time in which the kingdoms of Spain and Portugal were ruled under the same Crown, manuscripts, books and laws that reflect the growing concern of society for the naval theme arose. We found among their pages shipbuilding process descriptions of the more relevant vessels that responded to claims that arose at that time. This process brought together hull generation and shipbuilding and came from a secular tradition born in the Mediterranean. By means of simple geometric rules, the shipbuilder traced the central sections and profile of the ship, leaving the ends thereof (almost half of the length) to its good performance and experience. 3D computer modelling software by NURBs (Non-Uniform Rational B-spline) surfaces helps to reconstruct ships hulls from contemporary documents. Through a detailed texts interpretation and transcription processes, we manage to reach with a good degree of accuracy the ship hulls described in its pages. From there and through qualitative and quantitative analysis of the parameters obtained we can assess whether the solutions represented by ships gave response to the questions raised by the authors, the influence of external factors such as shipbuilding state regulations or identify their relationship to the origin and expansion of the theory that has determined the effects of Science in Naval Architecture. From the 1550 Venetian nao, inheritor of the secular Mediterranean building tradition, to the Royal Ordinances enacted in 1618, as nine hulls are reproduced, their line drawings are traced and exported for analysis hydrostatic. Further studies are needed to understand the factors that were part of shipbuilding technology development as the units of measure in use in shipyards, the different official regulations for calculating ship tonnage and the different geometric instruments to amend the design parameters. The work gives response to questions raised by archaeology in relation to the development of naval architecture highlighting that during this period were established the theoretical and practical foundations of what later became the science of naval engineering and raising new challenges for those wishing to continue the exciting task of scientific research of our History.
Resumo:
Technological innovation in all areas has led to the appearance in recent years of new metallic and pearlescent materials, yet no exhaustive studies have been conducted to assess their colorimetric capabilities. The chromatic variability of these special-effect pigments may largely be due to the three-dimensional effect of their curved shapes and orientations when they are directionally or diffusely illuminated. Our study examines goniochromatic colors using the optimal colors (MacAdam limits) associated with normal colors (photometric scale of relative spectral reflectance from 0 to 1) under certain conventional illuminants and other light sources. From a database of 91 metallic and interference samples and using a multi-gonio-spectrophotometer, we analyzed samples with lightness values of more than 100 and others with lightness values of less than 100, but with higher chromaticities than optimal colors, which places them beyond the MacAdam limits. Our study thus demonstrates the existence of chromatic perceptions beyond the normal solid color associated with these materials and independent of the light source. The challenge for future research, therefore, is to replicate and render these color appearances in current and future color reproduction technologies for computer graphics.
Resumo:
"COO-1469-0067."
Resumo:
Issued also as thesis (M.S.) University of Illinois.
Resumo:
Photocopy of typescript.
Resumo:
Thesis (M.S.)--University of Illinois at Urbana-Champaign.
Resumo:
"AEC report no. C00-2383-002."