7 resultados para Topology-based methods

em Universidad de Alicante


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Plane model extraction from three-dimensional point clouds is a necessary step in many different applications such as planar object reconstruction, indoor mapping and indoor localization. Different RANdom SAmple Consensus (RANSAC)-based methods have been proposed for this purpose in recent years. In this study, we propose a novel method-based on RANSAC called Multiplane Model Estimation, which can estimate multiple plane models simultaneously from a noisy point cloud using the knowledge extracted from a scene (or an object) in order to reconstruct it accurately. This method comprises two steps: first, it clusters the data into planar faces that preserve some constraints defined by knowledge related to the object (e.g., the angles between faces); and second, the models of the planes are estimated based on these data using a novel multi-constraint RANSAC. We performed experiments in the clustering and RANSAC stages, which showed that the proposed method performed better than state-of-the-art methods.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Predicting accurate bond length alternations (BLAs) in long conjugated oligomers has been a significant challenge for electronic-structure methods for many decades, made particularly important by the close relationships between BLA and the rich optoelectronic properties of π-delocalized systems. Here, we test the accuracy of recently developed, and increasingly popular, double hybrid (DH) functionals, positioned at the top of Jacobs Ladder of DFT methods of increasing sophistication, computational cost, and accuracy, due to incorporation of MP2 correlation energy. Our test systems comprise oligomeric series of polyacetylene, polymethineimine, and polysilaacetylene up to six units long. MP2 calculations reveal a pronounced shift in BLAs between the 6-31G(d) basis set used in many studies of BLA to date and the larger cc-pVTZ basis set, but only modest shifts between cc-pVTZ and aug-cc-pVQZ results. We hence perform new reference CCSD(T)/cc-pVTZ calculations for all three series of oligomers against which we assess the performance of several families of DH functionals based on BLYP, PBE, and TPSS, along with lower-rung relatives including global- and range-separated hybrids. Our results show that DH functionals systematically improve the accuracy of BLAs relative to single hybrid functionals. xDH-PBE0 (N4 scaling using SOS-MP2) emerges as a DH functional rivaling the BLA accuracy of SCS-MP2 (N5 scaling), which was found to offer the best compromise between computational cost and accuracy the last time the BLA accuracy of DFT- and wave function-based methods was systematically investigated. Interestingly, xDH-PBE0 (XYG3), which differs to other DHs in that its MP2 term uses PBE0 (B3LYP) orbitals that are not self-consistent with the DH functional, is an outlier of trends of decreasing average BLA errors with increasing fractions of MP2 correlation and HF exchange.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

One of the main challenges to be addressed in text summarization concerns the detection of redundant information. This paper presents a detailed analysis of three methods for achieving such goal. The proposed methods rely on different levels of language analysis: lexical, syntactic and semantic. Moreover, they are also analyzed for detecting relevance in texts. The results show that semantic-based methods are able to detect up to 90% of redundancy, compared to only the 19% of lexical-based ones. This is also reflected in the quality of the generated summaries, obtaining better summaries when employing syntactic- or semantic-based approaches to remove redundancy.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A MATLAB-based computer code has been developed for the simultaneous wavelet analysis and filtering of several environmental time series, particularly focused on the analyses of cave monitoring data. The continuous wavelet transform, the discrete wavelet transform and the discrete wavelet packet transform have been implemented to provide a fast and precise time–period examination of the time series at different period bands. Moreover, statistic methods to examine the relation between two signals have been included. Finally, the entropy of curves and splines based methods have also been developed for segmenting and modeling the analyzed time series. All these methods together provide a user-friendly and fast program for the environmental signal analysis, with useful, practical and understandable results.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Since the beginning of 3D computer vision problems, the use of techniques to reduce the data to make it treatable preserving the important aspects of the scene has been necessary. Currently, with the new low-cost RGB-D sensors, which provide a stream of color and 3D data of approximately 30 frames per second, this is getting more relevance. Many applications make use of these sensors and need a preprocessing to downsample the data in order to either reduce the processing time or improve the data (e.g., reducing noise or enhancing the important features). In this paper, we present a comparison of different downsampling techniques which are based on different principles. Concretely, five different downsampling methods are included: a bilinear-based method, a normal-based, a color-based, a combination of the normal and color-based samplings, and a growing neural gas (GNG)-based approach. For the comparison, two different models have been used acquired with the Blensor software. Moreover, to evaluate the effect of the downsampling in a real application, a 3D non-rigid registration is performed with the data sampled. From the experimentation we can conclude that depending on the purpose of the application some kernels of the sampling methods can improve drastically the results. Bilinear- and GNG-based methods provide homogeneous point clouds, but color-based and normal-based provide datasets with higher density of points in areas with specific features. In the non-rigid application, if a color-based sampled point cloud is used, it is possible to properly register two datasets for cases where intensity data are relevant in the model and outperform the results if only a homogeneous sampling is used.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Software for video-based multi-point frequency measuring and mapping: http://hdl.handle.net/10045/53429

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The research described in this thesis was motivated by the need of a robust model capable of representing 3D data obtained with 3D sensors, which are inherently noisy. In addition, time constraints have to be considered as these sensors are capable of providing a 3D data stream in real time. This thesis proposed the use of Self-Organizing Maps (SOMs) as a 3D representation model. In particular, we proposed the use of the Growing Neural Gas (GNG) network, which has been successfully used for clustering, pattern recognition and topology representation of multi-dimensional data. Until now, Self-Organizing Maps have been primarily computed offline and their application in 3D data has mainly focused on free noise models, without considering time constraints. It is proposed a hardware implementation leveraging the computing power of modern GPUs, which takes advantage of a new paradigm coined as General-Purpose Computing on Graphics Processing Units (GPGPU). The proposed methods were applied to different problem and applications in the area of computer vision such as the recognition and localization of objects, visual surveillance or 3D reconstruction.