15 resultados para Hilbert schemes of points Poincaré polynomial Betti numbers Goettsche formula
em Universidad de Alicante
Resumo:
In this paper, it is showed that, given an integer number n ≥ 2, each zero of an exponential polynomial of the form w1az1+w2az2+⋯+wnazn, with non-null complex numbers w 1,w 2,…,w n and a 1,a 2,…,a n , produces analytic solutions of the functional equation w 1 f(a 1 z) + w 2 f(a 2 z) + ... + w n f(a n z) = 0 on certain domains of C, which represents an extension of some existing results in the literature on this functional equation for the case of positive coefficients a j and w j.
Resumo:
In this paper we provide the proof of a practical point-wise characterization of the set RP defined by the closure set of the real projections of the zeros of an exponential polynomial P(z) = Σn j=1 cjewjz with real frequencies wj linearly independent over the rationals. As a consequence, we give a complete description of the set RP and prove its invariance with respect to the moduli of the c′ js, which allows us to determine exactly the gaps of RP and the extremes of the critical interval of P(z) by solving inequations with positive real numbers. Finally, we analyse the converse of this result of invariance.
Resumo:
Context. 4U 1538−52, an absorbed high mass X-ray binary with an orbital period of ~3.73 days, shows moderate orbital intensity modulations with a low level of counts during the eclipse. Several models have been proposed to explain the accretion at different orbital phases by a spherically symmetric stellar wind from the companion. Aims. The aim of this work is to study both the light curve and orbital phase spectroscopy of this source in the long term. In particular, we study the folded light curve and the changes in the spectral parameters with orbital phase to analyse the stellar wind of QV Nor, the mass donor of this binary system. Methods. We used all the observations made from the Gas Slit Camera on board MAXI of 4U 1538−52 covering many orbits continuously. We obtained the good interval times for all orbital phase ranges, which were the input for extracting our data. We estimated the orbital period of the system and then folded the light curves, and we fitted the X-ray spectra with the same model for every orbital phase spectrum. We also extracted the averaged spectrum of all the MAXI data available. Results. The MAXI spectra in the 2–20 keV energy range were fitted with an absorbed Comptonisation of cool photons on hot electrons. We found a strong orbital dependence of the absorption column density but neither the fluorescence iron emission line nor low energy excess were needed to fit the MAXI spectra. The variation in the spectral parameters over the binary orbit were used to examine the mode of accretion onto the neutron star in 4U 1538−52. We deduce a best value of Ṁ/v∞ = 0.65 × 10-9M⊙ yr-1/ (km s-1) for QV Nor.
Resumo:
The Iterative Closest Point algorithm (ICP) is commonly used in engineering applications to solve the rigid registration problem of partially overlapped point sets which are pre-aligned with a coarse estimate of their relative positions. This iterative algorithm is applied in many areas such as the medicine for volumetric reconstruction of tomography data, in robotics to reconstruct surfaces or scenes using range sensor information, in industrial systems for quality control of manufactured objects or even in biology to study the structure and folding of proteins. One of the algorithm’s main problems is its high computational complexity (quadratic in the number of points with the non-optimized original variant) in a context where high density point sets, acquired by high resolution scanners, are processed. Many variants have been proposed in the literature whose goal is the performance improvement either by reducing the number of points or the required iterations or even enhancing the complexity of the most expensive phase: the closest neighbor search. In spite of decreasing its complexity, some of the variants tend to have a negative impact on the final registration precision or the convergence domain thus limiting the possible application scenarios. The goal of this work is the improvement of the algorithm’s computational cost so that a wider range of computationally demanding problems from among the ones described before can be addressed. For that purpose, an experimental and mathematical convergence analysis and validation of point-to-point distance metrics has been performed taking into account those distances with lower computational cost than the Euclidean one, which is used as the de facto standard for the algorithm’s implementations in the literature. In that analysis, the functioning of the algorithm in diverse topological spaces, characterized by different metrics, has been studied to check the convergence, efficacy and cost of the method in order to determine the one which offers the best results. Given that the distance calculation represents a significant part of the whole set of computations performed by the algorithm, it is expected that any reduction of that operation affects significantly and positively the overall performance of the method. As a result, a performance improvement has been achieved by the application of those reduced cost metrics whose quality in terms of convergence and error has been analyzed and validated experimentally as comparable with respect to the Euclidean distance using a heterogeneous set of objects, scenarios and initial situations.
Resumo:
Since the beginning of 3D computer vision problems, the use of techniques to reduce the data to make it treatable preserving the important aspects of the scene has been necessary. Currently, with the new low-cost RGB-D sensors, which provide a stream of color and 3D data of approximately 30 frames per second, this is getting more relevance. Many applications make use of these sensors and need a preprocessing to downsample the data in order to either reduce the processing time or improve the data (e.g., reducing noise or enhancing the important features). In this paper, we present a comparison of different downsampling techniques which are based on different principles. Concretely, five different downsampling methods are included: a bilinear-based method, a normal-based, a color-based, a combination of the normal and color-based samplings, and a growing neural gas (GNG)-based approach. For the comparison, two different models have been used acquired with the Blensor software. Moreover, to evaluate the effect of the downsampling in a real application, a 3D non-rigid registration is performed with the data sampled. From the experimentation we can conclude that depending on the purpose of the application some kernels of the sampling methods can improve drastically the results. Bilinear- and GNG-based methods provide homogeneous point clouds, but color-based and normal-based provide datasets with higher density of points in areas with specific features. In the non-rigid application, if a color-based sampled point cloud is used, it is possible to properly register two datasets for cases where intensity data are relevant in the model and outperform the results if only a homogeneous sampling is used.
Resumo:
Purpose: To study the population of intrinsically photosensitive retinal ganglion cells (melanopsin-expressing RGCs, m+RGCs) in P23H-1 rats, a rat model of inherited photoreceptor degeneration. Methods: At postnatal (P) times P30, P365, and P540, retinas from P23H dystrophic rats (line 1, rapid degeneration; and line 3, slow degeneration) and Sprague Dawley (SD) rats (control) were dissected as whole-mounts and immunodetected for melanopsin and/or Brn3a. The dendritic arborization of m+RGCs and the numbers of Brn3a+RGCs and m+RGCs were quantified and their retinal distribution and coexpression analyzed. Results: In SD rats, aging did not affect the population of Brn3a+RGCs or m+RGCs or the percentage that showed coexpression (0.27%). Young P23H-1 rats had a significantly lower number of Brn3a+RGCs and showed a further decline with age. The population of m+RGCs in young P23H-1 rats was similar to that found in SD rats and decreased by 22.6% and 28.2% at P365 and P540, respectively, similarly to the decrease of the Brn3a+RGCs. At these ages the m+RGCs showed a decrease of their dendritic arborization parameters, which was similar in both the P23H-1 and P23H-3 lines. The percentage of coexpression of Brn3a was, however, already significantly higher at P30 (3.31%) and increased significantly with age (10.65% at P540). Conclusions: Inherited photoreceptor degeneration was followed by secondary loss of Brn3a+RGCs and m+RGCs. Surviving m+RGCs showed decreased dendritic arborization parameters and increased coexpression of Brn3a and melanopsin, phenotypic and molecular changes that may represent an effort to resist degeneration and/or preferential survival of m+RGCs capable of synthesizing Brn3a.
Resumo:
Examining a team’s performance from a physical point of view their momentum might indicate unexpected turning points in defeat or success. Physicists describe this value as to require some effort to be started, but also that it is relatively easy to keep it going once a sufficient level is reached (Reed and Hughes, 2006). Unlike football, rugby, handball and many more sports, a regular volleyball match is not limited by time but by points that need to be gathered. Every minute more than one point is won by either one team or the other. That means a series of successive points enlarges the gap between the teams making it more and more difficult to catch up with the leading one. This concept of gathering momentum, or the reverse in a performance, can give the coaches, athletes and sports scientists further insights into winning and losing performances. Momentum investigations also contain dependencies between performances or questions if future performances are reliant upon past streaks. Squash and volleyball share the characteristic of being played up to a certain amount of points. Squash was examined according to the momentum of players by Hughes et al. (2006). The initial aim was to expand normative profiles of elite squash players using momentum graphs of winners and errors to explore ‘turning points’ in a performance. Dynamic systems theory has enabled the definition of perturbations in sports exhibiting rhythms (Hughes et al., 2000; McGarry et al., 2002; Murray et al., 2008), and how players and teams cause these disruptions of rhythm can inform on the way they play, these techniques also contribute to profiling methods. Together with the analysis of one’s own performance it is essential to have an understanding of your oppositions’ tactical strengths and weaknesses. By modelling the oppositions’ performance it is possible to predict certain outcomes and patterns, and therefore intervene or change tactics before the critical incident occurs. The modelling of competitive sport is an informative analytic technique as it directs the attention of the modeller to the critical aspects of data that delineate successful performance (McGarry & Franks, 1996). Using tactical performance profiles to pull out and visualise these critical aspects of performance, players can build justified and sophisticated tactical plans. The area is discussed and reviewed, critically appraising the research completed in this element of Performance Analysis.
Resumo:
Several recent works deal with 3D data in mobile robotic problems, e.g. mapping or egomotion. Data comes from any kind of sensor such as stereo vision systems, time of flight cameras or 3D lasers, providing a huge amount of unorganized 3D data. In this paper, we describe an efficient method to build complete 3D models from a Growing Neural Gas (GNG). The GNG is applied to the 3D raw data and it reduces both the subjacent error and the number of points, keeping the topology of the 3D data. The GNG output is then used in a 3D feature extraction method. We have performed a deep study in which we quantitatively show that the use of GNG improves the 3D feature extraction method. We also show that our method can be applied to any kind of 3D data. The 3D features obtained are used as input in an Iterative Closest Point (ICP)-like method to compute the 6DoF movement performed by a mobile robot. A comparison with standard ICP is performed, showing that the use of GNG improves the results. Final results of 3D mapping from the egomotion calculated are also shown.
Resumo:
From a set of gonioapparent automotive samples from different manufacturers we selected 28 low-chroma color pairs with relatively small color differences predominantly in lightness. These color pairs were visually assessed with a gray scale at six different viewing angles by a panel of 10 observers. Using the Standardized Residual Sum of Squares (STRESS) index, the results of our visual experiment were tested against predictions made by 12 modern color-difference formulas. From a weighted STRESS index accounting for the uncertainty in visual assessments, the best prediction of our whole experiment was achieved using AUDI2000, CAM02-SCD, CAM02-UCS and OSA-GP-Euclidean color-difference formulas, which were no statistically significant different among them. A two-step optimization of the original AUDI2000 color-difference formula resulted in a modified AUDI2000 formula which performed both, significantly better than the original formula and below the experimental inter-observer variability. Nevertheless the proposal of a new revised AUDI2000 color-difference formula requires additional experimental data.
Resumo:
Mathematical morphology has been an area of intensive research over the last few years. Although many remarkable advances have been achieved throughout these years, there is still a great interest in accelerating morphological operations in order for them to be implemented in real-time systems. In this work, we present a new model for computing mathematical morphology operations, the so-called morphological trajectory model (MTM), in which a morphological filter will be divided into a sequence of basic operations. Then, a trajectory-based morphological operation (such as dilation, and erosion) is defined as the set of points resulting from the ordered application of the instant basic operations. The MTM approach allows working with different structuring elements, such as disks, and from the experiments, it can be extracted that our method is independent of the structuring element size and can be easily applied to industrial systems and high-resolution images.
Resumo:
The use of 3D data in mobile robotics applications provides valuable information about the robot’s environment but usually the huge amount of 3D information is unmanageable by the robot storage and computing capabilities. A data compression is necessary to store and manage this information but preserving as much information as possible. In this paper, we propose a 3D lossy compression system based on plane extraction which represent the points of each scene plane as a Delaunay triangulation and a set of points/area information. The compression system can be customized to achieve different data compression or accuracy ratios. It also supports a color segmentation stage to preserve original scene color information and provides a realistic scene reconstruction. The design of the method provides a fast scene reconstruction useful for further visualization or processing tasks.
Resumo:
Purpose – The purpose of this paper is to present a new geometric model based on the mathematical morphology paradigm, specialized to provide determinism to the classic morphological operations. The determinism is needed to model dynamic processes that require an order of application, as is the case for designing and manufacturing objects in CAD/CAM environments. Design/methodology/approach – The basic trajectory-based operation is the basis of the proposed morphological specialization. This operation allows the definition of morphological operators that obtain sequentially ordered sets of points from the boundary of the target objects, inexistent determinism in the classical morphological paradigm. From this basic operation, the complete set of morphological operators is redefined, incorporating the concept of boundary and determinism: trajectory-based erosion and dilation, and other morphological filtering operations. Findings – This new morphological framework allows the definition of complex three-dimensional objects, providing arithmetical support to generating machining trajectories, one of the most complex problems currently occurring in CAD/CAM. Originality/value – The model proposes the integration of the processes of design and manufacture, so that it avoids the problems of accuracy and integrity that present other classic geometric models that divide these processes in two phases. Furthermore, the morphological operative is based on points sets, so the geometric data structures and the operations are intrinsically simple and efficient. Another important value that no excessive computational resources are needed, because only the points in the boundary are processed.
Resumo:
%WL: porcentaje de pérdida de peso; %FL: porcentaje de pérdida de grasa; PNLWF: pacientes que pierden peso o grasa; PLWF: pacientes que pierden peso y grasa. Objetivo: evaluar si el %WL y el %FL en el tratamiento dietético, se vieron afectados por el género, la edad, el IMC y la asistencia a la consulta. Método: 4.700 consultas, 670 pacientes (IMC ≥ 25), en el sur-este de España (2006-12). Se utilizó la dieta equilibrada e hipocalórica. Dos tipos de pacientes: PNLWF y PLWF (91,9%). Resultados: en los PLWF, los hombres y los que asisten en mayor número de ocasiones a la consulta han mostrado una mayor pérdida, frente a las mujeres (%FL: 23,0 vs 14,3%, p = 0,000; %WL: 7,7 vs 6,6%, p = 0,020), y los que asisten con menor frecuencia (%FL: 19,1 vs 7,3%, p = 0,000; %WL: 7,8 vs 2,9%, p = 0,000). El análisis de regresión multinomial (PNLWF/ PLWF) indica que solo el asistir a más de mes y medio a la consulta es un factor que influye en la pérdida, OR 8,3 (IC 95% 4,5-15,1; p = 0,000). Conclusión: la medición de la grasa corporal proporciona una información adicional al peso perdido; la mayoría de los pacientes que asisten más de mes y medio obtienen un elevado %FL; la asistencia es un factor predictor de la pérdida; el %FL indica que el tratamiento dietético juega un papel principal en la resolución de esta patología; se recomienda diseñar esquemas prácticos del proceso de actuación de los nutricionistas en función del IMCi y el variable.
Resumo:
In this work, a modified version of the elastic bunch graph matching (EBGM) algorithm for face recognition is introduced. First, faces are detected by using a fuzzy skin detector based on the RGB color space. Then, the fiducial points for the facial graph are extracted automatically by adjusting a grid of points to the result of an edge detector. After that, the position of the nodes, their relation with their neighbors and their Gabor jets are calculated in order to obtain the feature vector defining each face. A self-organizing map (SOM) framework is shown afterwards. Thus, the calculation of the winning neuron and the recognition process are performed by using a similarity function that takes into account both the geometric and texture information of the facial graph. The set of experiments carried out for our SOM-EBGM method shows the accuracy of our proposal when compared with other state-of the-art methods.
Resumo:
Nowadays, new computers generation provides a high performance that enables to build computationally expensive computer vision applications applied to mobile robotics. Building a map of the environment is a common task of a robot and is an essential part to allow the robots to move through these environments. Traditionally, mobile robots used a combination of several sensors from different technologies. Lasers, sonars and contact sensors have been typically used in any mobile robotic architecture, however color cameras are an important sensor due to we want the robots to use the same information that humans to sense and move through the different environments. Color cameras are cheap and flexible but a lot of work need to be done to give robots enough visual understanding of the scenes. Computer vision algorithms are computational complex problems but nowadays robots have access to different and powerful architectures that can be used for mobile robotics purposes. The advent of low-cost RGB-D sensors like Microsoft Kinect which provide 3D colored point clouds at high frame rates made the computer vision even more relevant in the mobile robotics field. The combination of visual and 3D data allows the systems to use both computer vision and 3D processing and therefore to be aware of more details of the surrounding environment. The research described in this thesis was motivated by the need of scene mapping. Being aware of the surrounding environment is a key feature in many mobile robotics applications from simple robotic navigation to complex surveillance applications. In addition, the acquisition of a 3D model of the scenes is useful in many areas as video games scene modeling where well-known places are reconstructed and added to game systems or advertising where once you get the 3D model of one room the system can add furniture pieces using augmented reality techniques. In this thesis we perform an experimental study of the state-of-the-art registration methods to find which one fits better to our scene mapping purposes. Different methods are tested and analyzed on different scene distributions of visual and geometry appearance. In addition, this thesis proposes two methods for 3d data compression and representation of 3D maps. Our 3D representation proposal is based on the use of Growing Neural Gas (GNG) method. This Self-Organizing Maps (SOMs) has been successfully used for clustering, pattern recognition and topology representation of various kind of data. Until now, Self-Organizing Maps have been primarily computed offline and their application in 3D data has mainly focused on free noise models without considering time constraints. Self-organising neural models have the ability to provide a good representation of the input space. In particular, the Growing Neural Gas (GNG) is a suitable model because of its flexibility, rapid adaptation and excellent quality of representation. However, this type of learning is time consuming, specially for high-dimensional input data. Since real applications often work under time constraints, it is necessary to adapt the learning process in order to complete it in a predefined time. This thesis proposes a hardware implementation leveraging the computing power of modern GPUs which takes advantage of a new paradigm coined as General-Purpose Computing on Graphics Processing Units (GPGPU). Our proposed geometrical 3D compression method seeks to reduce the 3D information using plane detection as basic structure to compress the data. This is due to our target environments are man-made and therefore there are a lot of points that belong to a plane surface. Our proposed method is able to get good compression results in those man-made scenarios. The detected and compressed planes can be also used in other applications as surface reconstruction or plane-based registration algorithms. Finally, we have also demonstrated the goodness of the GPU technologies getting a high performance implementation of a CAD/CAM common technique called Virtual Digitizing.