13 resultados para sample covariance matrix
em Aston University Research Archive
Resumo:
A recently proposed colour based tracking algorithm has been established to track objects in real circumstances [Zivkovic, Z., Krose, B. 2004. An EM-like algorithm for color-histogram-based object tracking. In: Proc, IEEE Conf. on Computer Vision and Pattern Recognition, pp. 798-803]. To improve the performance of this technique in complex scenes, in this paper we propose a new algorithm for optimally adapting the ellipse outlining the objects of interest. This paper presents a Lagrangian based method to integrate a regularising component into the covariance matrix to be computed. Technically, we intend to reduce the residuals between the estimated probability distribution and the expected one. We argue that, by doing this, the shape of the ellipse can be properly adapted in the tracking stage. Experimental results show that the proposed method has favourable performance in shape adaption and object localisation.
Resumo:
The principled statistical application of Gaussian random field models used in geostatistics has historically been limited to data sets of a small size. This limitation is imposed by the requirement to store and invert the covariance matrix of all the samples to obtain a predictive distribution at unsampled locations, or to use likelihood-based covariance estimation. Various ad hoc approaches to solve this problem have been adopted, such as selecting a neighborhood region and/or a small number of observations to use in the kriging process, but these have no sound theoretical basis and it is unclear what information is being lost. In this article, we present a Bayesian method for estimating the posterior mean and covariance structures of a Gaussian random field using a sequential estimation algorithm. By imposing sparsity in a well-defined framework, the algorithm retains a subset of “basis vectors” that best represent the “true” posterior Gaussian random field model in the relative entropy sense. This allows a principled treatment of Gaussian random field models on very large data sets. The method is particularly appropriate when the Gaussian random field model is regarded as a latent variable model, which may be nonlinearly related to the observations. We show the application of the sequential, sparse Bayesian estimation in Gaussian random field models and discuss its merits and drawbacks.
Resumo:
Recently within the machine learning and spatial statistics communities many papers have explored the potential of reduced rank representations of the covariance matrix, often referred to as projected or fixed rank approaches. In such methods the covariance function of the posterior process is represented by a reduced rank approximation which is chosen such that there is minimal information loss. In this paper a sequential framework for inference in such projected processes is presented, where the observations are considered one at a time. We introduce a C++ library for carrying out such projected, sequential estimation which adds several novel features. In particular we have incorporated the ability to use a generic observation operator, or sensor model, to permit data fusion. We can also cope with a range of observation error characteristics, including non-Gaussian observation errors. Inference for the variogram parameters is based on maximum likelihood estimation. We illustrate the projected sequential method in application to synthetic and real data sets. We discuss the software implementation and suggest possible future extensions.
Resumo:
With the ability to collect and store increasingly large datasets on modern computers comes the need to be able to process the data in a way that can be useful to a Geostatistician or application scientist. Although the storage requirements only scale linearly with the number of observations in the dataset, the computational complexity in terms of memory and speed, scale quadratically and cubically respectively for likelihood-based Geostatistics. Various methods have been proposed and are extensively used in an attempt to overcome these complexity issues. This thesis introduces a number of principled techniques for treating large datasets with an emphasis on three main areas: reduced complexity covariance matrices, sparsity in the covariance matrix and parallel algorithms for distributed computation. These techniques are presented individually, but it is also shown how they can be combined to produce techniques for further improving computational efficiency.
Resumo:
Exploratory analysis of data seeks to find common patterns to gain insights into the structure and distribution of the data. In geochemistry it is a valuable means to gain insights into the complicated processes making up a petroleum system. Typically linear visualisation methods like principal components analysis, linked plots, or brushing are used. These methods can not directly be employed when dealing with missing data and they struggle to capture global non-linear structures in the data, however they can do so locally. This thesis discusses a complementary approach based on a non-linear probabilistic model. The generative topographic mapping (GTM) enables the visualisation of the effects of very many variables on a single plot, which is able to incorporate more structure than a two dimensional principal components plot. The model can deal with uncertainty, missing data and allows for the exploration of the non-linear structure in the data. In this thesis a novel approach to initialise the GTM with arbitrary projections is developed. This makes it possible to combine GTM with algorithms like Isomap and fit complex non-linear structure like the Swiss-roll. Another novel extension is the incorporation of prior knowledge about the structure of the covariance matrix. This extension greatly enhances the modelling capabilities of the algorithm resulting in better fit to the data and better imputation capabilities for missing data. Additionally an extensive benchmark study of the missing data imputation capabilities of GTM is performed. Further a novel approach, based on missing data, will be introduced to benchmark the fit of probabilistic visualisation algorithms on unlabelled data. Finally the work is complemented by evaluating the algorithms on real-life datasets from geochemical projects.
Resumo:
In this letter, we derive continuum equations for the generalization error of the Bayesian online algorithm (BOnA) for the one-layer perceptron with a spherical covariance matrix using the Rosenblatt potential and show, by numerical calculations, that the asymptotic performance of the algorithm is the same as the one for the optimal algorithm found by means of variational methods with the added advantage that the BOnA does not use any inaccessible information during learning. © 2007 IEEE.
Resumo:
Heterogeneous datasets arise naturally in most applications due to the use of a variety of sensors and measuring platforms. Such datasets can be heterogeneous in terms of the error characteristics and sensor models. Treating such data is most naturally accomplished using a Bayesian or model-based geostatistical approach; however, such methods generally scale rather badly with the size of dataset, and require computationally expensive Monte Carlo based inference. Recently within the machine learning and spatial statistics communities many papers have explored the potential of reduced rank representations of the covariance matrix, often referred to as projected or fixed rank approaches. In such methods the covariance function of the posterior process is represented by a reduced rank approximation which is chosen such that there is minimal information loss. In this paper a sequential Bayesian framework for inference in such projected processes is presented. The observations are considered one at a time which avoids the need for high dimensional integrals typically required in a Bayesian approach. A C++ library, gptk, which is part of the INTAMAP web service, is introduced which implements projected, sequential estimation and adds several novel features. In particular the library includes the ability to use a generic observation operator, or sensor model, to permit data fusion. It is also possible to cope with a range of observation error characteristics, including non-Gaussian observation errors. Inference for the covariance parameters is explored, including the impact of the projected process approximation on likelihood profiles. We illustrate the projected sequential method in application to synthetic and real datasets. Limitations and extensions are discussed. © 2010 Elsevier Ltd.
Resumo:
Matrix application continues to be a critical step in sample preparation for matrix-assisted laser desorption/ionization (MALDI) mass spectrometry imaging (MSI). Imaging of small molecules such as drugs and metabolites is particularly problematic because the commonly used washing steps to remove salts are usually omitted as they may also remove the analyte, and analyte spreading is more likely with conventional wet matrix application methods. We have developed a method which uses the application of matrix as a dry, finely divided powder, here referred to as dry matrix application, for the imaging of drug compounds. This appears to offer a complementary method to wet matrix application for the MALDI-MSI of small molecules, with the alternative matrix application techniques producing different ion profiles, and allows the visualization of compounds not observed using wet matrix application methods. We demonstrate its value in imaging clozapine from rat kidney and 4-bromophenyl-1,4-diazabicyclo(3.2.2)nonane-4-carboxylic acid from rat brain. In addition, exposure of the dry matrix coated sample to a saturated moist atmosphere appears to enhance the visualization of a different set of molecules.
Resumo:
Mass spectrometry imaging (MSI) is a powerful tool in metabolomics and proteomics for the spatial localization and identification of pharmaceuticals, metabolites, lipids, peptides and proteins in biological tissues. However, sample preparation remains a crucial variable in obtaining the most accurate distributions. Common washing steps used to remove salts, and solvent-based matrix application, allow analyte spreading to occur. Solvent-free matrix applications can reduce this risk, but increase the possibility of ionisation bias due to matrix adhesion to tissue sections. We report here the use of matrix-free MSI using laser desorption ionisation performed on a 12 T Fourier transform ion cyclotron resonance (FTICR) mass spectrometer. We used unprocessed tissue with no post-processing following thaw-mounting on matrix-assisted laser desorption ionisation (MALDI) indium-tin oxide (ITO) target plates. The identification and distribution of a range of phospholipids in mouse brain and kidney sections are presented and compared with previously published MALDI time-of-flight (TOF) MSI distributions.
Resumo:
The present study investigates the effect of different sample preparation methods on the pyrolysis behaviour of metal-added biomass; Willow samples were compared in the presence of two salts of zinc and lead containing sulphate and nitrate anions which were added to the wood samples with three different techniques as dry-mixing, impregnation and ion-exchange. The effect of acid and water wash as common demineralisation pre-treatments were also analysed to evaluate their roles in the thermal degradation of the biomass. Results from thermogravimetric analysis (TGA), Fourier transform infrared spectroscopy (FT-IR) and pyrolysis-mass spectrometry (Py-MS) measurements indicated that these pre-treatments change the matrix and the physical-chemical properties of wood. Results suggested that these structural changes increase the thermal stability of cellulose during pyrolysis. Sample preparation was also found to be a crucial factor during pyrolysis; different anions of metal salts changed the weight loss rate curves of wood material, which indicates changes in the primary degradation process of the biomass. Results also showed that dry-mixing, impregnation or ion-exchange influence the thermal behaviour of wood in different ways when a chosen metal salt was and added to the wood material. © 2011 Elsevier B.V. All rights reserved.
Resumo:
Poly(β-hydroxybutyrate), (PHB), is a biologically produced, biodegradable thennoplastic with commercial potential. In this work the qualitative and quantitative investigations of the structure and degradation of a previously unstudied, novel, fibrous form of PHB, were completed. This gel-spun PHB fibrous matrix, PHB(FM), which has a similar appearance to cotton wool, possesses a relatively complex structure which combines a large volume with a low mass and has potential for use as a wound scaffolding device. As a result of the intrinsic problems presented by this novel structure, a new experimental procedure was developed to analyze the degradation of the PHB to its monomer hydroxybutyric acid, (HBA). This procedure was used in an accelerated degradation model which accurately monitored the degradation of the undegraded and degraded fractions of a fibrous matrix and the degradation of its PHB component. The in vitro degradation mechanism was also monitored using phase contrast and scanning electron microscopy, differential scanning calorimetry, fibre diameter distributions and Fourier infra-red photoacoustic spectroscopy. The accelerated degradation model was used to predict the degradation of the samples in the physiological model and this provided a clearer picture as to the samples potential biodegradation as medical implantation devices. The degradation of the matrices was characterized by an initial penetration of the degradative medium and weakening of the fibre integrity due to cleavage of the ester linkages, this then led to the physical collapse of the fibres which increased the surface area to volume ratio of the sample and facilitated its degradation. Degradation in the later stages was reduced due to the experimental kinetics, compaction and degradation resistant material, most probably the highly crystalline regions of the PHB. The in vitro degradation of the PHB(FM) was influenced by blending with various polysaccharides, copolymerizing with poly(~-hydroxyvalerate), (PHV), and changes to the manufacturing process. The degradation was also detennined to be faster than that of conventional melt processed PHB based samples. It was concluded that the material factors such as processing, sample size and shape affected the degradation of PHB based samples with the major factor of sample surface area to volume ratio being of paramount importance in determining the degradation of a sample.
Resumo:
In this paper we discuss a fast Bayesian extension to kriging algorithms which has been used successfully for fast, automatic mapping in emergency conditions in the Spatial Interpolation Comparison 2004 (SIC2004) exercise. The application of kriging to automatic mapping raises several issues such as robustness, scalability, speed and parameter estimation. Various ad-hoc solutions have been proposed and used extensively but they lack a sound theoretical basis. In this paper we show how observations can be projected onto a representative subset of the data, without losing significant information. This allows the complexity of the algorithm to grow as O(n m 2), where n is the total number of observations and m is the size of the subset of the observations retained for prediction. The main contribution of this paper is to further extend this projective method through the application of space-limited covariance functions, which can be used as an alternative to the commonly used covariance models. In many real world applications the correlation between observations essentially vanishes beyond a certain separation distance. Thus it makes sense to use a covariance model that encompasses this belief since this leads to sparse covariance matrices for which optimised sparse matrix techniques can be used. In the presence of extreme values we show that space-limited covariance functions offer an additional benefit, they maintain the smoothness locally but at the same time lead to a more robust, and compact, global model. We show the performance of this technique coupled with the sparse extension to the kriging algorithm on synthetic data and outline a number of computational benefits such an approach brings. To test the relevance to automatic mapping we apply the method to the data used in a recent comparison of interpolation techniques (SIC2004) to map the levels of background ambient gamma radiation. © Springer-Verlag 2007.
Resumo:
The morphology, chemical composition, and mechanical properties in the surface region of α-irradiated polytetrafluoroethylene (PTFE) have been examined and compared to unirradiated specimens. Samples were irradiated with 5.5 MeV 4He2+ ions from a tandem accelerator to doses between 1 × 106 and 5 × 1010 Rad. Static time-of-flight secondary ion mass spectrometry (ToF-SIMS), using a 20 keV C60+ source, was employed to probe chemical changes as a function of a dose. Chemical images and high resolution spectra were collected and analyzed to reveal the effects of a particle radiation on the chemical structure. Residual gas analysis (RGA) was utilized to monitor the evolution of volatile species during vacuum irradiation of the samples. Scanning electron microscopy (SEM) was used to observe the morphological variation of samples with increasing a particle dose, and nanoindentation was engaged to determine the hardness and elastic modulus as a function of a dose. The data show that PTFE nominally retains its innate chemical structure and morphology at a doses <109 Rad. At α doses ≥109 Rad the polymer matrix experiences increased chemical degradation and morphological roughening which are accompanied by increased hardness and declining elasticity. At α doses >1010 Rad the polymer matrix suffers severe chemical degradation and material loss. Chemical degradation is observed in ToF-SIMS by detection of ions that are indicative of fragmentation, unsaturation, and functionalization of molecules in the PTFE matrix. The mass spectra also expose the subtle trends of crosslinking within the α-irradiated polymer matrix. ToF-SIMS images support the assertion that chemical degradation is the result of a particle irradiation and show morphological roughening of the sample with increased a dose. High resolution SEM images more clearly illustrate the morphological roughening and the mass loss that accompanies high doses of a particles. RGA confirms the supposition that the outcome of chemical degradation in the PTFE matrix with continuing irradiation is evolution of volatile species resulting in morphological roughening and mass loss. Finally, we reveal and discuss relationships between chemical structure and mechanical properties such as hardness and elastic modulus.