959 resultados para approximated inference


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Intracellular replication within specialized vacuoles and cell-to-cell spread in the tissue are essential for the virulence of Salmonella enterica. By observing infection dynamics at the single-cell level in vivo, we have discovered that the Salmonella pathogenicity island 2 (SPI-2) type 3 secretory system (T3SS) is dispensable for growth to high intracellular densities. This challenges the concept that intracellular replication absolutely requires proteins delivered by SPI-2 T3SS, which has been derived largely by inference from in vitro cell experiments and from unrefined measurement of net growth in mouse organs. Furthermore, we infer from our data that the SPI-2 T3SS mediates exit from infected cells, with consequent formation of new infection foci resulting in bacterial spread in the tissues. This suggests a new role for SPI-2 in vivo as a mediator of bacterial spread in the body. In addition, we demonstrate that very similar net growth rates of attenuated salmonellae in organs can be derived from very different underlying intracellular growth dynamics.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A fundamental problem in the analysis of structured relational data like graphs, networks, databases, and matrices is to extract a summary of the common structure underlying relations between individual entities. Relational data are typically encoded in the form of arrays; invariance to the ordering of rows and columns corresponds to exchangeable arrays. Results in probability theory due to Aldous, Hoover and Kallenberg show that exchangeable arrays can be represented in terms of a random measurable function which constitutes the natural model parameter in a Bayesian model. We obtain a flexible yet simple Bayesian nonparametric model by placing a Gaussian process prior on the parameter function. Efficient inference utilises elliptical slice sampling combined with a random sparse approximation to the Gaussian process. We demonstrate applications of the model to network data and clarify its relation to models in the literature, several of which emerge as special cases.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The brain extracts useful features from a maelstrom of sensory information, and a fundamental goal of theoretical neuroscience is to work out how it does so. One proposed feature extraction strategy is motivated by the observation that the meaning of sensory data, such as the identity of a moving visual object, is often more persistent than the activation of any single sensory receptor. This notion is embodied in the slow feature analysis (SFA) algorithm, which uses “slowness” as an heuristic by which to extract semantic information from multi-dimensional time-series. Here, we develop a probabilistic interpretation of this algorithm showing that inference and learning in the limiting case of a suitable probabilistic model yield exactly the results of SFA. Similar equivalences have proved useful in interpreting and extending comparable algorithms such as independent component analysis. For SFA, we use the equivalent probabilistic model as a conceptual spring-board, with which to motivate several novel extensions to the algorithm.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Natural sounds are structured on many time-scales. A typical segment of speech, for example, contains features that span four orders of magnitude: Sentences ($\sim1$s); phonemes ($\sim10$−$1$ s); glottal pulses ($\sim 10$−$2$s); and formants ($\sim 10$−$3$s). The auditory system uses information from each of these time-scales to solve complicated tasks such as auditory scene analysis [1]. One route toward understanding how auditory processing accomplishes this analysis is to build neuroscience-inspired algorithms which solve similar tasks and to compare the properties of these algorithms with properties of auditory processing. There is however a discord: Current machine-audition algorithms largely concentrate on the shorter time-scale structures in sounds, and the longer structures are ignored. The reason for this is two-fold. Firstly, it is a difficult technical problem to construct an algorithm that utilises both sorts of information. Secondly, it is computationally demanding to simultaneously process data both at high resolution (to extract short temporal information) and for long duration (to extract long temporal information). The contribution of this work is to develop a new statistical model for natural sounds that captures structure across a wide range of time-scales, and to provide efficient learning and inference algorithms. We demonstrate the success of this approach on a missing data task.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Variational methods are a key component of the approximate inference and learning toolbox. These methods fill an important middle ground, retaining distributional information about uncertainty in latent variables, unlike maximum a posteriori methods (MAP), and yet generally requiring less computational time than Monte Carlo Markov Chain methods. In particular the variational Expectation Maximisation (vEM) and variational Bayes algorithms, both involving variational optimisation of a free-energy, are widely used in time-series modelling. Here, we investigate the success of vEM in simple probabilistic time-series models. First we consider the inference step of vEM, and show that a consequence of the well-known compactness property of variational inference is a failure to propagate uncertainty in time, thus limiting the usefulness of the retained distributional information. In particular, the uncertainty may appear to be smallest precisely when the approximation is poorest. Second, we consider parameter learning and analytically reveal systematic biases in the parameters found by vEM. Surprisingly, simpler variational approximations (such a mean-field) can lead to less bias than more complicated structured approximations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We show how machine learning techniques based on Bayesian inference can be used to reach new levels of realism in the computer simulation of molecular materials, focusing here on water. We train our machine-learning algorithm using accurate, correlated quantum chemistry, and predict energies and forces in molecular aggregates ranging from clusters to solid and liquid phases. The widely used electronic-structure methods based on density-functional theory (DFT) give poor accuracy for molecular materials like water, and we show how our techniques can be used to generate systematically improvable corrections to DFT. The resulting corrected DFT scheme gives remarkably accurate predictions for the relative energies of small water clusters and of different ice structures, and greatly improves the description of the structure and dynamics of liquid water.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Copulas allow to learn marginal distributions separately from the multivariate dependence structure (copula) that links them together into a density function. Vine factorizations ease the learning of high-dimensional copulas by constructing a hierarchy of conditional bivariate copulas. However, to simplify inference, it is common to assume that each of these conditional bivariate copulas is independent from its conditioning variables. In this paper, we relax this assumption by discovering the latent functions that specify the shape of a conditional copula given its conditioning variables We learn these functions by following a Bayesian approach based on sparse Gaussian processes with expectation propagation for scalable, approximate inference. Experiments on real-world datasets show that, when modeling all conditional dependencies, we obtain better estimates of the underlying copula of the data.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Leading edge vortices are considered to be important in generating the high lift coefficients observed in insect flight and may therefore be relevant to micro-air vehicles. A potential flow model of an impulsively started flat plate, featuring a leading edge vortex (LEV) and a trailing edge vortex (TEV) is fitted to experimental data in order to provide insight into the mechanisms that influence the convection of the LEV and to study how the LEV contributes to lift. The potential flow model fits the experimental data best with no bound circulation, which is in accordance with Kelvin's circulation theorem. The lift-to-drag ratio is well approximated by the function 'cot α' for α > 15°, which supports the tentative conclusion that shortly after an impulsive start, at post-stall angles of attack, lift is caused non-circulatory forces and by the action of the LEV as opposed to bound circulation. Copyright © 2012 by C. W. Pitt Ford.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Many transductive inference algorithms assume that distributions over training and test estimates should be related, e.g. by providing a large margin of separation on both sets. We use this idea to design a transduction algorithm which can be used without modification for classification, regression, and structured estimation. At its heart we exploit the fact that for a good learner the distributions over the outputs on training and test sets should match. This is a classical two-sample problem which can be solved efficiently in its most general form by using distance measures in Hilbert Space. It turns out that a number of existing heuristics can be viewed as special cases of our approach.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The code provided here originally demonstrated the main algorithms from Rasmussen and Williams: Gaussian Processes for Machine Learning. It has since grown to allow more likelihood functions, further inference methods and a flexible framework for specifying GPs.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Building on recent developments in mixed methods, we discuss the methodological implications of critical realism and explore how these can guide dynamic mixed-methods research design in information systems. Specifically, we examine the core ontological assumptions of CR in order to gain some perspective on key epistemological issues such as causation and validity, and illustrate how these shape our logic of inference in the research process through what is known as retroduction. We demonstrate the value of a CR-led mixed-methods research approach by drawing on a study that examines the impact of ICT adoption in the financial services sector. In doing so, we provide insight into the interplay between qualitative and quantitative methods and the particular value of applying mixed methods guided by CR methodological principles. Our positioning of demi-regularities within the process of retroduction contributes a distinctive development in this regard. We argue that such a research design enables us to better address issues of validity and the development of more robust meta-inferences.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The fundamental aim of clustering algorithms is to partition data points. We consider tasks where the discovered partition is allowed to vary with some covariate such as space or time. One approach would be to use fragmentation-coagulation processes, but these, being Markov processes, are restricted to linear or tree structured covariate spaces. We define a partition-valued process on an arbitrary covariate space using Gaussian processes. We use the process to construct a multitask clustering model which partitions datapoints in a similar way across multiple data sources, and a time series model of network data which allows cluster assignments to vary over time. We describe sampling algorithms for inference and apply our method to defining cancer subtypes based on different types of cellular characteristics, finding regulatory modules from gene expression data from multiple human populations, and discovering time varying community structure in a social network.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Flapping wings often feature a leading-edge vortex (LEV) that is thought to enhance the lift generated by the wing. Here the lift on a wing featuring a leading-edge vortex is considered by performing experiments on a translating flat-plate aerofoil that is accelerated from rest in a water towing tank at a fixed angle of attack of 15°. The unsteady flow is investigated with dye flow visualization, particle image velocimetry (PIV) and force measurements. Leading-and trailing-edge vortex circulation and position are calculated directly from the velocity vectors obtained using PIV. In order to determine the most appropriate value of bound circulation, a two-dimensional potential flow model is employed and flow fields are calculated for a range of values of bound circulation. In this way, the value of bound circulation is selected to give the best fit between the experimental velocity field and the potential flow field. Early in the trajectory, the value of bound circulation calculated using this potential flow method is in accordance with Kelvin's circulation theorem, but differs from the values predicted by Wagner's growth of bound circulation and the Kutta condition. Later the Kutta condition is established but the bound circulation remains small; most of the circulation is contained instead in the LEVs. The growth of wake circulation can be approximated by Wagner's circulation curve. Superimposing the non-circulatory lift, approximated from the potential flow model, and Wagner's lift curve gives a first-order approximation of the measured lift. Lift is generated by inertial effects and the slow buildup of circulation, which is contained in shed vortices rather than bound circulation. © 2013 Cambridge University Press.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper proposes a hierarchical probabilistic model for ordinal matrix factorization. Unlike previous approaches, we model the ordinal nature of the data and take a principled approach to incorporating priors for the hidden variables. Two algorithms are presented for inference, one based on Gibbs sampling and one based on variational Bayes. Importantly, these algorithms may be implemented in the factorization of very large matrices with missing entries. The model is evaluated on a collaborative filtering task, where users have rated a collection of movies and the system is asked to predict their ratings for other movies. The Netflix data set is used for evaluation, which consists of around 100 million ratings. Using root mean-squared error (RMSE) as an evaluation metric, results show that the suggested model outperforms alternative factorization techniques. Results also show how Gibbs sampling outperforms variational Bayes on this task, despite the large number of ratings and model parameters. Matlab implementations of the proposed algorithms are available from cogsys.imm.dtu.dk/ordinalmatrixfactorization.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Inference for latent feature models is inherently difficult as the inference space grows exponentially with the size of the input data and number of latent features. In this work, we use Kurihara & Welling (2008)'s maximization-expectation framework to perform approximate MAP inference for linear-Gaussian latent feature models with an Indian Buffet Process (IBP) prior. This formulation yields a submodular function of the features that corresponds to a lower bound on the model evidence. By adding a constant to this function, we obtain a nonnegative submodular function that can be maximized via a greedy algorithm that obtains at least a one-third approximation to the optimal solution. Our inference method scales linearly with the size of the input data, and we show the efficacy of our method on the largest datasets currently analyzed using an IBP model.