971 resultados para Likelihood Functions
Resumo:
We propose a simple method of constructing quasi-likelihood functions for dependent data based on conditional-mean-variance relationships, and apply the method to estimating the fractal dimension from box-counting data. Simulation studies were carried out to compare this method with the traditional methods. We also applied this technique to real data from fishing grounds in the Gulf of Carpentaria, Australia
Resumo:
This paper presents a framework for performing real-time recursive estimation of landmarks’ visual appearance. Imaging data in its original high dimensional space is probabilistically mapped to a compressed low dimensional space through the definition of likelihood functions. The likelihoods are subsequently fused with prior information using a Bayesian update. This process produces a probabilistic estimate of the low dimensional representation of the landmark visual appearance. The overall filtering provides information complementary to the conventional position estimates which is used to enhance data association. In addition to robotics observations, the filter integrates human observations in the appearance estimates. The appearance tracks as computed by the filter allow landmark classification. The set of labels involved in the classification task is thought of as an observation space where human observations are made by selecting a label. The low dimensional appearance estimates returned by the filter allow for low cost communication in low bandwidth sensor networks. Deployment of the filter in such a network is demonstrated in an outdoor mapping application involving a human operator, a ground and an air vehicle.
Resumo:
In this paper we present a new simulation methodology in order to obtain exact or approximate Bayesian inference for models for low-valued count time series data that have computationally demanding likelihood functions. The algorithm fits within the framework of particle Markov chain Monte Carlo (PMCMC) methods. The particle filter requires only model simulations and, in this regard, our approach has connections with approximate Bayesian computation (ABC). However, an advantage of using the PMCMC approach in this setting is that simulated data can be matched with data observed one-at-a-time, rather than attempting to match on the full dataset simultaneously or on a low-dimensional non-sufficient summary statistic, which is common practice in ABC. For low-valued count time series data we find that it is often computationally feasible to match simulated data with observed data exactly. Our particle filter maintains $N$ particles by repeating the simulation until $N+1$ exact matches are obtained. Our algorithm creates an unbiased estimate of the likelihood, resulting in exact posterior inferences when included in an MCMC algorithm. In cases where exact matching is computationally prohibitive, a tolerance is introduced as per ABC. A novel aspect of our approach is that we introduce auxiliary variables into our particle filter so that partially observed and/or non-Markovian models can be accommodated. We demonstrate that Bayesian model choice problems can be easily handled in this framework.
Resumo:
Analytically or computationally intractable likelihood functions can arise in complex statistical inferential problems making them inaccessible to standard Bayesian inferential methods. Approximate Bayesian computation (ABC) methods address such inferential problems by replacing direct likelihood evaluations with repeated sampling from the model. ABC methods have been predominantly applied to parameter estimation problems and less to model choice problems due to the added difficulty of handling multiple model spaces. The ABC algorithm proposed here addresses model choice problems by extending Fearnhead and Prangle (2012, Journal of the Royal Statistical Society, Series B 74, 1–28) where the posterior mean of the model parameters estimated through regression formed the summary statistics used in the discrepancy measure. An additional stepwise multinomial logistic regression is performed on the model indicator variable in the regression step and the estimated model probabilities are incorporated into the set of summary statistics for model choice purposes. A reversible jump Markov chain Monte Carlo step is also included in the algorithm to increase model diversity for thorough exploration of the model space. This algorithm was applied to a validating example to demonstrate the robustness of the algorithm across a wide range of true model probabilities. Its subsequent use in three pathogen transmission examples of varying complexity illustrates the utility of the algorithm in inferring preference of particular transmission models for the pathogens.
Resumo:
Background: A knowledge of energy expenditure in infancy is required for the estimation of recommended daily amounts of food energy, for designing artificial infant feeds, and as a reference standard for studies of energy metabolism in disease states. Objectives: The objectives of this study were to construct centile reference charts for total energy expenditure (TEE) in infants across the first year of life. Methods: Repeated measures of TEE using the doubly labeled water technique were made in 162 infants at 1.5, 3, 6, 9 and 12 months. In total, 322 TEE measurements were obtained. The LMS method with maximum penalized likelihood was used to construct the centile reference charts. Centiles were constructed for TEE expressed as MJ/day and also expressed relative to body weight (BW) and fat-free mass (FFM). Results: TEE increased with age and was 1.40,1.86, 2.64, 3.07 and 3.65 MJ/day at 1.5, 3, 6, 9 and 12 months, respectively. The standard deviations were 0.43, 0.47, 0.52,0.66 and 0.88, respectively. TEE in MJ/kg increased from 0.29 to 0.36 and in MJ/day/kg FFM from 0.36 to 0.48. Conclusions: We have presented centile reference charts for TEE expressed as MJ/day and expressed relative to BW and FFM in infants across the first year of life. There was a wide variation or biological scatter in TEE values seen at all ages. We suggest that these centile charts may be used to assess and possibly quantify abnormal energy metabolism in disease states in infants.
Resumo:
OBJECTIVE: To further investigate a common variant (rs9939609) in the fat mass- and obesity-associated gene (FTO), which recent genome-wide association studies have shown to be associated with body mass index (BMI) and obesity. DESIGN: We examined the effect of this FTO variant on BMI in 3353 Australian adult male and female twins. RESULTS: The minor A allele of rs9939609 was associated with an increased BMI (P=0.0007). Each additional copy of the A allele was associated with a mean BMI increase of approximately 1.04 kg/m(2) (approximately 3.71 kg). Using variance components decomposition, we estimate that this single-nucleotide polymorphism accounts for approximately 3% of the genetic variance in BMI in our sample (approximately 2% of the total variance). By comparing intrapair variances of monozygotic twins of different genotypes we were able to perform a direct test of gene by environment (G x E) interaction in both sexes and gene by parity (G x P) interaction in women, but no evidence was found for either. CONCLUSIONS: In addition to supporting earlier findings that the rs9939609 variant in the FTO gene is associated with an increased BMI, our results indicate that the associated genetic effect does not interact with environment or parity.
Resumo:
We report the observation of electroweak single top quark production in 3.2 fb-1 of pp̅ collision data collected by the Collider Detector at Fermilab at √s=1.96 TeV. Candidate events in the W+jets topology with a leptonically decaying W boson are classified as signal-like by four parallel analyses based on likelihood functions, matrix elements, neural networks, and boosted decision trees. These results are combined using a super discriminant analysis based on genetically evolved neural networks in order to improve the sensitivity. This combined result is further combined with that of a search for a single top quark signal in an orthogonal sample of events with missing transverse energy plus jets and no charged lepton. We observe a signal consistent with the standard model prediction but inconsistent with the background-only model by 5.0 standard deviations, with a median expected sensitivity in excess of 5.9 standard deviations. We measure a production cross section of 2.3-0.5+0.6(stat+sys) pb, extract the value of the Cabibbo-Kobayashi-Maskawa matrix element |Vtb|=0.91-0.11+0.11(stat+sys)±0.07 (theory), and set a lower limit |Vtb|>0.71 at the 95% C.L., assuming mt=175 GeV/c2.
Resumo:
We report the observation of electroweak single top quark production in 3.2 fb-1 of ppbar collision data collected by the Collider Detector at Fermilab at sqrt{s}=1.96 TeV. Candidate events in the W+jets topology with a leptonically decaying W boson are classified as signal-like by four parallel analyses based on likelihood functions, matrix elements, neural networks, and boosted decision trees. These results are combined using a super discriminant analysis based on genetically evolved neural networks in order to improve the sensitivity. This combined result is further combined with that of a search for a single top quark signal in an orthogonal sample of events with missing transverse energy plus jets and no charged lepton. We observe a signal consistent with the standard model prediction but inconsistent with the background-only model by 5.0 standard deviations, with a median expected sensitivity in excess of 5.9 standard deviations. We measure a production cross section of 2.3+0.6-0.5(stat+sys) pb, extract the CKM matrix element value |Vtb|=0.91+0.11-0.11 (stat+sys)+-0.07(theory), and set a lower limit |Vtb|>0.71 at the 95% confidence level, assuming m_t=175 GeVc^2.
Resumo:
Although approximate Bayesian computation (ABC) has become a popular technique for performing parameter estimation when the likelihood functions are analytically intractable there has not as yet been a complete investigation of the theoretical properties of the resulting estimators. In this paper we give a theoretical analysis of the asymptotic properties of ABC based parameter estimators for hidden Markov models and show that ABC based estimators satisfy asymptotically biased versions of the standard results in the statistical literature.
Resumo:
English: We describe an age-structured statistical catch-at-length analysis (A-SCALA) based on the MULTIFAN-CL model of Fournier et al. (1998). The analysis is applied independently to both the yellowfin and the bigeye tuna populations of the eastern Pacific Ocean (EPO). We model the populations from 1975 to 1999, based on quarterly time steps. Only a single stock for each species is assumed for each analysis, but multiple fisheries that are spatially separate are modeled to allow for spatial differences in catchability and selectivity. The analysis allows for error in the effort-fishing mortality relationship, temporal trends in catchability, temporal variation in recruitment, relationships between the environment and recruitment and between the environment and catchability, and differences in selectivity and catchability among fisheries. The model is fit to total catch data and proportional catch-at-length data conditioned on effort. The A-SCALA method is a statistical approach, and therefore recognizes that the data collected from the fishery do not perfectly represent the population. Also, there is uncertainty in our knowledge about the dynamics of the system and uncertainty about how the observed data relate to the real population. The use of likelihood functions allow us to model the uncertainty in the data collected from the population, and the inclusion of estimable process error allows us to model the uncertainties in the dynamics of the system. The statistical approach allows for the calculation of confidence intervals and the testing of hypotheses. We use a Bayesian version of the maximum likelihood framework that includes distributional constraints on temporal variation in recruitment, the effort-fishing mortality relationship, and catchability. Curvature penalties for selectivity parameters and penalties on extreme fishing mortality rates are also included in the objective function. The mode of the joint posterior distribution is used as an estimate of the model parameters. Confidence intervals are calculated using the normal approximation method. It should be noted that the estimation method includes constraints and priors and therefore the confidence intervals are different from traditionally calculated confidence intervals. Management reference points are calculated, and forward projections are carried out to provide advice for making management decisions for the yellowfin and bigeye populations. Spanish: Describimos un análisis estadístico de captura a talla estructurado por edad, A-SCALA (del inglés age-structured statistical catch-at-length analysis), basado en el modelo MULTIFAN- CL de Fournier et al. (1998). Se aplica el análisis independientemente a las poblaciones de atunes aleta amarilla y patudo del Océano Pacífico oriental (OPO). Modelamos las poblaciones de 1975 a 1999, en pasos trimestrales. Se supone solamente una sola población para cada especie para cada análisis, pero se modelan pesquerías múltiples espacialmente separadas para tomar en cuenta diferencias espaciales en la capturabilidad y selectividad. El análisis toma en cuenta error en la relación esfuerzo-mortalidad por pesca, tendencias temporales en la capturabilidad, variación temporal en el reclutamiento, relaciones entre el medio ambiente y el reclutamiento y entre el medio ambiente y la capturabilidad, y diferencias en selectividad y capturabilidad entre pesquerías. Se ajusta el modelo a datos de captura total y a datos de captura a talla proporcional condicionados sobre esfuerzo. El método A-SCALA es un enfoque estadístico, y reconoce por lo tanto que los datos obtenidos de la pesca no representan la población perfectamente. Además, hay incertidumbre en nuestros conocimientos de la dinámica del sistema e incertidumbre sobre la relación entre los datos observados y la población real. El uso de funciones de verosimilitud nos permite modelar la incertidumbre en los datos obtenidos de la población, y la inclusión de un error de proceso estimable nos permite modelar las incertidumbres en la dinámica del sistema. El enfoque estadístico permite calcular intervalos de confianza y comprobar hipótesis. Usamos una versión bayesiana del marco de verosimilitud máxima que incluye constreñimientos distribucionales sobre la variación temporal en el reclutamiento, la relación esfuerzo-mortalidad por pesca, y la capturabilidad. Se incluyen también en la función objetivo penalidades por curvatura para los parámetros de selectividad y penalidades por tasas extremas de mortalidad por pesca. Se usa la moda de la distribución posterior conjunta como estimación de los parámetros del modelo. Se calculan los intervalos de confianza usando el método de aproximación normal. Cabe destacar que el método de estimación incluye constreñimientos y distribuciones previas y por lo tanto los intervalos de confianza son diferentes de los intervalos de confianza calculados de forma tradicional. Se calculan puntos de referencia para el ordenamiento, y se realizan proyecciones a futuro para asesorar la toma de decisiones para el ordenamiento de las poblaciones de aleta amarilla y patudo.
Resumo:
The GPML toolbox provides a wide range of functionality for Gaussian process (GP) inference and prediction. GPs are specified by mean and covariance functions; we offer a library of simple mean and covariance functions and mechanisms to compose more complex ones. Several likelihood functions are supported including Gaussian and heavy-tailed for regression as well as others suitable for classification. Finally, a range of inference methods is provided, including exact and variational inference, Expectation Propagation, and Laplace’s method dealing with non-Gaussian likelihoods and FITC for dealing with large regression tasks.
Resumo:
The code provided here originally demonstrated the main algorithms from Rasmussen and Williams: Gaussian Processes for Machine Learning. It has since grown to allow more likelihood functions, further inference methods and a flexible framework for specifying GPs.
Resumo:
The Rhizopus oryzae species complex is a group of zygomycete fungi that are common, cosmopolitan saprotrophs. Some strains are used beneficially for production of Asian fermented foods but they can also act as opportunistic human pathogens. Although R. oryzae reportedly has a heterothallic (+/-) mating system, most strains have not been observed to undergo sexual reproduction and the genetic structure of its mating locus has not been characterized. Here we report on the mating behavior and genetic structure of the mating locus for 54 isolates of the R. oryzae complex. All 54 strains have a mating locus similar in overall organization to Phycomyces blakesleeanus and Mucor circinelloides (Mucoromycotina, Zygomycota). In all of these fungi, the minus (-) allele features the SexM high mobility group (HMG) gene flanked by an RNA helicase gene and a TP transporter gene (TPT). Within the R. oryzae complex, the plus (+) mating allele includes an inserted region that codes for a BTB/POZ domain gene and the SexP HMG gene. Phylogenetic analyses of multiple genes, including the mating loci (HMG, TPT, RNA helicase), ITS1-5.8S-ITS2 rDNA, RPB2, and LDH genes, identified two distinct groups of strains. These correspond to previously described sibling species R. oryzae sensu stricto and R. delemar. Within each species, discordant gene phylogenies among multiple loci suggest an outcrossing population structure. The hypothesis of random-mating is also supported by a 50:50 ratio of plus and minus mating types in both cryptic species. When crossed with tester strains of the opposite mating type, most isolates of R. delemar failed to produce zygospores, while isolates of R. oryzae produced sterile zygospores. In spite of the reluctance of most strains to mate in vitro, the conserved sex locus structure and evidence for outcrossing suggest that a normal sexual cycle occurs in both species.
Resumo:
Population introduction is an important tool for ecosystem restoration. However, before introductions should be conducted, it is important to evaluate the genetic, phenotypic and ecological suitability of possible replacement populations. Careful genetic analysis is particularly important if it is suspected that the extirpated population was unique or genetically divergent. On the island of Martha's Vineyard, Massachusetts, the introduction of greater prairie chickens (Tympanuchus cupido pinnatus) to replace the extinct heath hen (T. cupido cupido) is being considered as part of an ecosystem restoration project. Martha's Vineyard was home to the last remaining heath hen population until its extinction in 1932. We conducted this study to aid in determining the suitability of greater prairie chickens as a possible replacement for the heath hen. We examined mitochondrial control region sequences from extant populations of all prairie grouse species (Tympanuchus) and from museum skin heath hen specimens. Our data suggest that the Martha's Vineyard heath hen population represents a divergent mitochondrial lineage. This result is attributable either to a long period of geographical isolation from other prairie grouse populations or to a population bottleneck resulting from human disturbance. The mtDNA diagnosability of the heath hen contrasts with the network of mtDNA haplotypes of other prairie grouse (T. cupido attwateri, T. pallidicinctus and T. phasianellus), which do not form distinguishable mtDNA groupings. Our findings suggest that the Martha's Vineyard heath hen was more genetically isolated than are current populations of prairie grouse and place the emphasis for future research on examining prairie grouse adaptations to different habitat types to assess ecological exchangeability between heath hens and greater prairie chickens.