984 resultados para Mathematical statistics.
Resumo:
We propose a distribution-free approach to the study of random geometric graphs. The distribution of vertices follows a Poisson point process with intensity function n f(center dot), where n is an element of N, and f is a probability density function on R-d. A vertex located at x connects via directed edges to other vertices that are within a cut-off distance r(n)(x). We prove strong law results for (i) the critical cut-off function so that almost surely, the graph does not contain any node with out-degree zero for sufficiently large n and (ii) the maximum and minimum vertex degrees. We also provide a characterization of the cut-off function for which the number of nodes with out-degree zero converges in distribution to a Poisson random variable. We illustrate this result for a class of densities with compact support that have at most polynomial rates of decay to zero. Finally, we state a sufficient condition for an enhanced version of the above graph to be almost surely connected eventually.
Resumo:
We apply the objective method of Aldous to the problem of finding the minimum-cost edge cover of the complete graph with random independent and identically distributed edge costs. The limit, as the number of vertices goes to infinity, of the expected minimum cost for this problem is known via a combinatorial approach of Hessler and Wastlund. We provide a proof of this result using the machinery of the objective method and local weak convergence, which was used to prove the (2) limit of the random assignment problem. A proof via the objective method is useful because it provides us with more information on the nature of the edge's incident on a typical root in the minimum-cost edge cover. We further show that a belief propagation algorithm converges asymptotically to the optimal solution. This can be applied in a computational linguistics problem of semantic projection. The belief propagation algorithm yields a near optimal solution with lesser complexity than the known best algorithms designed for optimality in worst-case settings.
Resumo:
We show that the density of eigenvalues for three classes of random matrix ensembles is determinantal. First we derive the density of eigenvalues of product of k independent n x n matrices with i.i.d. complex Gaussian entries with a few of matrices being inverted. In second example we calculate the same for (compatible) product of rectangular matrices with i.i.d. Gaussian entries and in last example we calculate for product of independent truncated unitary random matrices. We derive exact expressions for limiting expected empirical spectral distributions of above mentioned ensembles.
Resumo:
本文利用ICP-AES、EPMA、X-ray衍射等测试技术以及聚类分析和因子分析等多种数理统计方法,系统地对中太平洋海区富钴结壳的元素地球化学特征、矿物组成和微观组构进行了研究,并探讨了其成因机制,获得以下主要认识: 1中太平洋海区富钴结壳类型多样,均为水成成因,其矿物相主要由锰矿物相、铁矿物相和非金属矿物相组成;富钴结壳壳层发育多种原生和次生构造类型。 2磷酸盐化作用不仅强烈改变富钴结壳元素初始含量,而且造成富钴结壳某些元素间的相关性发生改变,这些敏感型元素对可用于指示富钴结壳是否发生磷酸盐化。在不同水深段内,未磷酸盐化型富钴结壳的主要元素随经向、纬向的变化趋势相似,表明其受水体化学障、表面生产力和物质来源等环境参数控制;而随水深的变化则具有区域一致性,表明水体化学具有区域成层性。 3未磷酸盐化富钴结壳稀土元素含量和轻重稀土分馏程度随水深发生规律变化,这种变化不仅与它们在海洋中的含量和行为有关,也与海洋背景颗粒的吸附有关;铈(Ce)在富钴结壳中基本上呈4价,且动力学因素控制了其富集过程,因此Ce异常不能用于指示富钴结壳形成环境的氧化程度。 4基于富钴结壳微层呈锯齿状且同一微层生长速率不同,提出了富钴结壳在各种基底表面生长以及后继发育过程受固液界面双电层控制的发育模式。在富钴结壳整个发育过程中,经历了从贫氧环境向富氧环境的转变,但微环境则呈富氧-低氧过程的交替。 关键词: 富钴结壳;中太平洋;元素地球化学;界面双电层
Resumo:
本文结合飞行器电源系统故障维护特点,利用统计过程控制理论,从数理统计的角度对电源系统测试数据进行分析,设计了电源故障预测系统。文中阐述了系统的工作原理和利用LabView软件的实现过程。实验结果表明此预测系统可以发现故障征兆及时维护,提高系统安全性。最后阐述了进一步研究可能会遇到的问题。
Resumo:
This article describes advances in statistical computation for large-scale data analysis in structured Bayesian mixture models via graphics processing unit (GPU) programming. The developments are partly motivated by computational challenges arising in fitting models of increasing heterogeneity to increasingly large datasets. An example context concerns common biological studies using high-throughput technologies generating many, very large datasets and requiring increasingly high-dimensional mixture models with large numbers of mixture components.We outline important strategies and processes for GPU computation in Bayesian simulation and optimization approaches, give examples of the benefits of GPU implementations in terms of processing speed and scale-up in ability to analyze large datasets, and provide a detailed, tutorial-style exposition that will benefit readers interested in developing GPU-based approaches in other statistical models. Novel, GPU-oriented approaches to modifying existing algorithms software design can lead to vast speed-up and, critically, enable statistical analyses that presently will not be performed due to compute time limitations in traditional computational environments. Supplementalmaterials are provided with all source code, example data, and details that will enable readers to implement and explore the GPU approach in this mixture modeling context. © 2010 American Statistical Association, Institute of Mathematical Statistics, and Interface Foundation of North America.
Resumo:
This paper studies the multiplicity-correction effect of standard Bayesian variable-selection priors in linear regression. Our first goal is to clarify when, and how, multiplicity correction happens automatically in Bayesian analysis, and to distinguish this correction from the Bayesian Ockham's-razor effect. Our second goal is to contrast empirical-Bayes and fully Bayesian approaches to variable selection through examples, theoretical results and simulations. Considerable differences between the two approaches are found. In particular, we prove a theorem that characterizes a surprising aymptotic discrepancy between fully Bayes and empirical Bayes. This discrepancy arises from a different source than the failure to account for hyperparameter uncertainty in the empirical-Bayes estimate. Indeed, even at the extreme, when the empirical-Bayes estimate converges asymptotically to the true variable-inclusion probability, the potential for a serious difference remains. © Institute of Mathematical Statistics, 2010.
Resumo:
© 2015, Institute of Mathematical Statistics. All rights reserved.In order to use persistence diagrams as a true statistical tool, it would be very useful to have a good notion of mean and variance for a set of diagrams. In [23], Mileyko and his collaborators made the first study of the properties of the Fréchet mean in (D
Resumo:
Given a probability distribution on an open book (a metric space obtained by gluing a disjoint union of copies of a half-space along their boundary hyperplanes), we define a precise concept of when the Fréchet mean (barycenter) is sticky. This nonclassical phenomenon is quantified by a law of large numbers (LLN) stating that the empirical mean eventually almost surely lies on the (codimension 1 and hence measure 0) spine that is the glued hyperplane, and a central limit theorem (CLT) stating that the limiting distribution is Gaussian and supported on the spine.We also state versions of the LLN and CLT for the cases where the mean is nonsticky (i.e., not lying on the spine) and partly sticky (i.e., is, on the spine but not sticky). © Institute of Mathematical Statistics, 2013.
Resumo:
© Institute of Mathematical Statistics, 2014.Motivated by recent findings in the field of consumer science, this paper evaluates the causal effect of debit cards on household consumption using population-based data from the Italy Survey on Household Income and Wealth (SHIW). Within the Rubin Causal Model, we focus on the estimand of population average treatment effect for the treated (PATT). We consider three existing estimators, based on regression, mixed matching and regression, propensity score weighting, and propose a new doubly-robust estimator. Semiparametric specification based on power series for the potential outcomes and the propensity score is adopted. Cross-validation is used to select the order of the power series. We conduct a simulation study to compare the performance of the estimators. The key assumptions, overlap and unconfoundedness, are systematically assessed and validated in the application. Our empirical results suggest statistically significant positive effects of debit cards on the monthly household spending in Italy.
Resumo:
Student’s t-distribution has found various applications in mathematical statistics. One of the main properties of the t-distribution is to converge to the normal distribution as the number of samples tends to infinity. In this paper, by using a Cauchy integral we introduce a generalization of the t-distribution function with four free parameters and show that it converges to the normal distribution again. We provide a comprehensive treatment of mathematical properties of this new distribution. Moreover, since the Fisher F-distribution has a close relationship with the t-distribution, we also introduce a generalization of the F-distribution and prove that it converges to the chi-square distribution as the number of samples tends to infinity. Finally some particular sub-cases of these distributions are considered.
Resumo:
Observations in daily practice are sometimes registered as positive values larger then a given threshold α. The sample space is in this case the interval (α,+∞), α > 0, which can be structured as a real Euclidean space in different ways. This fact opens the door to alternative statistical models depending not only on the assumed distribution function, but also on the metric which is considered as appropriate, i.e. the way differences are measured, and thus variability