944 resultados para Weighted Poisson distributions
Resumo:
In this article, we propose a new Bayesian flexible cure rate survival model, which generalises the stochastic model of Klebanov et al. [Klebanov LB, Rachev ST and Yakovlev AY. A stochastic-model of radiation carcinogenesis - latent time distributions and their properties. Math Biosci 1993; 113: 51-75], and has much in common with the destructive model formulated by Rodrigues et al. [Rodrigues J, de Castro M, Balakrishnan N and Cancho VG. Destructive weighted Poisson cure rate models. Technical Report, Universidade Federal de Sao Carlos, Sao Carlos-SP. Brazil, 2009 (accepted in Lifetime Data Analysis)]. In our approach, the accumulated number of lesions or altered cells follows a compound weighted Poisson distribution. This model is more flexible than the promotion time cure model in terms of dispersion. Moreover, it possesses an interesting and realistic interpretation of the biological mechanism of the occurrence of the event of interest as it includes a destructive process of tumour cells after an initial treatment or the capacity of an individual exposed to irradiation to repair altered cells that results in cancer induction. In other words, what is recorded is only the damaged portion of the original number of altered cells not eliminated by the treatment or repaired by the repair system of an individual. Markov Chain Monte Carlo (MCMC) methods are then used to develop Bayesian inference for the proposed model. Also, some discussions on the model selection and an illustration with a cutaneous melanoma data set analysed by Rodrigues et al. [Rodrigues J, de Castro M, Balakrishnan N and Cancho VG. Destructive weighted Poisson cure rate models. Technical Report, Universidade Federal de Sao Carlos, Sao Carlos-SP. Brazil, 2009 (accepted in Lifetime Data Analysis)] are presented.
Resumo:
"Vegeu el resum a l'inici del document del fitxer adjunt"
Resumo:
In this paper, we develop a flexible cure rate survival model by assuming the number of competing causes of the event of interest to follow a compound weighted Poisson distribution. This model is more flexible in terms of dispersion than the promotion time cure model. Moreover, it gives an interesting and realistic interpretation of the biological mechanism of the occurrence of event of interest as it includes a destructive process of the initial risk factors in a competitive scenario. In other words, what is recorded is only from the undamaged portion of the original number of risk factors.
Resumo:
In this paper, we develop a flexible cure rate survival model by assuming the number of competing causes of the event of interest to follow the Conway-Maxwell Poisson distribution. This model includes as special cases some of the well-known cure rate models discussed in the literature. Next, we discuss the maximum likelihood estimation of the parameters of this cure rate survival model. Finally, we illustrate the usefulness of this model by applying it to a real cutaneous melanoma data. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
The general properties of POISSON distributions and their relations to the binomial distribuitions are discussed. Two methods of statistical analysis are dealt with in detail: X2-test. In order to carry out the X2-test, the mean frequency and the theoretical frequencies for all classes are calculated. Than the observed and the calculated frequencies are compared, using the well nown formula: f(obs) - f(esp) 2; i(esp). When the expected frequencies are small, one must not forget that the value of X2 may only be calculated, if the expected frequencies are biger than 5. If smaller values should occur, the frequencies of neighboroughing classes must ge pooled. As a second test reintroduced by BRIEGER, consists in comparing the observed and expected error standard of the series. The observed error is calculated by the general formula: δ + Σ f . VK n-1 where n represents the number of cases. The theoretical error of a POISSON series with mean frequency m is always ± Vm. These two values may be compared either by dividing the observed by the theoretical error and using BRIEGER's tables for # or by dividing the respective variances and using SNEDECOR's tables for F. The degree of freedom for the observed error is one less the number of cases studied, and that of the theoretical error is always infinite. In carrying out these tests, one important point must never be overlloked. The values for the first class, even if no concrete cases of the type were observed, must always be zero, an dthe value of the subsequent classes must be 1, 2, 3, etc.. This is easily seen in some of the classical experiments. For instance in BORKEWITZ example of accidents in Prussian armee corps, the classes are: no, one, two, etc., accidents. When counting the frequency of bacteria, these values are: no, one, two, etc., bacteria or cultures of bacteria. Ins studies of plant diseases equally the frequencies are : no, one, two, etc., plants deseased. Howewer more complicated cases may occur. For instance, when analising the degree of polyembriony, frequently the case of "no polyembryony" corresponds to the occurrence of one embryo per each seed. Thus the classes are not: no, one, etc., embryo per seed, but they are: no additional embryo, one additional embryo, etc., per seed with at least one embryo. Another interestin case was found by BRIEGER in genetic studies on the number os rows in maize. Here the minimum number is of course not: no rows, but: no additional beyond eight rows. The next class is not: nine rows, but: 10 rows, since the row number varies always in pairs of rows. Thus the value of successive classes are: no additional pair of rows beyond 8, one additional pair (or 10 rows), two additional pairs (or 12 rows) etc.. The application of the methods is finally shown on the hand of three examples : the number of seeds per fruit in the oranges M Natal" and "Coco" and in "Calamondin". As shown in the text and the tables, the agreement with a POISSON series is very satisfactory in the first two cases. In the third case BRIEGER's error test indicated a significant reduction of variability, and the X2 test showed that there were two many fruits with 4 or 5 seeds and too few with more or with less seeds. Howewer the fact that no fruit was found without seed, may be taken to indicate that in Calamondin fruits are not fully parthenocarpic and may develop only with one seed at the least. Thus a new analysis was carried out, on another class basis. As value for the first class the following value was accepted: no additional seed beyond the indispensable minimum number of one seed, and for the later classes the values were: one, two, etc., additional seeds. Using this new basis for all calculations, a complete agreement of the observed and expected frequencies, of the correspondig POISSON series was obtained, thus proving that our hypothesis of the impossibility of obtaining fruits without any seed was correct for Calamondin while the other two oranges were completely parthenocarpic and fruits without seeds did occur.
Resumo:
It has been argued that by truncating the sample space of the negative binomial and of the inverse Gaussian-Poisson mixture models at zero, one is allowed to extend the parameter space of the model. Here that is proved to be the case for the more general three parameter Tweedie-Poisson mixture model. It is also proved that the distributions in the extended part of the parameter space are not the zero truncation of mixed poisson distributions and that, other than for the negative binomial, they are not mixtures of zero truncated Poisson distributions either. By extending the parameter space one can improve the fit when the frequency of one is larger and the right tail is heavier than is allowed by the unextended model. Considering the extended model also allows one to use the basic maximum likelihood based inference tools when parameter estimates fall in the extended part of the parameter space, and hence when the m.l.e. does not exist under the unextended model. This extended truncated Tweedie-Poisson model is proved to be useful in the analysis of words and species frequency count data.
Resumo:
In this paper, we proposed a flexible cure rate survival model by assuming the number of competing causes of the event of interest following the Conway-Maxwell distribution and the time for the event to follow the generalized gamma distribution. This distribution can be used to model survival data when the hazard rate function is increasing, decreasing, bathtub and unimodal-shaped including some distributions commonly used in lifetime analysis as particular cases. Some appropriate matrices are derived in order to evaluate local influence on the estimates of the parameters by considering different perturbations, and some global influence measurements are also investigated. Finally, data set from the medical area is analysed.
Resumo:
The Conway-Maxwell Poisson (COMP) distribution as an extension of the Poisson distribution is a popular model for analyzing counting data. For the first time, we introduce a new three parameter distribution, so-called the exponential-Conway-Maxwell Poisson (ECOMP) distribution, that contains as sub-models the exponential-geometric and exponential-Poisson distributions proposed by Adamidis and Loukas (Stat Probab Lett 39:35-42, 1998) and KuAY (Comput Stat Data Anal 51:4497-4509, 2007), respectively. The new density function can be expressed as a mixture of exponential density functions. Expansions for moments, moment generating function and some statistical measures are provided. The density function of the order statistics can also be expressed as a mixture of exponential densities. We derive two formulae for the moments of order statistics. The elements of the observed information matrix are provided. Two applications illustrate the usefulness of the new distribution to analyze positive data.
Resumo:
In species subject to individual and social learning, each individual is likely to express a certain number of different cultural traits acquired during its lifetime. If the process of trait innovation and transmission reaches a steady state in the population, the number of different cultural traits carried by an individual converges to some stationary distribution. We call this the trait-number distribution. In this paper, we derive the trait-number distributions for both individuals and populations when cultural traits are independent of each other. Our results suggest that as the number of cultural traits becomes large, the trait-number distributions approach Poisson distributions so that their means characterize cultural diversity in the population. We then analyse how the mean trait number varies at both the individual and population levels as a function of various demographic features, such as population size and subdivision, and social learning rules, such as conformism and anti-conformism. Diversity at the individual and population levels, as well as at the level of cultural homogeneity within groups, depends critically on the details of population demography and the individual and social learning rules.
Resumo:
Two simple and frequently used capture–recapture estimates of the population size are compared: Chao's lower-bound estimate and Zelterman's estimate allowing for contaminated distributions. In the Poisson case it is shown that if there are only counts of ones and twos, the estimator of Zelterman is always bounded above by Chao's estimator. If counts larger than two exist, the estimator of Zelterman is becoming larger than that of Chao's, if only the ratio of the frequencies of counts of twos and ones is small enough. A similar analysis is provided for the binomial case. For a two-component mixture of Poisson distributions the asymptotic bias of both estimators is derived and it is shown that the Zelterman estimator can experience large overestimation bias. A modified Zelterman estimator is suggested and also the bias-corrected version of Chao's estimator is considered. All four estimators are compared in a simulation study.
Resumo:
The paper concerns the design and analysis of serial dilution assays to estimate the infectivity of a sample of tissue when it is assumed that the sample contains a finite number of indivisible infectious units such that a subsample will be infectious if it contains one or more of these units. The aim of the study is to estimate the number of infectious units in the original sample. The standard approach to the analysis of data from such a study is based on the assumption of independence of aliquots both at the same dilution level and at different dilution levels, so that the numbers of infectious units in the aliquots follow independent Poisson distributions. An alternative approach is based on calculation of the expected value of the total number of samples tested that are not infectious. We derive the likelihood for the data on the basis of the discrete number of infectious units, enabling calculation of the maximum likelihood estimate and likelihood-based confidence intervals. We use the exact probabilities that are obtained to compare the maximum likelihood estimate with those given by the other methods in terms of bias and standard error and to compare the coverage of the confidence intervals. We show that the methods have very similar properties and conclude that for practical use the method that is based on the Poisson assumption is to be recommended, since it can be implemented by using standard statistical software. Finally we consider the design of serial dilution assays, concluding that it is important that neither the dilution factor nor the number of samples that remain untested should be too large.
Resumo:
Two simple and frequently used capture–recapture estimates of the population size are compared: Chao's lower-bound estimate and Zelterman's estimate allowing for contaminated distributions. In the Poisson case it is shown that if there are only counts of ones and twos, the estimator of Zelterman is always bounded above by Chao's estimator. If counts larger than two exist, the estimator of Zelterman is becoming larger than that of Chao's, if only the ratio of the frequencies of counts of twos and ones is small enough. A similar analysis is provided for the binomial case. For a two-component mixture of Poisson distributions the asymptotic bias of both estimators is derived and it is shown that the Zelterman estimator can experience large overestimation bias. A modified Zelterman estimator is suggested and also the bias-corrected version of Chao's estimator is considered. All four estimators are compared in a simulation study.
Resumo:
Estimation of a population size by means of capture-recapture techniques is an important problem occurring in many areas of life and social sciences. We consider the frequencies of frequencies situation, where a count variable is used to summarize how often a unit has been identified in the target population of interest. The distribution of this count variable is zero-truncated since zero identifications do not occur in the sample. As an application we consider the surveillance of scrapie in Great Britain. In this case study holdings with scrapie that are not identified (zero counts) do not enter the surveillance database. The count variable of interest is the number of scrapie cases per holding. For count distributions a common model is the Poisson distribution and, to adjust for potential heterogeneity, a discrete mixture of Poisson distributions is used. Mixtures of Poissons usually provide an excellent fit as will be demonstrated in the application of interest. However, as it has been recently demonstrated, mixtures also suffer under the so-called boundary problem, resulting in overestimation of population size. It is suggested here to select the mixture model on the basis of the Bayesian Information Criterion. This strategy is further refined by employing a bagging procedure leading to a series of estimates of population size. Using the median of this series, highly influential size estimates are avoided. In limited simulation studies it is shown that the procedure leads to estimates with remarkable small bias.
Resumo:
We analyse in a common framework the properties of the Voronoi tessellations resulting from regular 2D and 3D crystals and those of tessellations generated by Poisson distributions of points, thus joining on symmetry breaking processes and the approach to uniform random distributions of seeds. We perturb crystalline structures in 2D and 3D with a spatial Gaussian noise whose adimensional strength is α and analyse the statistical properties of the cells of the resulting Voronoi tessellations using an ensemble approach. In 2D we consider triangular, square and hexagonal regular lattices, resulting into hexagonal, square and triangular tessellations, respectively. In 3D we consider the simple cubic (SC), body-centred cubic (BCC), and face-centred cubic (FCC) crystals, whose corresponding Voronoi cells are the cube, the truncated octahedron, and the rhombic dodecahedron, respectively. In 2D, for all values α>0, hexagons constitute the most common class of cells. Noise destroys the triangular and square tessellations, which are structurally unstable, as their topological properties are discontinuous in α=0. On the contrary, the honeycomb hexagonal tessellation is topologically stable and, experimentally, all Voronoi cells are hexagonal for small but finite noise with α<0.12. Basically, the same happens in the 3D case, where only the tessellation of the BCC crystal is topologically stable even against noise of small but finite intensity. In both 2D and 3D cases, already for a moderate amount of Gaussian noise (α>0.5), memory of the specific initial unperturbed state is lost, because the statistical properties of the three perturbed regular tessellations are indistinguishable. When α>2, results converge to those of Poisson-Voronoi tessellations. In 2D, while the isoperimetric ratio increases with noise for the perturbed hexagonal tessellation, for the perturbed triangular and square tessellations it is optimised for specific value of noise intensity. The same applies in 3D, where noise degrades the isoperimetric ratio for perturbed FCC and BCC lattices, whereas the opposite holds for perturbed SCC lattices. This allows for formulating a weaker form of the Kelvin conjecture. By analysing jointly the statistical properties of the area and of the volume of the cells, we discover that also the cells shape heavily fluctuates when noise is introduced in the system. In 2D, the geometrical properties of n-sided cells change with α until the Poisson-Voronoi limit is reached for α>2; in this limit the Desch law for perimeters is shown to be not valid and a square root dependence on n is established, which agrees with exact asymptotic results. Anomalous scaling relations are observed between the perimeter and the area in the 2D and between the areas and the volumes of the cells in 3D: except for the hexagonal (2D) and FCC structure (3D), this applies also for infinitesimal noise. In the Poisson-Voronoi limit, the anomalous exponent is about 0.17 in both the 2D and 3D case. A positive anomaly in the scaling indicates that large cells preferentially feature large isoperimetric quotients. As the number of faces is strongly correlated with the sphericity (cells with more faces are bulkier), in 3D it is shown that the anomalous scaling is heavily reduced when we perform power law fits separately on cells with a specific number of faces.
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)