35 resultados para MAXIMUM PENALIZED LIKELIHOOD ESTIMATES
Resumo:
[1] In many practical situations where spatial rainfall estimates are needed, rainfall occurs as a spatially intermittent phenomenon. An efficient geostatistical method for rainfall estimation in the case of intermittency has previously been published and comprises the estimation of two independent components: a binary random function for modeling the intermittency and a continuous random function that models the rainfall inside the rainy areas. The final rainfall estimates are obtained as the product of the estimates of these two random functions. However the published approach does not contain a method for estimation of uncertainties. The contribution of this paper is the presentation of the indicator maximum likelihood estimator from which the local conditional distribution of the rainfall value at any location may be derived using an ensemble approach. From the conditional distribution, representations of uncertainty such as the estimation variance and confidence intervals can be obtained. An approximation to the variance can be calculated more simply by assuming rainfall intensity is independent of location within the rainy area. The methodology has been validated using simulated and real rainfall data sets. The results of these case studies show good agreement between predicted uncertainties and measured errors obtained from the validation data.
Resumo:
We use geomagnetic activity data to study the rise and fall over the past century of the solar wind flow speed VSW, the interplanetary magnetic field strength B, and the open solar flux FS. Our estimates include allowance for the kinematic effect of longitudinal structure in the solar wind flow speed. As well as solar cycle variations, all three parameters show a long-term rise during the first half of the 20th century followed by peaks around 1955 and 1986 and then a recent decline. Cosmogenic isotope data reveal that this constitutes a grand maximum of solar activity which began in 1920, using the definition that such grand maxima are when 25-year averages of the heliospheric modulation potential exceeds 600 MV. Extrapolating the linear declines seen in all three parameters since 1985, yields predictions that the grand maximum will end in the years 2013, 2014, or 2027 using VSW, FS, or B, respectively. These estimates are consistent with predictions based on the probability distribution of the durations of past grand solar maxima seen in cosmogenic isotope data. The data contradict any suggestions of a floor to the open solar flux: we show that the solar minimum open solar flux, kinematically corrected to allow for the excess flux effect, has halved over the past two solar cycles.
Resumo:
The article considers screening human populations with two screening tests. If any of the two tests is positive, then full evaluation of the disease status is undertaken; however, if both diagnostic tests are negative, then disease status remains unknown. This procedure leads to a data constellation in which, for each disease status, the 2 × 2 table associated with the two diagnostic tests used in screening has exactly one empty, unknown cell. To estimate the unobserved cell counts, previous approaches assume independence of the two diagnostic tests and use specific models, including the special mixture model of Walter or unconstrained capture–recapture estimates. Often, as is also demonstrated in this article by means of a simple test, the independence of the two screening tests is not supported by the data. Two new estimators are suggested that allow associations of the screening test, although the form of association must be assumed to be homogeneous over disease status. These estimators are modifications of the simple capture–recapture estimator and easy to construct. The estimators are investigated for several screening studies with fully evaluated disease status in which the superior behavior of the new estimators compared to the previous conventional ones can be shown. Finally, the performance of the new estimators is compared with maximum likelihood estimators, which are more difficult to obtain in these models. The results indicate the loss of efficiency as minor.
Resumo:
The likelihood for the Logit model is modified, so as to take account of uncertainty associated with mis-reporting in stated preference experiments estimating willingness to pay (WTP). Monte Carlo results demonstrate the bias imparted to estimates where there is mis-reporting. The approach is applied to a data set examining consumer preferences for food produced employing a nonpesticide technology. Our modified approach leads to WTP that are substantially downwardly revised.
Resumo:
Background: The present paper investigates the question of a suitable basic model for the number of scrapie cases in a holding and applications of this knowledge to the estimation of scrapie-ffected holding population sizes and adequacy of control measures within holding. Is the number of scrapie cases proportional to the size of the holding in which case it should be incorporated into the parameter of the error distribution for the scrapie counts? Or, is there a different - potentially more complex - relationship between case count and holding size in which case the information about the size of the holding should be better incorporated as a covariate in the modeling? Methods: We show that this question can be appropriately addressed via a simple zero-truncated Poisson model in which the hypothesis of proportionality enters as a special offset-model. Model comparisons can be achieved by means of likelihood ratio testing. The procedure is illustrated by means of surveillance data on classical scrapie in Great Britain. Furthermore, the model with the best fit is used to estimate the size of the scrapie-affected holding population in Great Britain by means of two capture-recapture estimators: the Poisson estimator and the generalized Zelterman estimator. Results: No evidence could be found for the hypothesis of proportionality. In fact, there is some evidence that this relationship follows a curved line which increases for small holdings up to a maximum after which it declines again. Furthermore, it is pointed out how crucial the correct model choice is when applied to capture-recapture estimation on the basis of zero-truncated Poisson models as well as on the basis of the generalized Zelterman estimator. Estimators based on the proportionality model return very different and unreasonable estimates for the population sizes. Conclusion: Our results stress the importance of an adequate modelling approach to the association between holding size and the number of cases of classical scrapie within holding. Reporting artefacts and speculative biological effects are hypothesized as the underlying causes of the observed curved relationship. The lack of adjustment for these artefacts might well render ineffective the current strategies for the control of the disease.
Resumo:
The article considers screening human populations with two screening tests. If any of the two tests is positive, then full evaluation of the disease status is undertaken; however, if both diagnostic tests are negative, then disease status remains unknown. This procedure leads to a data constellation in which, for each disease status, the 2 x 2 table associated with the two diagnostic tests used in screening has exactly one empty, unknown cell. To estimate the unobserved cell counts, previous approaches assume independence of the two diagnostic tests and use specific models, including the special mixture model of Walter or unconstrained capture-recapture estimates. Often, as is also demonstrated in this article by means of a simple test, the independence of the two screening tests is not supported by the data. Two new estimators are suggested that allow associations of the screening test, although the form of association must be assumed to be homogeneous over disease status. These estimators are modifications of the simple capture-recapture estimator and easy to construct. The estimators are investigated for several screening studies with fully evaluated disease status in which the superior behavior of the new estimators compared to the previous conventional ones can be shown. Finally, the performance of the new estimators is compared with maximum likelihood estimators, which are more difficult to obtain in these models. The results indicate the loss of efficiency as minor.
Resumo:
Estimation of population size with missing zero-class is an important problem that is encountered in epidemiological assessment studies. Fitting a Poisson model to the observed data by the method of maximum likelihood and estimation of the population size based on this fit is an approach that has been widely used for this purpose. In practice, however, the Poisson assumption is seldom satisfied. Zelterman (1988) has proposed a robust estimator for unclustered data that works well in a wide class of distributions applicable for count data. In the work presented here, we extend this estimator to clustered data. The estimator requires fitting a zero-truncated homogeneous Poisson model by maximum likelihood and thereby using a Horvitz-Thompson estimator of population size. This was found to work well, when the data follow the hypothesized homogeneous Poisson model. However, when the true distribution deviates from the hypothesized model, the population size was found to be underestimated. In the search of a more robust estimator, we focused on three models that use all clusters with exactly one case, those clusters with exactly two cases and those with exactly three cases to estimate the probability of the zero-class and thereby use data collected on all the clusters in the Horvitz-Thompson estimator of population size. Loss in efficiency associated with gain in robustness was examined based on a simulation study. As a trade-off between gain in robustness and loss in efficiency, the model that uses data collected on clusters with at most three cases to estimate the probability of the zero-class was found to be preferred in general. In applications, we recommend obtaining estimates from all three models and making a choice considering the estimates from the three models, robustness and the loss in efficiency. (© 2008 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim)
Resumo:
This paper considers the problem of estimation when one of a number of populations, assumed normal with known common variance, is selected on the basis of it having the largest observed mean. Conditional on selection of the population, the observed mean is a biased estimate of the true mean. This problem arises in the analysis of clinical trials in which selection is made between a number of experimental treatments that are compared with each other either with or without an additional control treatment. Attempts to obtain approximately unbiased estimates in this setting have been proposed by Shen [2001. An improved method of evaluating drug effect in a multiple dose clinical trial. Statist. Medicine 20, 1913–1929] and Stallard and Todd [2005. Point estimates and confidence regions for sequential trials involving selection. J. Statist. Plann. Inference 135, 402–419]. This paper explores the problem in the simple setting in which two experimental treatments are compared in a single analysis. It is shown that in this case the estimate of Stallard and Todd is the maximum-likelihood estimate (m.l.e.), and this is compared with the estimate proposed by Shen. In particular, it is shown that the m.l.e. has infinite expectation whatever the true value of the mean being estimated. We show that there is no conditionally unbiased estimator, and propose a new family of approximately conditionally unbiased estimators, comparing these with the estimators suggested by Shen.
Resumo:
The evolutionary history of gains and losses of vegetative reproductive propagules (soredia) in Porpidia s.l., a group of lichen-forming ascomycetes, was clarified using Bayesian Markov chain Monte Carlo (MCMC) approaches to monophyly tests and a combined MCMC and maximum likelihood approach to ancestral character state reconstructions. The MCMC framework provided confidence estimates for the reconstructions of relationships and ancestral character states, which formed the basis for tests of evolutionary hypotheses. Monophyly tests rejected all hypotheses that predicted any clustering of reproductive modes in extant taxa. In addition, a nearest-neighbor statistic could not reject the hypothesis that the vegetative reproductive mode is randomly distributed throughout the group. These results show that transitions between presence and absence of the vegetative reproductive mode within Porpidia s.l. occurred several times and independently of each other. Likelihood reconstructions of ancestral character states at selected nodes suggest that - contrary to previous thought - the ancestor to Porpidia s.l. already possessed the vegetative reproductive mode. Furthermore, transition rates are reconstructed asymmetrically with the vegetative reproductive mode being gained at a much lower rate than it is lost. A cautious note has to be added, because a simulation study showed that the ancestral character state reconstructions were highly dependent on taxon sampling. However, our central conclusions, particularly the higher rate of change from vegetative reproductive mode present to absent than vice versa within Porpidia s.l., were found to be broadly independent of taxon sampling. [Ancestral character state reconstructions; Ascomycota, Bayesian inference; hypothesis testing; likelihood; MCMC; Porpidia; reproductive systems]
Resumo:
Objectives: To assess the potential source of variation that surgeon may add to patient outcome in a clinical trial of surgical procedures. Methods: Two large (n = 1380) parallel multicentre randomized surgical trials were undertaken to compare laparoscopically assisted hysterectomy with conventional methods of abdominal and vaginal hysterectomy; involving 43 surgeons. The primary end point of the trial was the occurrence of at least one major complication. Patients were nested within surgeons giving the data set a hierarchical structure. A total of 10% of patients had at least one major complication, that is, a sparse binary outcome variable. A linear mixed logistic regression model (with logit link function) was used to model the probability of a major complication, with surgeon fitted as a random effect. Models were fitted using the method of maximum likelihood in SAS((R)). Results: There were many convergence problems. These were resolved using a variety of approaches including; treating all effects as fixed for the initial model building; modelling the variance of a parameter on a logarithmic scale and centring of continuous covariates. The initial model building process indicated no significant 'type of operation' across surgeon interaction effect in either trial, the 'type of operation' term was highly significant in the abdominal trial, and the 'surgeon' term was not significant in either trial. Conclusions: The analysis did not find a surgeon effect but it is difficult to conclude that there was not a difference between surgeons. The statistical test may have lacked sufficient power, the variance estimates were small with large standard errors, indicating that the precision of the variance estimates may be questionable.
Resumo:
Microsatellites are widely used in genetic analyses, many of which require reliable estimates of microsatellite mutation rates, yet the factors determining mutation rates are uncertain. The most straightforward and conclusive method by which to study mutation is direct observation of allele transmissions in parent-child pairs, and studies of this type suggest a positive, possibly exponential, relationship between mutation rate and allele size, together with a bias toward length increase. Except for microsatellites on the Y chromosome, however, previous analyses have not made full use of available data and may have introduced bias: mutations have been identified only where child genotypes could not be generated by transmission from parents' genotypes, so that the probability that a mutation is detected depends on the distribution of allele lengths and varies with allele length. We introduce a likelihood-based approach that has two key advantages over existing methods. First, we can make formal comparisons between competing models of microsatellite evolution; second, we obtain asymptotically unbiased and efficient parameter estimates. Application to data composed of 118,866 parent-offspring transmissions of AC microsatellites supports the hypothesis that mutation rate increases exponentially with microsatellite length, with a suggestion that contractions become more likely than expansions as length increases. This would lead to a stationary distribution for allele length maintained by mutational balance. There is no evidence that contractions and expansions differ in their step size distributions.
Resumo:
We use microwave retrievals of upper tropospheric humidity (UTH) to estimate the impact of clear-sky-only sampling by infrared instruments on the distribution, variability and trends in UTH. Our method isolates the impact of the clear-sky-only sampling, without convolving errors from other sources. On daily time scales IR-sampled UTH contains large data gaps in convectively active areas, with only about 20-30 % of the tropics (30 S 30 N) being sampled. This results in a dry bias of about -9 %RH in the area-weighted tropical daily UTH time series. On monthly scales, maximum clear-sky bias (CSB) is up to -30 %RH over convectively active areas. The magnitude of CSB shows significant correlations with UTH itself (-0.5) and also with the variability in UTH (-0.6). We also show that IR-sampled UTH time series have higher interannual variability and smaller trends compared to microwave sampling. We argue that a significant part of the smaller trend results from the contrasting influence of diurnal drift in the satellite measurements on the wet and dry regions of the tropics.
Resumo:
Understanding the surface O3 response over a “receptor” region to emission changes over a foreign “source” region is key to evaluating the potential gains from an international approach to abate ozone (O3) pollution. We apply an ensemble of 21 global and hemispheric chemical transport models to estimate the spatial average surface O3 response over east Asia (EA), Europe (EU), North America (NA), and south Asia (SA) to 20% decreases in anthropogenic emissions of the O3 precursors, NOx, NMVOC, and CO (individually and combined), from each of these regions. We find that the ensemble mean surface O3 concentrations in the base case (year 2001) simulation matches available observations throughout the year over EU but overestimates them by >10 ppb during summer and early fall over the eastern United States and Japan. The sum of the O3 responses to NOx, CO, and NMVOC decreases separately is approximately equal to that from a simultaneous reduction of all precursors. We define a continental-scale “import sensitivity” as the ratio of the O3 response to the 20% reductions in foreign versus “domestic” (i.e., over the source region itself) emissions. For example, the combined reduction of emissions from the three foreign regions produces an ensemble spatial mean decrease of 0.6 ppb over EU (0.4 ppb from NA), less than the 0.8 ppb from the reduction of EU emissions, leading to an import sensitivity ratio of 0.7. The ensemble mean surface O3 response to foreign emissions is largest in spring and late fall (0.7–0.9 ppb decrease in all regions from the combined precursor reductions in the three foreign regions), with import sensitivities ranging from 0.5 to 1.1 (responses to domestic emission reductions are 0.8–1.6 ppb). High O3 values are much more sensitive to domestic emissions than to foreign emissions, as indicated by lower import sensitivities of 0.2 to 0.3 during July in EA, EU, and NA when O3 levels are typically highest and by the weaker relative response of annual incidences of daily maximum 8-h average O3 above 60 ppb to emission reductions in a foreign region (<10–20% of that to domestic) as compared to the annual mean response (up to 50% of that to domestic). Applying the ensemble annual mean results to changes in anthropogenic emissions from 1996 to 2002, we estimate a Northern Hemispheric increase in background surface O3 of about 0.1 ppb a−1, at the low end of the 0.1–0.5 ppb a−1 derived from observations. From an additional simulation in which global atmospheric methane was reduced, we infer that 20% reductions in anthropogenic methane emissions from a foreign source region would yield an O3 response in a receptor region that roughly equals that produced by combined 20% reductions of anthropogenic NOx, NMVOC, and CO emissions from the foreign source region.
Resumo:
Quantitative estimates of temperature and precipitation change during the late Pleistocene and Holocene have been difficult to obtain for much of the lowland Neotropics. Using two published lacustrine pollen records and a climate-vegetation model based on the modern abundance distributions of 154 Neotropical plant families, we demonstrate how family-level counts of fossil pollen can be used to quantitatively reconstruct tropical paleoclimate and provide needed information on historic patterns of climatic change. With this family-level analysis, we show that one area of the lowland tropics, northeastern Bolivia, experienced cooling (1–3 °C) and drying (400 mm/yr), relative to present, during the late Pleistocene (50,000–12,000 calendar years before present [cal. yr B.P.]). Immediately prior to the Last Glacial Maximum (LGM, ca. 21,000 cal. yr B.P.), we observe a distinct transition from cooler temperatures and variable precipitation to a period of warmer temperatures and relative dryness that extends to the middle Holocene (5000–3000 cal. yr B.P.). This prolonged reduction in precipitation occurs against the backdrop of increasing atmospheric CO2 concentrations, indicating that the presence of mixed savanna and dry-forest communities in northeastern Bolivia durng the LGM was not solely the result of low CO2 levels, as suggested previously, but also lower precipitation. The results of our analysis demonstrate the potential for using the distribution and abundance structure of modern Neotropical plant families to infer paleoclimate from the fossil pollen record.
Resumo:
The National Center for Atmospheric Research-Community Climate System Model (NCAR-CCSM) is used in a coupled atmosphere–ocean–sea-ice simulation of the Last Glacial Maximum (LGM, around 21,000 years ago) climate. In the tropics, the simulation shows a moderate cooling of 3 °C over land and 2 °C in the ocean in zonal average. This cooling is about 1 °C cooler than the CLIMAP sea surface temperatures (SSTs) but consistent with recent estimates of both land and sea surface temperature changes. Subtropical waters are cooled by 2–2.5 °C, also in agreement with recent estimates. The simulated oceanic thermohaline circulation at the LGM is not only shallower but also weaker than the modern with a migration of deep-water formation site in the North Atlantic as suggested by the paleoceanographic evidences. The simulated northward flow of Antarctic Bottom Water (AABW) is enhanced. These deep circulation changes are attributable to the increased surface density flux in the Southern Ocean caused by sea-ice expansion at the LGM. Both the Gulf Stream and the Kuroshio are intensified due to the overall increase of wind stress over the subtropical oceans. The intensified zonal wind stress and southward shift of its maximum in the Southern Ocean effectively enhances the transport of the Antarctic Circumpolar Current (ACC) by more than 50%. Simulated SSTs are lowered by up to 8 °C in the midlatitudes. Simulated conditions in the North Atlantic are warmer and with less sea-ice than indicated by CLIMAP again, in agreement with more recent estimates. The increased meridional SST gradient at the LGM results in an enhanced Hadley Circulation and increased midlatitude storm track precipitation. The increased baroclinic storm activity also intensifies the meridional atmospheric heat transport. A sensitivity experiment shows that about half of the simulated tropical cooling at the LGM originates from reduced atmospheric concentrations of greenhouse gases.