937 resultados para Bayesian smoothing
Resumo:
In this paper, we present various diagnostic methods for polyhazard models. Polyhazard models are a flexible family for fitting lifetime data. Their main advantage over the single hazard models, such as the Weibull and the log-logistic models, is to include a large amount of nonmonotone hazard shapes, as bathtub and multimodal curves. Some influence methods, such as the local influence and total local influence of an individual are derived, analyzed and discussed. A discussion of the computation of the likelihood displacement as well as the normal curvature in the local influence method are presented. Finally, an example with real data is given for illustration.
Resumo:
In a previous study, we observed no spatial genetic structure in Mexican populations of the parasitoids Chelonus insularis Cresson (Hymenoptera: Braconidae) and Campoletis sonorensis Cameron (Hymenoptera: Ichneumonidae) by using microsatellite markers In the current study, we Investigated whether for these important parasitoids of the fall armyworm (Lepidoptera: Noctuidae) there is any genetic structure at a larger scale Insects of both species were collected across the American continent and their phylogeography was Investigated using both nuclear and mitochondria] markers Our results suggest an ancient north-south migration of C insularis, whereas no clear pattern] could be determined for C sonorensis. Nonetheless, the resulting topology indicated the existence of a cryptic taxon within this later species. a few Canadian specimens determined as C. sonorensis branch outside a clack composed of the Argentinean Chelonus grioti Blanchard, the Brazilian Chelonus flavicincta Ashmead, and the rest of the C sonorensis individuals The individuals revealing the cryptic taxon were collected from Thichoplusia in (Hubner) (Lepidoptera. Noctuidae) on tomato (Lycopersicon spp) and may represent a biotype that has adapted to the early season phenology of its host. Overall, the loosely defined spatial genetic structure previously shown at a local fine scale also was found at the larger scale, for both species Dispersal of these insects may be partly driven by wind as suggested by genetic similarities between Individuals coming from very distant locations.
Resumo:
This article presents a statistical model of agricultural yield data based on a set of hierarchical Bayesian models that allows joint modeling of temporal and spatial autocorrelation. This method captures a comprehensive range of the various uncertainties involved in predicting crop insurance premium rates as opposed to the more traditional ad hoc, two-stage methods that are typically based on independent estimation and prediction. A panel data set of county-average yield data was analyzed for 290 counties in the State of Parana (Brazil) for the period of 1990 through 2002. Posterior predictive criteria are used to evaluate different model specifications. This article provides substantial improvements in the statistical and actuarial methods often applied to the calculation of insurance premium rates. These improvements are especially relevant to situations where data are limited.
Resumo:
A total of 152,145 weekly test-day milk yield records from 7317 first lactations of Holstein cows distributed in 93 herds in southeastern Brazil were analyzed. Test-day milk yields were classified into 44 weekly classes of DIM. The contemporary groups were defined as herd-year-week of test-day. The model included direct additive genetic, permanent environmental and residual effects as random and fixed effects of contemporary group and age of cow at calving as covariable, linear and quadratic effects. Mean trends were modeled by a cubic regression on orthogonal polynomials of DIM. Additive genetic and permanent environmental random effects were estimated by random regression on orthogonal Legendre polynomials. Residual variances were modeled using third to seventh-order variance functions or a step function with 1, 6,13,17 and 44 variance classes. Results from Akaike`s and Schwarz`s Bayesian information criterion suggested that a model considering a 7th-order Legendre polynomial for additive effect, a 12th-order polynomial for permanent environment effect and a step function with 6 classes for residual variances, fitted best. However, a parsimonious model, with a 6th-order Legendre polynomial for additive effects and a 7th-order polynomial for permanent environmental effects, yielded very similar genetic parameter estimates. (C) 2008 Elsevier B.V. All rights reserved.
Resumo:
Genetic recombination can produce heterogeneous phylogenetic histories within a set of homologous genes. Delineating recombination events is important in the study of molecular evolution, as inference of such events provides a clearer picture of the phylogenetic relationships among different gene sequences or genomes. Nevertheless, detecting recombination events can be a daunting task, as the performance of different recombination-detecting approaches can vary, depending on evolutionary events that take place after recombination. We recently evaluated the effects of post-recombination events on the prediction accuracy of recombination-detecting approaches using simulated nucleotide sequence data. The main conclusion, supported by other studies, is that one should not depend on a single method when searching for recombination events. In this paper, we introduce a two-phase strategy, applying three statistical measures to detect the occurrence of recombination events, and a Bayesian phylogenetic approach in delineating breakpoints of such events in nucleotide sequences. We evaluate the performance of these approaches using simulated data, and demonstrate the applicability of this strategy to empirical data. The two-phase strategy proves to be time-efficient when applied to large datasets, and yields high-confidence results.
Resumo:
The received view of an ad hoc hypothesis is that it accounts for only the observation(s) it was designed to account for, and so non-adhocness is generally held to be necessary or important for an introduced hypothesis or modification to a theory. Attempts by Popper and several others to convincingly explicate this view, however, prove to be unsuccessful or of doubtful value, and familiar and firmer criteria for evaluating the hypotheses or modified theories so classified are characteristically available. These points are obscured largely because the received view fails to adequately separate psychology from methodology or to recognise ambiguities in the use of 'ad hoc'.
Resumo:
The generalized Gibbs sampler (GGS) is a recently developed Markov chain Monte Carlo (MCMC) technique that enables Gibbs-like sampling of state spaces that lack a convenient representation in terms of a fixed coordinate system. This paper describes a new sampler, called the tree sampler, which uses the GGS to sample from a state space consisting of phylogenetic trees. The tree sampler is useful for a wide range of phylogenetic applications, including Bayesian, maximum likelihood, and maximum parsimony methods. A fast new algorithm to search for a maximum parsimony phylogeny is presented, using the tree sampler in the context of simulated annealing. The mathematics underlying the algorithm is explained and its time complexity is analyzed. The method is tested on two large data sets consisting of 123 sequences and 500 sequences, respectively. The new algorithm is shown to compare very favorably in terms of speed and accuracy to the program DNAPARS from the PHYLIP package.
Resumo:
Genetic recombination can produce heterogeneous phylogenetic histories within a set of homologous genes. Delineating recombination events is important in the study of molecular evolution, as inference of such events provides a clearer picture of the phylogenetic relationships among different gene sequences or genomes. Nevertheless, detecting recombination events can be a daunting task, as the performance of different recombination-detecting approaches can vary, depending on evolutionary events that take place after recombination. We previously evaluated the effects of post-recombination events on the prediction accuracy of recombination-detecting approaches using simulated nucleotide sequence data. The main conclusion, supported by other studies, is that one should not depend on a single method when searching for recombination events. In this paper, we introduce a two-phase strategy, applying three statistical measures to detect the occurrence of recombination events, and a Bayesian phylogenetic approach to delineate breakpoints of such events in nucleotide sequences. We evaluate the performance of these approaches using simulated data, and demonstrate the applicability of this strategy to empirical data. The two-phase strategy proves to be time-efficient when applied to large datasets, and yields high-confidence results.
Resumo:
Data mining is the process to identify valid, implicit, previously unknown, potentially useful and understandable information from large databases. It is an important step in the process of knowledge discovery in databases, (Olaru & Wehenkel, 1999). In a data mining process, input data can be structured, seme-structured, or unstructured. Data can be in text, categorical or numerical values. One of the important characteristics of data mining is its ability to deal data with large volume, distributed, time variant, noisy, and high dimensionality. A large number of data mining algorithms have been developed for different applications. For example, association rules mining can be useful for market basket problems, clustering algorithms can be used to discover trends in unsupervised learning problems, classification algorithms can be applied in decision-making problems, and sequential and time series mining algorithms can be used in predicting events, fault detection, and other supervised learning problems (Vapnik, 1999). Classification is among the most important tasks in the data mining, particularly for data mining applications into engineering fields. Together with regression, classification is mainly for predictive modelling. So far, there have been a number of classification algorithms in practice. According to (Sebastiani, 2002), the main classification algorithms can be categorized as: decision tree and rule based approach such as C4.5 (Quinlan, 1996); probability methods such as Bayesian classifier (Lewis, 1998); on-line methods such as Winnow (Littlestone, 1988) and CVFDT (Hulten 2001), neural networks methods (Rumelhart, Hinton & Wiliams, 1986); example-based methods such as k-nearest neighbors (Duda & Hart, 1973), and SVM (Cortes & Vapnik, 1995). Other important techniques for classification tasks include Associative Classification (Liu et al, 1998) and Ensemble Classification (Tumer, 1996).
Resumo:
A significant problem in the collection of responses to potentially sensitive questions, such as relating to illegal, immoral or embarrassing activities, is non-sampling error due to refusal to respond or false responses. Eichhorn & Hayre (1983) suggested the use of scrambled responses to reduce this form of bias. This paper considers a linear regression model in which the dependent variable is unobserved but for which the sum or product with a scrambling random variable of known distribution, is known. The performance of two likelihood-based estimators is investigated, namely of a Bayesian estimator achieved through a Markov chain Monte Carlo (MCMC) sampling scheme, and a classical maximum-likelihood estimator. These two estimators and an estimator suggested by Singh, Joarder & King (1996) are compared. Monte Carlo results show that the Bayesian estimator outperforms the classical estimators in almost all cases, and the relative performance of the Bayesian estimator improves as the responses become more scrambled.
Resumo:
Item noise models of recognition assert that interference at retrieval is generated by the words from the study list. Context noise models of recognition assert that interference at retrieval is generated by the contexts in which the test word has appeared. The authors introduce the bind cue decide model of episodic memory, a Bayesian context noise model, and demonstrate how it can account for data from the item noise and dual-processing approaches to recognition memory. From the item noise perspective, list strength and list length effects, the mirror effect for word frequency and concreteness, and the effects of the similarity of other words in a list are considered. From the dual-processing perspective, process dissociation data on the effects of length, temporal separation of lists, strength, and diagnosticity of context are examined. The authors conclude that the context noise approach to recognition is a viable alternative to existing approaches.
Gender differences in the relationship between depression and suicidal ideation in young adolescents
Resumo:
Objective: This study examined the risk relationship between depressive symptomatology and suicidal ideation for young adolescent males and females. Method: A large cohort of students in their first year of high school completed the Center for Epidemiological Studies Depression Scale (CES-D) and the Adolescent Suicide Questionnaire. The risk relationship between depressive symptomatology and suicidal ideation was modelled using non-parametric kernel-smoothing techniques. Results: Suicidal ideation was more frequently reported by females compared with males which was partly explained by females having higher mean depression scores. At moderate levels of depression females also had a significantly higher risk of suicidal ideation compared with males and this increased risk contributed to the overall higher levels of female ideation. Conclusions: The risk relationship between depressive symptomatology and suicidal ideation is different for young adolescent males and females. The results indicate that moderate levels of depressive symptomatology can be associated with suicidal ideation (especially among young females) and that for these young people a suicide risk assessment is required.
Resumo:
Intelligent design theorist William Dembski has proposed an explanatory filter for distinguishing between events due to chance, lawful regularity or design. We show that if Dembski's filter were adopted as a scientific heuristic, some classical developments in science would not be rational, and that Dembski's assertion that the filter reliably identifies rarefied design requires ignoring the state of background knowledge. If background information changes even slightly, the filter's conclusion will vary wildly. Dembski fails to overcome Hume's objections to arguments from design.
Resumo:
The phylogeny of the Australian legume genus Daviesia was estimated using sequences of the internal transcribed spacers of nuclear ribosomal DNA. Partial congruence was found with previous analyses using morphology, including strong support for monophyly of the genus and for a sister group relationship between the clade D. pachyloma and the rest of the genus. A previously unplaced bird-pollinated species, anceps + D. D. epiphyllum, was well supported as sister to the only other bird-pollinated species in the genus, D. speciosa, indicating a single origin of bird pollination in their common ancestor. Other morphological groups within Daviesia were not supported and require reassessment. A strong and previously unreported sister clade of Daviesia consists of the two monotypic genera Erichsenia and Viminaria. These share phyllode-like leaves and indehiscent fruits. The evolutionary history of cord roots, which have anomalous secondary thickening, was explored using parsimony. Cord roots are limited to three separate clades but have a complex history involving a small number of gains (most likely 0-3) and losses (0-5). The anomalous structure of cord roots ( adventitious vascular strands embedded in a parenchymatous matrix) may facilitate nutrient storage, and the roots may be contractile. Both functions may be related to a postfire resprouting adaptation. Alternatively, cord roots may be an adaptation to the low-nutrient lateritic soils of Western Australia. However, tests for association between root type, soil type, and growth habit were equivocal, depending on whether the variables were treated as phylogenetically dependent (insignificant) or independent ( significant).
Resumo:
We model and calibrate the arguments in favor and against short-term and long-term debt. These arguments broadly include: maturity premium, sustainability, and service smoothing. We use a dynamic-equilibrium model with tax distortions and government outlays uncertainty, and model maturity as the fraction of debt that needs to be rolled over every period. In the model, the benefits of defaulting are tempered by higher future interest rates. We then calibrate our artificial economy and solve for the optimal debt maturity for Brazil as an example of a developing country and the US as an example of a mature economy. We obtain that the calibrated costs from defaulting on long-term debt more than offset costs associated with short-term debt. Therefore, short-term debt implies higher welfare levels.