941 resultados para Bayesian rationality
Resumo:
In this paper, we present various diagnostic methods for polyhazard models. Polyhazard models are a flexible family for fitting lifetime data. Their main advantage over the single hazard models, such as the Weibull and the log-logistic models, is to include a large amount of nonmonotone hazard shapes, as bathtub and multimodal curves. Some influence methods, such as the local influence and total local influence of an individual are derived, analyzed and discussed. A discussion of the computation of the likelihood displacement as well as the normal curvature in the local influence method are presented. Finally, an example with real data is given for illustration.
Resumo:
In a previous study, we observed no spatial genetic structure in Mexican populations of the parasitoids Chelonus insularis Cresson (Hymenoptera: Braconidae) and Campoletis sonorensis Cameron (Hymenoptera: Ichneumonidae) by using microsatellite markers In the current study, we Investigated whether for these important parasitoids of the fall armyworm (Lepidoptera: Noctuidae) there is any genetic structure at a larger scale Insects of both species were collected across the American continent and their phylogeography was Investigated using both nuclear and mitochondria] markers Our results suggest an ancient north-south migration of C insularis, whereas no clear pattern] could be determined for C sonorensis. Nonetheless, the resulting topology indicated the existence of a cryptic taxon within this later species. a few Canadian specimens determined as C. sonorensis branch outside a clack composed of the Argentinean Chelonus grioti Blanchard, the Brazilian Chelonus flavicincta Ashmead, and the rest of the C sonorensis individuals The individuals revealing the cryptic taxon were collected from Thichoplusia in (Hubner) (Lepidoptera. Noctuidae) on tomato (Lycopersicon spp) and may represent a biotype that has adapted to the early season phenology of its host. Overall, the loosely defined spatial genetic structure previously shown at a local fine scale also was found at the larger scale, for both species Dispersal of these insects may be partly driven by wind as suggested by genetic similarities between Individuals coming from very distant locations.
Resumo:
This article presents a statistical model of agricultural yield data based on a set of hierarchical Bayesian models that allows joint modeling of temporal and spatial autocorrelation. This method captures a comprehensive range of the various uncertainties involved in predicting crop insurance premium rates as opposed to the more traditional ad hoc, two-stage methods that are typically based on independent estimation and prediction. A panel data set of county-average yield data was analyzed for 290 counties in the State of Parana (Brazil) for the period of 1990 through 2002. Posterior predictive criteria are used to evaluate different model specifications. This article provides substantial improvements in the statistical and actuarial methods often applied to the calculation of insurance premium rates. These improvements are especially relevant to situations where data are limited.
Resumo:
A total of 152,145 weekly test-day milk yield records from 7317 first lactations of Holstein cows distributed in 93 herds in southeastern Brazil were analyzed. Test-day milk yields were classified into 44 weekly classes of DIM. The contemporary groups were defined as herd-year-week of test-day. The model included direct additive genetic, permanent environmental and residual effects as random and fixed effects of contemporary group and age of cow at calving as covariable, linear and quadratic effects. Mean trends were modeled by a cubic regression on orthogonal polynomials of DIM. Additive genetic and permanent environmental random effects were estimated by random regression on orthogonal Legendre polynomials. Residual variances were modeled using third to seventh-order variance functions or a step function with 1, 6,13,17 and 44 variance classes. Results from Akaike`s and Schwarz`s Bayesian information criterion suggested that a model considering a 7th-order Legendre polynomial for additive effect, a 12th-order polynomial for permanent environment effect and a step function with 6 classes for residual variances, fitted best. However, a parsimonious model, with a 6th-order Legendre polynomial for additive effects and a 7th-order polynomial for permanent environmental effects, yielded very similar genetic parameter estimates. (C) 2008 Elsevier B.V. All rights reserved.
Resumo:
In his last papers about deontic logic, von Wright sustained that there is no genuine logic of norms. We argue in this paper that this striking statement by the father of deontic logic should not be understood as a death sentence to the subject. Rather, it indicates a profound change in von Wright`s understanding about the epistemic and ontological role of logic in the field of norms. Instead of a logical constructivism of deontic systems revealing a necessary structure of prescriptive discourse, which marked his earlier efforts, he adopted the view that such systems should be seem as mere objects of comparison, i.e. as providing practical standards of rationality for norm-giving activity. Within such view he proposed an interpretation of standard deontic logic in such a way to free deontic logicians from the philosophical difficulties related to the so-called Jorgensen`s dilemma and deontic paradoxes. This effort, as we claim in the present paper, is an application of Wittgenstein`s therapeutic method to dissolve philosophical difficulties caused by the use of logical tools to model relations between norms.
Resumo:
Genetic recombination can produce heterogeneous phylogenetic histories within a set of homologous genes. Delineating recombination events is important in the study of molecular evolution, as inference of such events provides a clearer picture of the phylogenetic relationships among different gene sequences or genomes. Nevertheless, detecting recombination events can be a daunting task, as the performance of different recombination-detecting approaches can vary, depending on evolutionary events that take place after recombination. We recently evaluated the effects of post-recombination events on the prediction accuracy of recombination-detecting approaches using simulated nucleotide sequence data. The main conclusion, supported by other studies, is that one should not depend on a single method when searching for recombination events. In this paper, we introduce a two-phase strategy, applying three statistical measures to detect the occurrence of recombination events, and a Bayesian phylogenetic approach in delineating breakpoints of such events in nucleotide sequences. We evaluate the performance of these approaches using simulated data, and demonstrate the applicability of this strategy to empirical data. The two-phase strategy proves to be time-efficient when applied to large datasets, and yields high-confidence results.
Resumo:
This opening chapter provides an overview of organizational behavior theory and research and the paradigms that have dominated the field to date. Running through a discussion of rational notions of organizational behavior, to concepts of bounded rationality and most recently the call for bounded emotionality perspectives, we identify for the reader what a bounded emotionality perspective adds to the understanding of organizations. We then provide an overview of the remaining chapters in the book and how they contribute to the book's objectives.
Resumo:
The received view of an ad hoc hypothesis is that it accounts for only the observation(s) it was designed to account for, and so non-adhocness is generally held to be necessary or important for an introduced hypothesis or modification to a theory. Attempts by Popper and several others to convincingly explicate this view, however, prove to be unsuccessful or of doubtful value, and familiar and firmer criteria for evaluating the hypotheses or modified theories so classified are characteristically available. These points are obscured largely because the received view fails to adequately separate psychology from methodology or to recognise ambiguities in the use of 'ad hoc'.
Resumo:
The discovery of Neptune in September 1846 is a good example of scientific rationality but it has proven to be surprisingly difficult to explain how this is so from the perspective of Karl Popper, Thomas Kuhn and several other commentators who have been influenced by these thinkers. I try briefly to explain how to avoid these difficulties and understand the achievement of the astronomers who predicted the location of the new planet, Urbain J. J. Leverrier and John Couch Adams.
Resumo:
Knowledge of residual perturbations in the orbit of Uranus in the early 1840s did not lead to the refutation of Newton's law of gravitation but instead to the discovery of Neptune in 1846. Karl Popper asserts that this case is atypical of science and that the law of gravitation was at least prima facie falsified by these perturbations. I argue that these assertions are the product of a false, a priori methodological position I call, 'Weak Popperian Falsificationism' (WPF). Further, on the evidence the law was not prima facie false and was not generally considered so by astronomers at the time. Many of Popper's commentators (Kuhn, Lakatos, Feyerabend and others) presuppose WPF and their views on this case and its implications for scientific rationality and method suffer from this same defect.
Resumo:
The generalized Gibbs sampler (GGS) is a recently developed Markov chain Monte Carlo (MCMC) technique that enables Gibbs-like sampling of state spaces that lack a convenient representation in terms of a fixed coordinate system. This paper describes a new sampler, called the tree sampler, which uses the GGS to sample from a state space consisting of phylogenetic trees. The tree sampler is useful for a wide range of phylogenetic applications, including Bayesian, maximum likelihood, and maximum parsimony methods. A fast new algorithm to search for a maximum parsimony phylogeny is presented, using the tree sampler in the context of simulated annealing. The mathematics underlying the algorithm is explained and its time complexity is analyzed. The method is tested on two large data sets consisting of 123 sequences and 500 sequences, respectively. The new algorithm is shown to compare very favorably in terms of speed and accuracy to the program DNAPARS from the PHYLIP package.
Resumo:
Genetic recombination can produce heterogeneous phylogenetic histories within a set of homologous genes. Delineating recombination events is important in the study of molecular evolution, as inference of such events provides a clearer picture of the phylogenetic relationships among different gene sequences or genomes. Nevertheless, detecting recombination events can be a daunting task, as the performance of different recombination-detecting approaches can vary, depending on evolutionary events that take place after recombination. We previously evaluated the effects of post-recombination events on the prediction accuracy of recombination-detecting approaches using simulated nucleotide sequence data. The main conclusion, supported by other studies, is that one should not depend on a single method when searching for recombination events. In this paper, we introduce a two-phase strategy, applying three statistical measures to detect the occurrence of recombination events, and a Bayesian phylogenetic approach to delineate breakpoints of such events in nucleotide sequences. We evaluate the performance of these approaches using simulated data, and demonstrate the applicability of this strategy to empirical data. The two-phase strategy proves to be time-efficient when applied to large datasets, and yields high-confidence results.
Resumo:
Data mining is the process to identify valid, implicit, previously unknown, potentially useful and understandable information from large databases. It is an important step in the process of knowledge discovery in databases, (Olaru & Wehenkel, 1999). In a data mining process, input data can be structured, seme-structured, or unstructured. Data can be in text, categorical or numerical values. One of the important characteristics of data mining is its ability to deal data with large volume, distributed, time variant, noisy, and high dimensionality. A large number of data mining algorithms have been developed for different applications. For example, association rules mining can be useful for market basket problems, clustering algorithms can be used to discover trends in unsupervised learning problems, classification algorithms can be applied in decision-making problems, and sequential and time series mining algorithms can be used in predicting events, fault detection, and other supervised learning problems (Vapnik, 1999). Classification is among the most important tasks in the data mining, particularly for data mining applications into engineering fields. Together with regression, classification is mainly for predictive modelling. So far, there have been a number of classification algorithms in practice. According to (Sebastiani, 2002), the main classification algorithms can be categorized as: decision tree and rule based approach such as C4.5 (Quinlan, 1996); probability methods such as Bayesian classifier (Lewis, 1998); on-line methods such as Winnow (Littlestone, 1988) and CVFDT (Hulten 2001), neural networks methods (Rumelhart, Hinton & Wiliams, 1986); example-based methods such as k-nearest neighbors (Duda & Hart, 1973), and SVM (Cortes & Vapnik, 1995). Other important techniques for classification tasks include Associative Classification (Liu et al, 1998) and Ensemble Classification (Tumer, 1996).
Resumo:
A significant problem in the collection of responses to potentially sensitive questions, such as relating to illegal, immoral or embarrassing activities, is non-sampling error due to refusal to respond or false responses. Eichhorn & Hayre (1983) suggested the use of scrambled responses to reduce this form of bias. This paper considers a linear regression model in which the dependent variable is unobserved but for which the sum or product with a scrambling random variable of known distribution, is known. The performance of two likelihood-based estimators is investigated, namely of a Bayesian estimator achieved through a Markov chain Monte Carlo (MCMC) sampling scheme, and a classical maximum-likelihood estimator. These two estimators and an estimator suggested by Singh, Joarder & King (1996) are compared. Monte Carlo results show that the Bayesian estimator outperforms the classical estimators in almost all cases, and the relative performance of the Bayesian estimator improves as the responses become more scrambled.
Resumo:
The British agricultural sector is either already in or rapidly approaching some sort of crisis. Two features are particularly significant in the political response to the current situation. First, there is an increasingly neoliberal approach to agricultural policy. Sec end, agricultural policy per se is being subsumed with wider rural policies. In this paper we question the rationality of both these trends, both theoretically through 'new wave regulation theory' and by relating the British situation to the recent experiences of the agricultural sectors in Australia and New Zealand.