976 resultados para Bayesian models
Resumo:
This article presents a statistical model of agricultural yield data based on a set of hierarchical Bayesian models that allows joint modeling of temporal and spatial autocorrelation. This method captures a comprehensive range of the various uncertainties involved in predicting crop insurance premium rates as opposed to the more traditional ad hoc, two-stage methods that are typically based on independent estimation and prediction. A panel data set of county-average yield data was analyzed for 290 counties in the State of Parana (Brazil) for the period of 1990 through 2002. Posterior predictive criteria are used to evaluate different model specifications. This article provides substantial improvements in the statistical and actuarial methods often applied to the calculation of insurance premium rates. These improvements are especially relevant to situations where data are limited.
Resumo:
Els rius i rieres mediterranis són ecosistemes que es caracteritzen per fortes oscil•lacions de cabal i temperatura al llarg de l’any. Aquestes oscil•lacions provoquen canvis ambientals en l'hàbitat i en els recursos que afecten directament o indirecta la biota que habita aquests ecosistemes, la qual, per tant, ha de presentar adaptacions a aquestes oscil•lacions ambientals. L'escenari actual de canvi climàtic preveu una intensificació dels fenòmens de sequera i augment de temperatura. Entendre com la biota dels rius respon a aquestes fluctuacions és de gran importància per poder anticipar les respostes d'aquests sistemes als imminents canvis ambientals així com per gestionar adequadament els recursos hídrics en un futur. Els objectius principals d'aquesta tesi eren: caracteritzar estructural i funcionalment dues rieres intermitents mediterrànies al llarg dels diferents períodes característics del cicle anual i veure els efectes d'un augment de la sequera; veure com aquests efectes podien afectar l'ecosistema ripari circumdant i establir com diferències en la qualitat de la matèria orgànica derivades del canvi climàtic pot afectar el fitness i desenvolupament dels invertebrats. Aquests objectius s'han pogut complir només parcialment, ja que adversitats climàtiques van impedir finalitzar amb èxit la manipulació del cabal al camp i la resolució d'algunes dades no ha estat prou bona com per aplicar els models corresponents. Aquests contratemps s'han solucionat amb la incorporació de dos nous experiments (un encara s'ha de realitzar), fet que ha fet enlentir la finalització de la tesi.
Resumo:
Attrition in longitudinal studies can lead to biased results. The study is motivated by the unexpected observation that alcohol consumption decreased despite increased availability, which may be due to sample attrition of heavy drinkers. Several imputation methods have been proposed, but rarely compared in longitudinal studies of alcohol consumption. The imputation of consumption level measurements is computationally particularly challenging due to alcohol consumption being a semi-continuous variable (dichotomous drinking status and continuous volume among drinkers), and the non-normality of data in the continuous part. Data come from a longitudinal study in Denmark with four waves (2003-2006) and 1771 individuals at baseline. Five techniques for missing data are compared: Last value carried forward (LVCF) was used as a single, and Hotdeck, Heckman modelling, multivariate imputation by chained equations (MICE), and a Bayesian approach as multiple imputation methods. Predictive mean matching was used to account for non-normality, where instead of imputing regression estimates, "real" observed values from similar cases are imputed. Methods were also compared by means of a simulated dataset. The simulation showed that the Bayesian approach yielded the most unbiased estimates for imputation. The finding of no increase in consumption levels despite a higher availability remained unaltered. Copyright (C) 2011 John Wiley & Sons, Ltd.
Resumo:
��In a sign that researchers are grappling with therapy development, the 4th annual conference on Clinical Trials in Alzheimer's Disease was filled beyond its venue's capacity, drawing 522 researchers from around the globe. Held 3-5 November 2011 in San Diego, CTAD is the brainchild of Paul Aisen, Jacques Touchon, Bruno Vellas, and Michael Weiner. The conference posted no ringing trial successes. Instead, scientists worked on methodological aspects they hope will improve future trials' chances. They discussed Bayesian models, simulated placebos, and biomarker data standards. They presented alternative outcome measures to the ADAS-cog, ranging widely from composite scales that are sensitive early on to continuous measures that encompass a patients' day-to-day variability. They focused on EEG, and on a collective effort to develop patient-reported outcomes. Highlights include:Whence and Where To: History and Future of AD Therapy Trials��Webinar: Evolution of AD Trials��Nutrient Formulation Appears to Grease Memory Function��Door Slams on RAGE��Clinical Trials: Making "Protocols From Hell" Less Burdensome��EEG: Coming in From the Margins of Alzheimer's Research?��EEG: Old Method to Lend New Help in AD Drug Development?������
Resumo:
Background: Most mortality atlases show static maps from count data aggregated over time. This procedure has several methodological problems and serious limitations for decision making in Public Health. The evaluation of health outcomes, including mortality, should be approached from a dynamic time perspective that is specific for each gender and age group. At the moment, researches in Spain do not provide a dynamic image of the population’s mortality status from a spatio-temporal point of view. The aim of this paper is to describe the spatial distribution of mortality from all causes in small areas of Andalusia (Southern Spain) and evolution over time from 1981 to 2006. Methods: A small-area ecological study was devised using the municipality as the unit for analysis. Two spatiotemporal hierarchical Bayesian models were estimated for each age group and gender. One of these was used to estimate the specific mortality rate, together with its time trends, and the other to estimate the specific rate ratio for each municipality compared with Spain as a whole. Results: More than 97% of the municipalities showed a diminishing or flat mortality trend in all gender and age groups. In 2006, over 95% of municipalities showed male and female mortality specific rates similar or significantly lower than Spanish rates for all age groups below 65. Systematically, municipalities in Western Andalusia showed significant male and female mortality excess from 1981 to 2006 only in age groups over 65. Conclusions: The study shows a dynamic geographical distribution of mortality, with a different pattern for each year, gender and age group. This information will contribute towards a reflection on the past, present and future of mortality in Andalusia.
Resumo:
Objective: To study the linkage between material deprivation and mortality from all causes, for men and women separately, in the capital cities of the provinces in Andalusia and Catalonia (Spain). Methods: A small-area ecological study was devised using the census section as the unit for analysis. 188 983 Deaths occurring in the capital cities of the Andalusian provinces and 109 478 deaths recorded in the Catalan capital cities were examined. Principal components factorial analysis was used to devise a material deprivation index comprising the percentage of manual labourers, unemployment and illiteracy. A hierarchical Bayesian model was used to study the relationship between mortality and area deprivation. Main results: In most cities, results show an increased male mortality risk in the most deprived areas in relation to the least depressed. In Andalusia, the relative risks between the highest and lowest deprivation decile ranged from 1.24 (Malaga) to 1.40 (Granada), with 95% credibility intervals showing a significant excess risk. In Catalonia, relative risks ranged between 1.08 (Girona) and 1.50 (Tarragona). No evidence was found for an excess of female mortality in most deprived areas in either of the autonomous communities. Conclusions: Within cities, gender-related differences were revealed when deprivation was correlated geographically with mortality rates. These differences were found from an ecological perspective. Further research is needed in order to validate these results from an individual approach. The idea to be analysed is to identify those factors that explain these differences at an individual level.
Resumo:
Until now, mortality atlases have been static. Most of them describe the geographical distribution of mortality using count data aggregated over time and standardized mortality rates. However, this methodology has several limitations. Count data aggregated over time produce a bias in the estimation of death rates. Moreover, this practice difficult the study of temporal changes in geographical distribution of mortality. On the other hand, using standardized mortality hamper to check differences in mortality among groups. The Interactive Mortality Atlas in Andalusia (AIMA) is an alternative to conventional static atlases. It is a dynamic Geographical Information System that allows visualizing in web-site more than 12.000 maps and 338.00 graphics related to the spatio-temporal distribution of the main death causes in Andalusia by age and sex groups from 1981. The objective of this paper is to describe the methods used for AIMA development, to show technical specifications and to present their interactivity. The system is available from the link products in www.demap.es. AIMA is the first interactive GIS that have been developed in Spain with these characteristics. Spatio-temporal Hierarchical Bayesian Models were used for statistical data analysis. The results were integrated into web-site using a PHP environment and a dynamic cartography in Flash. Thematic maps in AIMA demonstrate that the geographical distribution of mortality is dynamic, with differences among year, age and sex groups. The information nowadays provided by AIMA and the future updating will contribute to reflect on the past, the present and the future of population health in Andalusia.
Resumo:
Les sociétés modernes dépendent de plus en plus sur les systèmes informatiques et ainsi, il y a de plus en plus de pression sur les équipes de développement pour produire des logiciels de bonne qualité. Plusieurs compagnies utilisent des modèles de qualité, des suites de programmes qui analysent et évaluent la qualité d'autres programmes, mais la construction de modèles de qualité est difficile parce qu'il existe plusieurs questions qui n'ont pas été répondues dans la littérature. Nous avons étudié les pratiques de modélisation de la qualité auprès d'une grande entreprise et avons identifié les trois dimensions où une recherche additionnelle est désirable : Le support de la subjectivité de la qualité, les techniques pour faire le suivi de la qualité lors de l'évolution des logiciels, et la composition de la qualité entre différents niveaux d'abstraction. Concernant la subjectivité, nous avons proposé l'utilisation de modèles bayésiens parce qu'ils sont capables de traiter des données ambiguës. Nous avons appliqué nos modèles au problème de la détection des défauts de conception. Dans une étude de deux logiciels libres, nous avons trouvé que notre approche est supérieure aux techniques décrites dans l'état de l'art, qui sont basées sur des règles. Pour supporter l'évolution des logiciels, nous avons considéré que les scores produits par un modèle de qualité sont des signaux qui peuvent être analysés en utilisant des techniques d'exploration de données pour identifier des patrons d'évolution de la qualité. Nous avons étudié comment les défauts de conception apparaissent et disparaissent des logiciels. Un logiciel est typiquement conçu comme une hiérarchie de composants, mais les modèles de qualité ne tiennent pas compte de cette organisation. Dans la dernière partie de la dissertation, nous présentons un modèle de qualité à deux niveaux. Ces modèles ont trois parties: un modèle au niveau du composant, un modèle qui évalue l'importance de chacun des composants, et un autre qui évalue la qualité d'un composé en combinant la qualité de ses composants. L'approche a été testée sur la prédiction de classes à fort changement à partir de la qualité des méthodes. Nous avons trouvé que nos modèles à deux niveaux permettent une meilleure identification des classes à fort changement. Pour terminer, nous avons appliqué nos modèles à deux niveaux pour l'évaluation de la navigabilité des sites web à partir de la qualité des pages. Nos modèles étaient capables de distinguer entre des sites de très bonne qualité et des sites choisis aléatoirement. Au cours de la dissertation, nous présentons non seulement des problèmes théoriques et leurs solutions, mais nous avons également mené des expériences pour démontrer les avantages et les limitations de nos solutions. Nos résultats indiquent qu'on peut espérer améliorer l'état de l'art dans les trois dimensions présentées. En particulier, notre travail sur la composition de la qualité et la modélisation de l'importance est le premier à cibler ce problème. Nous croyons que nos modèles à deux niveaux sont un point de départ intéressant pour des travaux de recherche plus approfondis.
Resumo:
The eruption of the volcano at Thera (Santorini) in the Aegean Sea undoubtedly had a profound influence on the civilizations of the surrounding region. The date of the eruption has been a subject of much controversy because it must be linked into the established and intricate archaeological phasings of both the prehistoric Aegean and the wider east Mediterranean. Radiocarbon dating of material from the volcanic destruction layer itself can provide some evidence for the date of the eruption, but because of the shape of the calibration curve for the relevant period, the value of such dates relies on there being no biases in the data sets. However, by dating the material from phases earlier and later than the eruption, some of the problems of the calibration data set can be circumvented and the chronology for the region can be resolved with more certainty. In this paper, we draw together the evidence we have accumulated so far, including new data on the destruction layer itself and for the preceding cultural horizon at Thera, and from associated layers at Miletos in western Turkey. Using Bayesian models to synthesize the data and to identify outliers, we conclude from the most reliable C-14 evidence (and using the INTCAL98 calibration data set) that the eruption of Thera occurred between 1663 and 1599 BC.
Resumo:
Undeniably, anticipation plays a crucial role in cognition. By what means, to what extent, and what it achieves remain open questions. In a recent BBS target article, Clark (in press) depicts an integrative model of the brain that builds on hierarchical Bayesian models of neural processing (Rao and Ballard, 1999; Friston, 2005; Brown et al., 2011), and their most recent formulation using the free-energy principle borrowed from thermodynamics (Feldman and Friston, 2010; Friston, 2010; Friston et al., 2010). Hierarchical generative models of cognition, such as those described by Clark, presuppose the manipulation of representations and internal models of the world, in as much detail as is perceptually available. Perhaps surprisingly, Clark acknowledges the existence of a “virtual version of the sensory data” (p. 4), but with no reference to some of the historical debates that shaped cognitive science, related to the storage, manipulation, and retrieval of representations in a cognitive system (Shanahan, 1997), or accounting for the emergence of intentionality within such a system (Searle, 1980; Preston and Bishop, 2002). Instead of demonstrating how this Bayesian framework responds to these foundational questions, Clark describes the structure and the functional properties of an action-oriented, multi-level system that is meant to combine perception, learning, and experience (Niedenthal, 2007).
Resumo:
The Delaware River provides half of New York City's drinking water, is a habitat for wild trout, American shad and the federally endangered dwarf wedge mussel. It has suffered four 100‐year floods in the last seven years. A drought during the 1960s stands as a warning of the potential vulnerability of the New York City area to severe water shortages if a similar drought were to recur. The water releases from three New York City dams on the Delaware River's headwaters impact not only the reliability of the city’s water supply, but also the potential impact of floods, and the quality of the aquatic habitat in the upper river. The goal of this work is to influence the Delaware River water release policies (FFMP/OST) to further benefit river habitat and fisheries without increasing New York City's drought risk, or the flood risk to down basin residents. The Delaware water release policies are constrained by the dictates of two US Supreme Court Decrees (1931 and 1954) and the need for unanimity among four states: New York, New Jersey, Pennsylvania, and Delaware ‐‐ and New York City. Coordination of their activities and the operation under the existing decrees is provided by the Delaware River Basin Commission (DRBC). Questions such as the probability of the system approaching drought state based on the current FFMP plan and the severity of the 1960s drought are addressed using long record paleo‐reconstructions of flows. For this study, we developed reconstructed total annual flows (water year) for 3 reservoir inflows using regional tree rings going back upto 1754 (a total of 246 years). The reconstructed flows are used with a simple reservoir model to quantify droughts. We observe that the 1960s drought is by far the worst drought based on 246 years of simulations (since 1754).
Resumo:
The portfolio theory is a field of study devoted to investigate the decision-making by investors of resources. The purpose of this process is to reduce risk through diversification and thus guarantee a return. Nevertheless, the classical Mean-Variance has been criticized regarding its parameters and it is observed that the use of variance and covariance has sensitivity to the market and parameter estimation. In order to reduce the estimation errors, the Bayesian models have more flexibility in modeling, capable of insert quantitative and qualitative parameters about the behavior of the market as a way of reducing errors. Observing this, the present study aimed to formulate a new matrix model using Bayesian inference as a way to replace the covariance in the MV model, called MCB - Covariance Bayesian model. To evaluate the model, some hypotheses were analyzed using the method ex post facto and sensitivity analysis. The benchmarks used as reference were: (1) the classical Mean Variance, (2) the Bovespa index's market, and (3) in addition 94 investment funds. The returns earned during the period May 2002 to December 2009 demonstrated the superiority of MCB in relation to the classical model MV and the Bovespa Index, but taking a little more diversifiable risk that the MV. The robust analysis of the model, considering the time horizon, found returns near the Bovespa index, taking less risk than the market. Finally, in relation to the index of Mao, the model showed satisfactory, return and risk, especially in longer maturities. Some considerations were made, as well as suggestions for further work
Resumo:
In epidemiological work, outcomes are frequently non-normal, sample sizes may be large, and effects are often small. To relate health outcomes to geographic risk factors, fast and powerful methods for fitting spatial models, particularly for non-normal data, are required. We focus on binary outcomes, with the risk surface a smooth function of space. We compare penalized likelihood models, including the penalized quasi-likelihood (PQL) approach, and Bayesian models based on fit, speed, and ease of implementation. A Bayesian model using a spectral basis representation of the spatial surface provides the best tradeoff of sensitivity and specificity in simulations, detecting real spatial features while limiting overfitting and being more efficient computationally than other Bayesian approaches. One of the contributions of this work is further development of this underused representation. The spectral basis model outperforms the penalized likelihood methods, which are prone to overfitting, but is slower to fit and not as easily implemented. Conclusions based on a real dataset of cancer cases in Taiwan are similar albeit less conclusive with respect to comparing the approaches. The success of the spectral basis with binary data and similar results with count data suggest that it may be generally useful in spatial models and more complicated hierarchical models.
Resumo:
Acknowledgements This study was part of the Tursiops Project of the Dolphin Research Centre of Caprera, La Maddalena. Financial and logistical support was provided by the Centro Turistico Studentesco (CTS) and by the National Park of the Archipelago de La Maddalena. We thank the Natural Reserve of Bocche di Bonifacio for the support provided during data collection. The authors thank the numerous volunteers of the Caprera Dolphin Research Centre and especially Marco Ferraro, Mirko Ugo, Angela Pira and Maurizio Piras whose assistance during field observation and skills as a boat driver were invaluable.
Resumo:
Gene clustering is a useful exploratory technique to group together genes with similar expression levels under distinct cell cycle phases or distinct conditions. It helps the biologist to identify potentially meaningful relationships between genes. In this study, we propose a clustering method based on multivariate normal mixture models, where the number of clusters is predicted via sequential hypothesis tests: at each step, the method considers a mixture model of m components (m = 2 in the first step) and tests if in fact it should be m - 1. If the hypothesis is rejected, m is increased and a new test is carried out. The method continues (increasing m) until the hypothesis is accepted. The theoretical core of the method is the full Bayesian significance test, an intuitive Bayesian approach, which needs no model complexity penalization nor positive probabilities for sharp hypotheses. Numerical experiments were based on a cDNA microarray dataset consisting of expression levels of 205 genes belonging to four functional categories, for 10 distinct strains of Saccharomyces cerevisiae. To analyze the method's sensitivity to data dimension, we performed principal components analysis on the original dataset and predicted the number of classes using 2 to 10 principal components. Compared to Mclust (model-based clustering), our method shows more consistent results.