972 resultados para Bayesian modelling
Resumo:
This project was a step forward in applying statistical methods and models to provide new insights for more informed decision-making at large spatial scales. The model has been designed to address complicated effects of ecological processes that govern the state of populations and uncertainties inherent in large spatio-temporal datasets. Specifically, the thesis contributes to better understanding and management of the Great Barrier Reef.
Resumo:
Large fruited spotted gum eucalypt Corymbia henryi occurs sympatrically with small fruited spotted gum Corymbia citriodora subspecies variegata over a large portion of its range on the east coast of Australia. The two taxa are interfertile, have overlapping flowering times and share a common set of insect and vertebrate pollinators. Previous genetic analysis of both taxa from two geographically remote sites suggested that the two were morphotypes rather than genetically distinct species. In this study we further explore this hypothesis of genic species by expanding sampling broadly through their sympatric locations and examine local-scale spatial genetic structure in stands that differ in species and age composition. Delineation of populations at five microsatellite loci, using an individual-based approach and Bayesian modelling, as well as clustering of individuals based on allele frequencies showed the two species to be molecularly homogeneous. Genetic structure aligned largely with geographic areas of origin, and followed an isolation-by-distance model, where proximal populations were generally less differentiated than more distant ones. At the stand level, spotted gums also generally showed little structure consistent with the high levels of gene flow inferred across the species range. Disturbances in the uniformity of structuring were detected, however, and attributed to localised events giving rise to even aged stands, probably due to regeneration from a few individuals following fire.
Resumo:
This paper presents a new series of AMS dates on ultrafiltered bone gelatin extracted from identified cutmarked or humanly-modified bones and teeth from the site of Abri Pataud, in the French Dordogne. The sequence of 32 new determinations provides a coherent and reliable chronology from the site's early Upper Palaeolithic levels 5-14, excavated by Hallam Movius. The results show that there were some problems with the previous series of dates, with many underestimating the real age. The new results, when calibrated and modelled using a Bayesian statistical method, allow detailed understanding of the pace of cultural changes within the Aurignacian I and II levels of the site, something not achievable before. In the future, the sequence of dates will allow wider comparison to similarly dated contexts elsewhere in Europe. High precision dating is only possible by using large suites of AMS dates from humanly-modified material within well understood archaeological sequences modelled using a Bayesian statistical method. © 2011.
Resumo:
Higham et al (2010) published a large series of new dates from the key French Palaeolithic site of the Grotte du Renne at Arcy-sur-Cure. The site is important because it is one of only two sites in Europe in which Châtelperronian lithic remains co-occur with Neanderthal human remains. A large series of dates from the Mousterian, Châtelperronian, Aurignacian and Gravettian levels of the site was obtained. The 14C results showed great variability, which Higham et al (2010) interpreted as most likely to be due to mixing of archaeological material in the site. In contrast, Caron et al (2011) suggested that the site stratigraphy is well preserved and that the problem with the variability in the radiocarbon ages was due to unremoved contamination in the dated bone. In this paper we address their critique of the original Higham et al (2010) paper
Resumo:
This paper outlines the results of a programme of radiocarbon dating and Bayesian modelling relating to an Early Bronze Age barrow cemetery at Over, Cambridgeshire. In total, 43 dates were obtained, enabling the first high-resolution independent chronology (relating to both burial and architectural events) to be constructed for a site of this kind. The results suggest that the three main turf-mound barrows were probably constructed and used successively rather than simultaneously, that the shift from inhumation to cremation seen on the site was not a straightforward progression, and that the four main ‘types’ of cremation burial in evidence were used throughout the life of the site. Overall, variability in terms of burial practice appears to have been a key feature of the site. The paper also considers the light that the fine-grained chronology developed can shed on recent much wider discussions of memory and time within Early Bronze Age barrows
Resumo:
Results of extensive site reconnaissance on the Isles of Tiree, Coll and north-west Mull, Inner Hebrides are presented. Pollen-stratigraphic records were compiled from a profile from Glen Aros, north-west Mull and from two profiles on Coll located at Loch an t-Sagairt and Caolas an Eilean. Quantification of microscopic charcoal provided records that were used to facilitate a preliminary evaluation of the causal driving mechanisms of vegetation change. Bayesian modelling of radiocarbon dates was used to construct preliminary chronological frameworks for these records. Basal sedimentary deposits at Glen Aros contain pollen records that correspond with vegetation succession typical of the early Holocene dating to c. 11,370 cal BP. Woodland development is a key feature of the pollen records dating to the early Holocene, while records from Loch an t-Sagairt show that blanket mire communities were widespread in north-west Coll by c. 9800 cal BP. The Corylus-rise is dated to c. 10,710 cal BP at Glen Aros and c. 9905 cal BP at Loch an t-Sagairt, with records indicating extensive cover of hazel woodland with birch. All of the major arboreal taxa were recorded, though Quercus and Ulmus were nowhere widespread. Analysis of wood charcoal remains from a Mesolithic site at Fiskary Bay, Coll indicate that Salix and Populus are likely to be under-represented in the pollen records. Reconstructed isopoll maps appear to underplay the importance of alder in western Scotland during the mid-Holocene. Alder-rise expansions in microscopic charcoal dating to c. 7300 cal BP at Glen Aros and c. 6510 to 5830 cal BP on Coll provide records of significance to the issue of human-induced burning related to the expansion of alder in Britain. Increasing frequencies in microscopic charcoal are correlated with mid-Holocene records of increasing aridity in western Scotland after c. 7490 cal BP at Glen Aros, 6760 cal BP at Loch an t-Sagairt and 6590 cal BP at Caolas an Eilean, while several phases of increasing bog surface wetness were detected in the Loch an t-Sagairt archive during the Holocene. At least five phases of small-scale woodland disturbance during the Mesolithic period were identified in the Glen Aros profile dating to c. 11,650 cal BP, 9300 cal BP, 7840 cal BP, 7040 cal BP and 6100 cal BP. The timing of the third phase is coincident with evidence of Mesolithic settlement at Creit Dhu, north-west Mull. Three phases of small-scale woodland disturbance were detected at Loch an t-Sagairt dating to c. 9270 cal BP, 8770 cal BP and 8270 cal BP, all of which overlap chronologically with evidence of Mesolithic activity at Fiskary Bay, Coll. A number of these episodes are aligned chronologically with phases of Holocene climate variability such as the 8.2 K event.
Resumo:
Forecasting, for obvious reasons, often become the most important goal to be achieved. For spatially extended systems (e.g. atmospheric system) where the local nonlinearities lead to the most unpredictable chaotic evolution, it is highly desirable to have a simple diagnostic tool to identify regions of predictable behaviour. In this paper, we discuss the use of the bred vector (BV) dimension, a recently introduced statistics, to identify the regimes where a finite time forecast is feasible. Using the tools from dynamical systems theory and Bayesian modelling, we show the finite time predictability in two-dimensional coupled map lattices in the regions of low BV dimension. © Indian Academy of Sciences.
Resumo:
Eukaryotic genomes display segmental patterns of variation in various properties, including GC content and degree of evolutionary conservation. DNA segmentation algorithms are aimed at identifying statistically significant boundaries between such segments. Such algorithms may provide a means of discovering new classes of functional elements in eukaryotic genomes. This paper presents a model and an algorithm for Bayesian DNA segmentation and considers the feasibility of using it to segment whole eukaryotic genomes. The algorithm is tested on a range of simulated and real DNA sequences, and the following conclusions are drawn. Firstly, the algorithm correctly identifies non-segmented sequence, and can thus be used to reject the null hypothesis of uniformity in the property of interest. Secondly, estimates of the number and locations of change-points produced by the algorithm are robust to variations in algorithm parameters and initial starting conditions and correspond to real features in the data. Thirdly, the algorithm is successfully used to segment human chromosome 1 according to GC content, thus demonstrating the feasibility of Bayesian segmentation of eukaryotic genomes. The software described in this paper is available from the author's website (www.uq.edu.au/similar to uqjkeith/) or upon request to the author.
Resumo:
There have been many models developed by scientists to assist decision-makers in making socio-economic and environmental decisions. It is now recognised that there is a shift in the dominant paradigm to making decisions with stakeholders, rather than making decisions for stakeholders. Our paper investigates two case studies where group model building has been undertaken for maintaining biodiversity in Australia. The first case study focuses on preservation and management of green spaces and biodiversity in metropolitan Melbourne under the umbrella of the Melbourne 2030 planning strategy. A geographical information system is used to collate a number of spatial datasets encompassing a range of cultural and natural assets data layers including: existing open spaces, waterways, threatened fauna and flora, ecological vegetation covers, registered cultural heritage sites, and existing land parcel zoning. Group model building is incorporated into the study through eliciting weightings and ratings of importance for each datasets from urban planners to formulate different urban green system scenarios. The second case study focuses on modelling ecoregions from spatial datasets for the state of Queensland. The modelling combines collaborative expert knowledge and a vast amount of environmental data to build biogeographical classifications of regions. An information elicitation process is used to capture expert knowledge of ecoregions as geographical descriptions, and to transform this into prior probability distributions that characterise regions in terms of environmental variables. This prior information is combined with measured data on the environmental variables within a Bayesian modelling technique to produce the final classified regions. We describe how linked views between descriptive information, mapping and statistical plots are used to decide upon representative regions that satisfy a number of criteria for biodiversity and conservation. This paper discusses the advantages and problems encountered when undertaking group model building. Future research will extend the group model building approach to include interested individuals and community groups.
Resumo:
Ecological problems are typically multi faceted and need to be addressed from a scientific and a management perspective. There is a wealth of modelling and simulation software available, each designed to address a particular aspect of the issue of concern. Choosing the appropriate tool, making sense of the disparate outputs, and taking decisions when little or no empirical data is available, are everyday challenges facing the ecologist and environmental manager. Bayesian Networks provide a statistical modelling framework that enables analysis and integration of information in its own right as well as integration of a variety of models addressing different aspects of a common overall problem. There has been increased interest in the use of BNs to model environmental systems and issues of concern. However, the development of more sophisticated BNs, utilising dynamic and object oriented (OO) features, is still at the frontier of ecological research. Such features are particularly appealing in an ecological context, since the underlying facts are often spatial and temporal in nature. This thesis focuses on an integrated BN approach which facilitates OO modelling. Our research devises a new heuristic method, the Iterative Bayesian Network Development Cycle (IBNDC), for the development of BN models within a multi-field and multi-expert context. Expert elicitation is a popular method used to quantify BNs when data is sparse, but expert knowledge is abundant. The resulting BNs need to be substantiated and validated taking this uncertainty into account. Our research demonstrates the application of the IBNDC approach to support these aspects of BN modelling. The complex nature of environmental issues makes them ideal case studies for the proposed integrated approach to modelling. Moreover, they lend themselves to a series of integrated sub-networks describing different scientific components, combining scientific and management perspectives, or pooling similar contributions developed in different locations by different research groups. In southern Africa the two largest free-ranging cheetah (Acinonyx jubatus) populations are in Namibia and Botswana, where the majority of cheetahs are located outside protected areas. Consequently, cheetah conservation in these two countries is focussed primarily on the free-ranging populations as well as the mitigation of conflict between humans and cheetahs. In contrast, in neighbouring South Africa, the majority of cheetahs are found in fenced reserves. Nonetheless, conflict between humans and cheetahs remains an issue here. Conservation effort in South Africa is also focussed on managing the geographically isolated cheetah populations as one large meta-population. Relocation is one option among a suite of tools used to resolve human-cheetah conflict in southern Africa. Successfully relocating captured problem cheetahs, and maintaining a viable free-ranging cheetah population, are two environmental issues in cheetah conservation forming the first case study in this thesis. The second case study involves the initiation of blooms of Lyngbya majuscula, a blue-green algae, in Deception Bay, Australia. L. majuscula is a toxic algal bloom which has severe health, ecological and economic impacts on the community located in the vicinity of this algal bloom. Deception Bay is an important tourist destination with its proximity to Brisbane, Australia’s third largest city. Lyngbya is one of several algae considered to be a Harmful Algal Bloom (HAB). This group of algae includes other widespread blooms such as red tides. The occurrence of Lyngbya blooms is not a local phenomenon, but blooms of this toxic weed occur in coastal waters worldwide. With the increase in frequency and extent of these HAB blooms, it is important to gain a better understanding of the underlying factors contributing to the initiation and sustenance of these blooms. This knowledge will contribute to better management practices and the identification of those management actions which could prevent or diminish the severity of these blooms.
Resumo:
This article explores the use of probabilistic classification, namely finite mixture modelling, for identification of complex disease phenotypes, given cross-sectional data. In particular, if focuses on posterior probabilities of subgroup membership, a standard output of finite mixture modelling, and how the quantification of uncertainty in these probabilities can lead to more detailed analyses. Using a Bayesian approach, we describe two practical uses of this uncertainty: (i) as a means of describing a person’s membership to a single or multiple latent subgroups and (ii) as a means of describing identified subgroups by patient-centred covariates not included in model estimation. These proposed uses are demonstrated on a case study in Parkinson’s disease (PD), where latent subgroups are identified using multiple symptoms from the Unified Parkinson’s Disease Rating Scale (UPDRS).
Resumo:
Mixture models are a flexible tool for unsupervised clustering that have found popularity in a vast array of research areas. In studies of medicine, the use of mixtures holds the potential to greatly enhance our understanding of patient responses through the identification of clinically meaningful clusters that, given the complexity of many data sources, may otherwise by intangible. Furthermore, when developed in the Bayesian framework, mixture models provide a natural means for capturing and propagating uncertainty in different aspects of a clustering solution, arguably resulting in richer analyses of the population under study. This thesis aims to investigate the use of Bayesian mixture models in analysing varied and detailed sources of patient information collected in the study of complex disease. The first aim of this thesis is to showcase the flexibility of mixture models in modelling markedly different types of data. In particular, we examine three common variants on the mixture model, namely, finite mixtures, Dirichlet Process mixtures and hidden Markov models. Beyond the development and application of these models to different sources of data, this thesis also focuses on modelling different aspects relating to uncertainty in clustering. Examples of clustering uncertainty considered are uncertainty in a patient’s true cluster membership and accounting for uncertainty in the true number of clusters present. Finally, this thesis aims to address and propose solutions to the task of comparing clustering solutions, whether this be comparing patients or observations assigned to different subgroups or comparing clustering solutions over multiple datasets. To address these aims, we consider a case study in Parkinson’s disease (PD), a complex and commonly diagnosed neurodegenerative disorder. In particular, two commonly collected sources of patient information are considered. The first source of data are on symptoms associated with PD, recorded using the Unified Parkinson’s Disease Rating Scale (UPDRS) and constitutes the first half of this thesis. The second half of this thesis is dedicated to the analysis of microelectrode recordings collected during Deep Brain Stimulation (DBS), a popular palliative treatment for advanced PD. Analysis of this second source of data centers on the problems of unsupervised detection and sorting of action potentials or "spikes" in recordings of multiple cell activity, providing valuable information on real time neural activity in the brain.
Resumo:
Reliable pollutant build-up prediction plays a critical role in the accuracy of urban stormwater quality modelling outcomes. However, water quality data collection is resource demanding compared to streamflow data monitoring, where a greater quantity of data is generally available. Consequently, available water quality data sets span only relatively short time scales unlike water quantity data. Therefore, the ability to take due consideration of the variability associated with pollutant processes and natural phenomena is constrained. This in turn gives rise to uncertainty in the modelling outcomes as research has shown that pollutant loadings on catchment surfaces and rainfall within an area can vary considerably over space and time scales. Therefore, the assessment of model uncertainty is an essential element of informed decision making in urban stormwater management. This paper presents the application of a range of regression approaches such as ordinary least squares regression, weighted least squares Regression and Bayesian Weighted Least Squares Regression for the estimation of uncertainty associated with pollutant build-up prediction using limited data sets. The study outcomes confirmed that the use of ordinary least squares regression with fixed model inputs and limited observational data may not provide realistic estimates. The stochastic nature of the dependent and independent variables need to be taken into consideration in pollutant build-up prediction. It was found that the use of the Bayesian approach along with the Monte Carlo simulation technique provides a powerful tool, which attempts to make the best use of the available knowledge in the prediction and thereby presents a practical solution to counteract the limitations which are otherwise imposed on water quality modelling.
Resumo:
This paper presents a novel framework for the modelling of passenger facilitation in a complex environment. The research is motivated by the challenges in the airport complex system, where there are multiple stakeholders, differing operational objectives and complex interactions and interdependencies between different parts of the airport system. Traditional methods for airport terminal modelling do not explicitly address the need for understanding causal relationships in a dynamic environment. Additionally, existing Bayesian Network (BN) models, which provide a means for capturing causal relationships, only present a static snapshot of a system. A method to integrate a BN complex systems model with stochastic queuing theory is developed based on the properties of the Poisson and Exponential distributions. The resultant Hybrid Queue-based Bayesian Network (HQBN) framework enables the simulation of arbitrary factors, their relationships, and their effects on passenger flow and vice versa. A case study implementation of the framework is demonstrated on the inbound passenger facilitation process at Brisbane International Airport. The predicted outputs of the model, in terms of cumulative passenger flow at intermediary and end points in the inbound process, are found to have an $R^2$ goodness of fit of 0.9994 and 0.9982 respectively over a 10 hour test period. The utility of the framework is demonstrated on a number of usage scenarios including real time monitoring and `what-if' analysis. This framework provides the ability to analyse and simulate a dynamic complex system, and can be applied to other socio-technical systems such as hospitals.