920 resultados para probability distribution


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The influence of hydrological dynamics on vegetation distribution and the structuring of wetland environments is of growing interest as wetlands are modified by human action and the increasing threat from climate change. Hydrological properties have long been considered a driving force in structuring wetland communities. We link hydrological dynamics with vegetation distribution across Everglades National Park (ENP) using two publicly available datasets to study the probability structure of the frequency, duration, and depth of inundation events along with their relationship to vegetation distribution. This study is among the first to show hydrologic structuring of vegetation communities at wide spatial and temporal scales, as results indicate that the percentage of time a location is inundated and its mean depth are the principal structuring variables to which individual communities respond. For example, sawgrass, the most abundant vegetation type within the ENP, is found across a wide range of time inundated percentages and mean depths. Meanwhile, other communities like pine savanna or red mangrove scrub are more restricted in their distribution and found disproportionately at particular depths and inundations. These results, along with the probabilistic structure of hydropatterns, potentially allow for the evaluation of climate change impacts on wetland vegetation community structure and distribution.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The three-parameter lognormal distribution is the extension of the two-parameter lognormal distribution to meet the need of the biological, sociological, and other fields. Numerous research papers have been published for the parameter estimation problems for the lognormal distributions. The inclusion of the location parameter brings in some technical difficulties for the parameter estimation problems, especially for the interval estimation. This paper proposes a method for constructing exact confidence intervals and exact upper confidence limits for the location parameter of the three-parameter lognormal distribution. The point estimation problem is discussed as well. The performance of the point estimator is compared with the maximum likelihood estimator, which is widely used in practice. Simulation result shows that the proposed method is less biased in estimating the location parameter. The large sample size case is discussed in the paper.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis proposes some confidence intervals for the mean of a positively skewed distribution. The following confidence intervals are considered: Student-t, Johnson-t, median-t, mad-t, bootstrap-t, BCA, T1 , T3 and six new confidence intervals, the median bootstrap-t, mad bootstrap-t, median T1, mad T1 , median T3 and the mad T3. A simulation study has been conducted and average widths, coefficient of variation of widths, and coverage probabilities were recorded and compared across confidence intervals. To compare confidence intervals, the width and coverage probabilities were compared so that smaller widths indicated a better confidence interval when coverage probabilities were the same. Results showed that the median T1 and median T3 outperformed other confidence intervals in terms of coverage probability and the mad bootstrap-t, mad-t, and mad T3 outperformed others in terms of width. Some real life data are considered to illustrate the findings of the thesis.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Human use of the oceans is increasingly in conflict with conservation of endangered species. Methods for managing the spatial and temporal placement of industries such as military, fishing, transportation and offshore energy, have historically been post hoc; i.e. the time and place of human activity is often already determined before assessment of environmental impacts. In this dissertation, I build robust species distribution models in two case study areas, US Atlantic (Best et al. 2012) and British Columbia (Best et al. 2015), predicting presence and abundance respectively, from scientific surveys. These models are then applied to novel decision frameworks for preemptively suggesting optimal placement of human activities in space and time to minimize ecological impacts: siting for offshore wind energy development, and routing ships to minimize risk of striking whales. Both decision frameworks relate the tradeoff between conservation risk and industry profit with synchronized variable and map views as online spatial decision support systems.

For siting offshore wind energy development (OWED) in the U.S. Atlantic (chapter 4), bird density maps are combined across species with weights of OWED sensitivity to collision and displacement and 10 km2 sites are compared against OWED profitability based on average annual wind speed at 90m hub heights and distance to transmission grid. A spatial decision support system enables toggling between the map and tradeoff plot views by site. A selected site can be inspected for sensitivity to a cetaceans throughout the year, so as to capture months of the year which minimize episodic impacts of pre-operational activities such as seismic airgun surveying and pile driving.

Routing ships to avoid whale strikes (chapter 5) can be similarly viewed as a tradeoff, but is a different problem spatially. A cumulative cost surface is generated from density surface maps and conservation status of cetaceans, before applying as a resistance surface to calculate least-cost routes between start and end locations, i.e. ports and entrance locations to study areas. Varying a multiplier to the cost surface enables calculation of multiple routes with different costs to conservation of cetaceans versus cost to transportation industry, measured as distance. Similar to the siting chapter, a spatial decisions support system enables toggling between the map and tradeoff plot view of proposed routes. The user can also input arbitrary start and end locations to calculate the tradeoff on the fly.

Essential to the input of these decision frameworks are distributions of the species. The two preceding chapters comprise species distribution models from two case study areas, U.S. Atlantic (chapter 2) and British Columbia (chapter 3), predicting presence and density, respectively. Although density is preferred to estimate potential biological removal, per Marine Mammal Protection Act requirements in the U.S., all the necessary parameters, especially distance and angle of observation, are less readily available across publicly mined datasets.

In the case of predicting cetacean presence in the U.S. Atlantic (chapter 2), I extracted datasets from the online OBIS-SEAMAP geo-database, and integrated scientific surveys conducted by ship (n=36) and aircraft (n=16), weighting a Generalized Additive Model by minutes surveyed within space-time grid cells to harmonize effort between the two survey platforms. For each of 16 cetacean species guilds, I predicted the probability of occurrence from static environmental variables (water depth, distance to shore, distance to continental shelf break) and time-varying conditions (monthly sea-surface temperature). To generate maps of presence vs. absence, Receiver Operator Characteristic (ROC) curves were used to define the optimal threshold that minimizes false positive and false negative error rates. I integrated model outputs, including tables (species in guilds, input surveys) and plots (fit of environmental variables, ROC curve), into an online spatial decision support system, allowing for easy navigation of models by taxon, region, season, and data provider.

For predicting cetacean density within the inner waters of British Columbia (chapter 3), I calculated density from systematic, line-transect marine mammal surveys over multiple years and seasons (summer 2004, 2005, 2008, and spring/autumn 2007) conducted by Raincoast Conservation Foundation. Abundance estimates were calculated using two different methods: Conventional Distance Sampling (CDS) and Density Surface Modelling (DSM). CDS generates a single density estimate for each stratum, whereas DSM explicitly models spatial variation and offers potential for greater precision by incorporating environmental predictors. Although DSM yields a more relevant product for the purposes of marine spatial planning, CDS has proven to be useful in cases where there are fewer observations available for seasonal and inter-annual comparison, particularly for the scarcely observed elephant seal. Abundance estimates are provided on a stratum-specific basis. Steller sea lions and harbour seals are further differentiated by ‘hauled out’ and ‘in water’. This analysis updates previous estimates (Williams & Thomas 2007) by including additional years of effort, providing greater spatial precision with the DSM method over CDS, novel reporting for spring and autumn seasons (rather than summer alone), and providing new abundance estimates for Steller sea lion and northern elephant seal. In addition to providing a baseline of marine mammal abundance and distribution, against which future changes can be compared, this information offers the opportunity to assess the risks posed to marine mammals by existing and emerging threats, such as fisheries bycatch, ship strikes, and increased oil spill and ocean noise issues associated with increases of container ship and oil tanker traffic in British Columbia’s continental shelf waters.

Starting with marine animal observations at specific coordinates and times, I combine these data with environmental data, often satellite derived, to produce seascape predictions generalizable in space and time. These habitat-based models enable prediction of encounter rates and, in the case of density surface models, abundance that can then be applied to management scenarios. Specific human activities, OWED and shipping, are then compared within a tradeoff decision support framework, enabling interchangeable map and tradeoff plot views. These products make complex processes transparent for gaming conservation, industry and stakeholders towards optimal marine spatial management, fundamental to the tenets of marine spatial planning, ecosystem-based management and dynamic ocean management.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Dirichlet distribution is a multivariate generalization of the Beta distribution. It is an important multivariate continuous distribution in probability and statistics. In this report, we review the Dirichlet distribution and study its properties, including statistical and information-theoretic quantities involving this distribution. Also, relationships between the Dirichlet distribution and other distributions are discussed. There are some different ways to think about generating random variables with a Dirichlet distribution. The stick-breaking approach and the Pólya urn method are discussed. In Bayesian statistics, the Dirichlet distribution and the generalized Dirichlet distribution can both be a conjugate prior for the Multinomial distribution. The Dirichlet distribution has many applications in different fields. We focus on the unsupervised learning of a finite mixture model based on the Dirichlet distribution. The Initialization Algorithm and Dirichlet Mixture Estimation Algorithm are both reviewed for estimating the parameters of a Dirichlet mixture. Three experimental results are shown for the estimation of artificial histograms, summarization of image databases and human skin detection.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Aims. Projected rotational velocities (ve sin i) have been estimated for 334 targets in the VLT-FLAMES Tarantula Survey that do not manifest significant radial velocity variations and are not supergiants. They have spectral types from approximately O9.5 to B3. The estimates have been analysed to infer the underlying rotational velocity distribution, which is critical for understanding the evolution of massive stars. Methods. Projected rotational velocities were deduced from the Fourier transforms of spectral lines, with upper limits also being obtained from profile fitting. For the narrower lined stars, metal and non-diffuse helium lines were adopted, and for the broader lined stars, both non-diffuse and diffuse helium lines; the estimates obtained using the different sets of lines are in good agreement. The uncertainty in the mean estimates is typically 4% for most targets. The iterative deconvolution procedure of Lucy has been used to deduce the probability density distribution of the rotational velocities. Results. Projected rotational velocities range up to approximately 450 kms-1 and show a bi-modal structure. This is also present in the inferred rotational velocity distribution with 25% of the sample having 0 <ve <100 km s-1 and the high velocity component having ve ∼ 250 km s-1. There is no evidence from the spatial and radial velocity distributions of the two components that they represent either field and cluster populations or different episodes of star formation. Be-type stars have also been identified. Conclusions. The bi-modal rotational velocity distribution in our sample resembles that found for late-B and early-A type stars.While magnetic braking appears to be a possible mechanism for producing the low-velocity component, we can not rule out alternative explanations. © ESO 2013.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Schedules can be built in a similar way to a human scheduler by using a set of rules that involve domain knowledge. This paper presents an Estimation of Distribution Algorithm (EDA) for the nurse scheduling problem, which involves choosing a suitable scheduling rule from a set for the assignment of each nurse. Unlike previous work that used Genetic Algorithms (GAs) to implement implicit learning, the learning in the proposed algorithm is explicit, i.e. we identify and mix building blocks directly. The EDA is applied to implement such explicit learning by building a Bayesian network of the joint distribution of solutions. The conditional probability of each variable in the network is computed according to an initial set of promising solutions. Subsequently, each new instance for each variable is generated by using the corresponding conditional probabilities, until all variables have been generated, i.e. in our case, a new rule string has been obtained. Another set of rule strings will be generated in this way, some of which will replace previous strings based on fitness selection. If stopping conditions are not met, the conditional probabilities for all nodes in the Bayesian network are updated again using the current set of promising rule strings. Computational results from 52 real data instances demonstrate the success of this approach. It is also suggested that the learning mechanism in the proposed approach might be suitable for other scheduling problems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Vector-borne disease emergence in recent decades has been associated with different environmental drivers including changes in habitat, hosts and climate. Lyme borreliosis is among the most important vector-borne diseases in the Northern hemisphere and is an emerging disease in Scotland. Transmitted by Ixodid tick vectors between large numbers of wild vertebrate host species, Lyme borreliosis is caused by bacteria from the Borrelia burgdorferi sensu lato species group. Ecological studies can inform how environmental factors such as host abundance and community composition, habitat and landscape heterogeneity contribute to spatial and temporal variation in risk from B. burgdorferi s.l. In this thesis a range of approaches were used to investigate the effects of vertebrate host communities and individual host species as drivers of B. burgdorferi s.l. dynamics and its tick vector Ixodes ricinus. Host species differ in reservoir competence for B. burgdorferi s.l. and as hosts for ticks. Deer are incompetent transmission hosts for B. burgdorferi s.l. but are significant hosts of all life-stages of I. ricinus. Rodents and birds are important transmission hosts of B. burgdorferi s.l. and common hosts of immature life-stages of I. ricinus. In this thesis, surveys of woodland sites revealed variable effects of deer density on B. burgdorferi prevalence, from no effect (Chapter 2) to a possible ‘dilution’ effect resulting in lower prevalence at higher deer densities (Chapter 3). An invasive species in Scotland, the grey squirrel (Sciurus carolinensis), was found to host diverse genotypes of B. burgdorferi s.l. and may act as a spill-over host for strains maintained by native host species (Chapter 4). Habitat fragmentation may alter the dynamics of B. burgdorferi s.l. via effects on the host community and host movements. In this thesis, there was lack of persistence of the rodent associated genospecies of B. burgdorferi s.l. within a naturally fragmented landscape (Chapter 3). Rodent host biology, particularly population cycles and dispersal ability are likely to affect pathogen persistence and recolonization in fragmented habitats. Heterogeneity in disease dynamics can occur spatially and temporally due to differences in the host community, habitat and climatic factors. Higher numbers of I. ricinus nymphs, and a higher probability of detecting a nymph infected with B. burgdorferi s.l., were found in areas with warmer climates estimated by growing degree days (Chapter 2). The ground vegetation type associated with the highest number of I. ricinus nymphs varied between studies in this thesis (Chapter 2 & 3) and does not appear to be a reliable predictor across large areas. B. burgdorferi s.l. prevalence and genospecies composition was highly variable for the same sites sampled in subsequent years (Chapter 2). This suggests that dynamic variables such as reservoir host densities and deer should be measured as well as more static habitat and climatic factors to understand the drivers of B. burgdorferi s.l. infection in ticks. Heterogeneity in parasite loads amongst hosts is a common finding which has implications for disease ecology and management. Using a 17-year data set for tick infestations in a wild bird community in Scotland, different effects of age and sex on tick burdens were found among four species of passerine bird (Chapter 5). There were also different rates of decline in tick burdens among bird species in response to a long term decrease in questing tick pressure over the study. Species specific patterns may be driven by differences in behaviour and immunity and highlight the importance of comparative approaches. Combining whole genome sequencing (WGS) and population genetics approaches offers a novel approach to identify ecological drivers of pathogen populations. An initial analysis of WGS from B. burgdorferi s.s. isolates sampled 16 years apart suggests that there is a signal of measurable evolution (Chapter 6). This suggests demographic analyses may be applied to understand ecological and evolutionary processes of these bacteria. This work shows how host communities, habitat and climatic factors can affect the local transmission dynamics of B. burgdorferi s.l. and the potential risk of infection to humans. Spatial and temporal heterogeneity in pathogen dynamics poses challenges for the prediction of risk. New tools such as WGS of the pathogen (Chapter 6) and blood meal analysis techniques will add power to future studies on the ecology and evolution of B. burgdorferi s.l.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Schedules can be built in a similar way to a human scheduler by using a set of rules that involve domain knowledge. This paper presents an Estimation of Distribution Algorithm (EDA) for the nurse scheduling problem, which involves choosing a suitable scheduling rule from a set for the assignment of each nurse. Unlike previous work that used Genetic Algorithms (GAs) to implement implicit learning, the learning in the proposed algorithm is explicit, i.e. we identify and mix building blocks directly. The EDA is applied to implement such explicit learning by building a Bayesian network of the joint distribution of solutions. The conditional probability of each variable in the network is computed according to an initial set of promising solutions. Subsequently, each new instance for each variable is generated by using the corresponding conditional probabilities, until all variables have been generated, i.e. in our case, a new rule string has been obtained. Another set of rule strings will be generated in this way, some of which will replace previous strings based on fitness selection. If stopping conditions are not met, the conditional probabilities for all nodes in the Bayesian network are updated again using the current set of promising rule strings. Computational results from 52 real data instances demonstrate the success of this approach. It is also suggested that the learning mechanism in the proposed approach might be suitable for other scheduling problems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Schedules can be built in a similar way to a human scheduler by using a set of rules that involve domain knowledge. This paper presents an Estimation of Distribution Algorithm (EDA) for the nurse scheduling problem, which involves choosing a suitable scheduling rule from a set for the assignment of each nurse. Unlike previous work that used Genetic Algorithms (GAs) to implement implicit learning, the learning in the proposed algorithm is explicit, i.e. we identify and mix building blocks directly. The EDA is applied to implement such explicit learning by building a Bayesian network of the joint distribution of solutions. The conditional probability of each variable in the network is computed according to an initial set of promising solutions. Subsequently, each new instance for each variable is generated by using the corresponding conditional probabilities, until all variables have been generated, i.e. in our case, a new rule string has been obtained. Another set of rule strings will be generated in this way, some of which will replace previous strings based on fitness selection. If stopping conditions are not met, the conditional probabilities for all nodes in the Bayesian network are updated again using the current set of promising rule strings. Computational results from 52 real data instances demonstrate the success of this approach. It is also suggested that the learning mechanism in the proposed approach might be suitable for other scheduling problems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Growing models have been widely used for clustering or topology learning. Traditionally these models work on stationary environments, grow incrementally and adapt their nodes to a given distribution based on global parameters. In this paper, we present an enhanced unsupervised self-organising network for the modelling of visual objects. We first develop a framework for building non-rigid shapes using the growth mechanism of the self-organising maps, and then we define an optimal number of nodes without overfitting or underfitting the network based on the knowledge obtained from information-theoretic considerations. We present experimental results for hands and we quantitatively evaluate the matching capabilities of the proposed method with the topographic product.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this work, we report a 20-ns constant pressure molecular dynamics simulation of prilocaine (PLC), in amine-amide local anesthetic, in a hydrated liquid crystal bilayer of 1-palmitoyl-2-oleoyl-sn-glycero-3-phosphatidylcholine. The partition of PLC induces the lateral expansion of the bilayer and a concomitant contraction in its thickness. PLC molecules are preferentially found in the hydrophobic acyl chains region, with a maximum probability at similar to 12 angstrom from the center of the bilayer (between the C(4) and C(5) methylene groups). A decrease in the acyl chain segmental order parameter, vertical bar S-CD vertical bar, compared to neat bilayers, is found, in good agreement with experimental H-2-NMR studies. The decrease in vertical bar S-CD vertical bar induced by PLC is attributed to a larger accessible volume per lipid in the acyl chain region. (C) 2008 Wiley Periodicals, Inc.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Several deterministic and probabilistic methods are used to evaluate the probability of seismically induced liquefaction of a soil. The probabilistic models usually possess some uncertainty in that model and uncertainties in the parameters used to develop that model. These model uncertainties vary from one statistical model to another. Most of the model uncertainties are epistemic, and can be addressed through appropriate knowledge of the statistical model. One such epistemic model uncertainty in evaluating liquefaction potential using a probabilistic model such as logistic regression is sampling bias. Sampling bias is the difference between the class distribution in the sample used for developing the statistical model and the true population distribution of liquefaction and non-liquefaction instances. Recent studies have shown that sampling bias can significantly affect the predicted probability using a statistical model. To address this epistemic uncertainty, a new approach was developed for evaluating the probability of seismically-induced soil liquefaction, in which a logistic regression model in combination with Hosmer-Lemeshow statistic was used. This approach was used to estimate the population (true) distribution of liquefaction to non-liquefaction instances of standard penetration test (SPT) and cone penetration test (CPT) based most updated case histories. Apart from this, other model uncertainties such as distribution of explanatory variables and significance of explanatory variables were also addressed using KS test and Wald statistic respectively. Moreover, based on estimated population distribution, logistic regression equations were proposed to calculate the probability of liquefaction for both SPT and CPT based case history. Additionally, the proposed probability curves were compared with existing probability curves based on SPT and CPT case histories.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This report discusses the calculation of analytic second-order bias techniques for the maximum likelihood estimates (for short, MLEs) of the unknown parameters of the distribution in quality and reliability analysis. It is well-known that the MLEs are widely used to estimate the unknown parameters of the probability distributions due to their various desirable properties; for example, the MLEs are asymptotically unbiased, consistent, and asymptotically normal. However, many of these properties depend on an extremely large sample sizes. Those properties, such as unbiasedness, may not be valid for small or even moderate sample sizes, which are more practical in real data applications. Therefore, some bias-corrected techniques for the MLEs are desired in practice, especially when the sample size is small. Two commonly used popular techniques to reduce the bias of the MLEs, are ‘preventive’ and ‘corrective’ approaches. They both can reduce the bias of the MLEs to order O(n−2), whereas the ‘preventive’ approach does not have an explicit closed form expression. Consequently, we mainly focus on the ‘corrective’ approach in this report. To illustrate the importance of the bias-correction in practice, we apply the bias-corrected method to two popular lifetime distributions: the inverse Lindley distribution and the weighted Lindley distribution. Numerical studies based on the two distributions show that the considered bias-corrected technique is highly recommended over other commonly used estimators without bias-correction. Therefore, special attention should be paid when we estimate the unknown parameters of the probability distributions under the scenario in which the sample size is small or moderate.