936 resultados para Maximum entropy
Resumo:
Using surface charts at 0330GMT, the movement df the monsoon trough during the months June to September 1990 al two fixed longitudes, namely 79 degrees E and 85 degrees E, is studied. The probability distribution of trough position shows that the median, mean and mode occur at progressively more northern latitudes, especially at 85 degrees E, with a pronounced mode that is close to the northern-most limit reached by the trough. A spectral analysis of the fluctuating latitudinal position of the trough is carried out using FFT and the Maximum Entropy Method (MEM). Both methods show significant peaks around 7.5 and 2.6 days, and a less significant one around 40-50 days. The two peaks at the shorter period are more prominent at the eastern longitude. MEM shows an additional peak around 15 days. A study of the weather systems that occurred during the season shows them to have a duration around 3 days and an interval between systems of around 9 days, suggesting a possible correlation with the dominant short periods observed in the spectrum of trough position.
Resumo:
Doppler weather radars with fast scanning rates must estimate spectral moments based on a small number of echo samples. This paper concerns the estimation of mean Doppler velocity in a coherent radar using a short complex time series. Specific results are presented based on 16 samples. A wide range of signal-to-noise ratios are considered, and attention is given to ease of implementation. It is shown that FFT estimators fare poorly in low SNR and/or high spectrum-width situations. Several variants of a vector pulse-pair processor are postulated and an algorithm is developed for the resolution of phase angle ambiguity. This processor is found to be better than conventional processors at very low SNR values. A feasible approximation to the maximum entropy estimator is derived as well as a technique utilizing the maximization of the periodogram. It is found that a vector pulse-pair processor operating with four lags for clear air observation and a single lag (pulse-pair mode) for storm observation may be a good way to estimate Doppler velocities over the entire gamut of weather phenomena.
Resumo:
Species distribution modelling (SDM) typically analyses species’ presence together with some form of absence information. Ideally absences comprise observations or are inferred from comprehensive sampling. When such information is not available, then pseudo-absences are often generated from the background locations within the study region of interest containing the presences, or else absence is implied through the comparison of presences to the whole study region, e.g. as is the case in Maximum Entropy (MaxEnt) or Poisson point process modelling. However, the choice of which absence information to include can be both challenging and highly influential on SDM predictions (e.g. Oksanen and Minchin, 2002). In practice, the use of pseudo- or implied absences often leads to an imbalance where absences far outnumber presences. This leaves analysis highly susceptible to ‘naughty-noughts’: absences that occur beyond the envelope of the species, which can exert strong influence on the model and its predictions (Austin and Meyers, 1996). Also known as ‘excess zeros’, naughty noughts can be estimated via an overall proportion in simple hurdle or mixture models (Martin et al., 2005). However, absences, especially those that occur beyond the species envelope, can often be more diverse than presences. Here we consider an extension to excess zero models. The two-staged approach first exploits the compartmentalisation provided by classification trees (CTs) (as in O’Leary, 2008) to identify multiple sources of naughty noughts and simultaneously delineate several species envelopes. Then SDMs can be fit separately within each envelope, and for this stage, we examine both CTs (as in Falk et al., 2014) and the popular MaxEnt (Elith et al., 2006). We introduce a wider range of model performance measures to improve treatment of naughty noughts in SDM. We retain an overall measure of model performance, the area under the curve (AUC) of the Receiver-Operating Curve (ROC), but focus on its constituent measures of false negative rate (FNR) and false positive rate (FPR), and how these relate to the threshold in the predicted probability of presence that delimits predicted presence from absence. We also propose error rates more relevant to users of predictions: false omission rate (FOR), the chance that a predicted absence corresponds to (and hence wastes) an observed presence, and the false discovery rate (FDR), reflecting those predicted (or potential) presences that correspond to absence. A high FDR may be desirable since it could help target future search efforts, whereas zero or low FOR is desirable since it indicates none of the (often valuable) presences have been ignored in the SDM. For illustration, we chose Bradypus variegatus, a species that has previously been published as an exemplar species for MaxEnt, proposed by Phillips et al. (2006). We used CTs to increasingly refine the species envelope, starting with the whole study region (E0), eliminating more and more potential naughty noughts (E1–E3). When combined with an SDM fit within the species envelope, the best CT SDM had similar AUC and FPR to the best MaxEnt SDM, but otherwise performed better. The FNR and FOR were greatly reduced, suggesting that CTs handle absences better. Interestingly, MaxEnt predictions showed low discriminatory performance, with the most common predicted probability of presence being in the same range (0.00-0.20) for both true absences and presences. In summary, this example shows that SDMs can be improved by introducing an initial hurdle to identify naughty noughts and partition the envelope before applying SDMs. This improvement was barely detectable via AUC and FPR yet visible in FOR, FNR, and the comparison of predicted probability of presence distribution for pres/absence.
Resumo:
The quality of species distribution models (SDMs) relies to a large degree on the quality of the input data, from bioclimatic indices to environmental and habitat descriptors (Austin, 2002). Recent reviews of SDM techniques, have sought to optimize predictive performance e.g. Elith et al., 2006. In general SDMs employ one of three approaches to variable selection. The simplest approach relies on the expert to select the variables, as in environmental niche models Nix, 1986 or a generalized linear model without variable selection (Miller and Franklin, 2002). A second approach explicitly incorporates variable selection into model fitting, which allows examination of particular combinations of variables. Examples include generalized linear or additive models with variable selection (Hastie et al. 2002); or classification trees with complexity or model based pruning (Breiman et al., 1984, Zeileis, 2008). A third approach uses model averaging, to summarize the overall contribution of a variable, without considering particular combinations. Examples include neural networks, boosted or bagged regression trees and Maximum Entropy as compared in Elith et al. 2006. Typically, users of SDMs will either consider a small number of variable sets, via the first approach, or else supply all of the candidate variables (often numbering more than a hundred) to the second or third approaches. Bayesian SDMs exist, with several methods for eliciting and encoding priors on model parameters (see review in Low Choy et al. 2010). However few methods have been published for informative variable selection; one example is Bayesian trees (O’Leary 2008). Here we report an elicitation protocol that helps makes explicit a priori expert judgements on the quality of candidate variables. This protocol can be flexibly applied to any of the three approaches to variable selection, described above, Bayesian or otherwise. We demonstrate how this information can be obtained then used to guide variable selection in classical or machine learning SDMs, or to define priors within Bayesian SDMs.
Resumo:
Given the limited resources available for weed management, a strategic approach is required to give the best bang for your buck. The current study incorporates: (1) a model ensemble approach to identify areas of uncertainty and commonality regarding a species invasive potential, (2) current distribution of the invaded species, and (3) connectivity of systems to identify target regions and focus efforts for more effective management. Uncertainty in the prediction of suitable habitat for H. amplexicaulis (study species) in Australia was addressed in an ensemble-forecasting approach to compare distributional scenarios from four models (CLIMATCH; CLIMEX; boosted regression trees [BRT]; maximum entropy [Maxent]). Models were built using subsets of occurrence and environmental data. Catchment risk was determined through incorporating habitat suitability, the current abundance and distribution of H. amplexicaulis, and catchment connectivity. Our results indicate geographic differences between predictions of different approaches. Despite these differences a number of catchments in northern, central, and southern Australia were identified as high risk of invasion or further spread by all models suggesting they should be given priority for the management of H. amplexicaulis. The study also highlighted the utility of ensemble approaches in indentifying areas of uncertainty and commonality regarding the species invasive potential.
Resumo:
We present some results on multicarrier analysis of magnetotransport data, Both synthetic as well as data from narrow gap Hg0.8Cd0.2Te samples are used to demonstrate applicability of various algorithms vs. nonlinear least square fitting, Quantitative Mobility Spectrum Analysis (QMSA) and Maximum Entropy Mobility Spectrum Analysis (MEMSA). Comments are made from our experience oil these algorithms, and, on the inversion procedure from experimental R/sigma-B to S-mu specifically with least square fitting as an example. Amongst the conclusions drawn are: (i) Experimentally measured resistivity (R-xx, R-xy) should also be used instead of just the inverted conductivity (sigma(xx), sigma(xy)) to fit data to semiclassical expressions for better fits especially at higher B. (ii) High magnetic field is necessary to extract low mobility carrier parameters. (iii) Provided the error in data is not large, better estimates to carrier parameters of remaining carrier species can be obtained at any stage by subtracting highest mobility carrier contribution to sigma from the experimental data and fitting with the remaining carriers. (iv)Even in presence of high electric field, an approximate multicarrier expression can be used to guess the carrier mobilities and their variations before solving the full Boltzmann equation.
Resumo:
We present a theory of multichannel disordered conductors by directly studying the statistical distribution of the transfer matrix for the full system. The theory is based on the general properties of the scattering system: flux conservation, time-reversal invariance, and the appropriate combination requirement when two wires are put together. The distribution associated with systems of very small length is then selected on the basis of a maximum-entropy criterion; a fixed value is assumed for the diffusion coefficient that characterizes the evolution of the distribution as the length increases. We obtain a diffusion equation for the probability distribution and compute the average of a few relevant quantities.
Resumo:
It is well known that in the time-domain acquisition of NMR data, signal-to-noise (S/N) improves as the square root of the number of transients accumulated. However, the amplitude of the measured signal varies during the time of detection, having a functional form dependent on the coherence detected. Matching the time spent signal averaging to the expected amplitude of the signal observed should also improve the detected signal-to-noise. Following this reasoning, Barna et al. (J Magn. Reson.75, 384, 1987) demonstrated the utility of exponential sampling in one- and two-dimensional NMR, using maximum-entropy methods to analyze the data. It is proposed here that for two-dimensional experiments the exponential sampling be replaced by exponential averaging. The data thus collected can be analyzed by standard fast-Fourier-transform routines. We demonstrate the utility of exponential averaging in 2D NOESY spectra of the protein ubiquitin, in which an enhanced SIN is observed. It is also shown that the method acquires delayed double-quantum-filtered COSY without phase distortion.
Resumo:
We report studies of magnetocaloric effect in lanthanum cobaltate doped with different Sr-concentrations, La1-xSrxCoO3 (0.05 less than or equal to x less than or equal to 0.4). The study has revealed that La0.6Sr0.4CoO3, which exhibits a moderately large value of maximum entropy change of about 1.45 J/kg/K in 1.5 T DC-fieid around its Curie temperature (of 235 K), can be used as an active magnetic refrigerant (AMR) material at similar to 2.35 K. (C) 1999 Elsevier Science B.V. Al rights reserved.
Resumo:
Predictive distribution modelling of Berberis aristata DC, a rare threatened plant with high medicinal values has been done with an aim to understand its potential distribution zones in Indian Himalayan region. Bioclimatic and topographic variables were used to develop the distribution model with the help of three different algorithms viz. GeneticAlgorithm for Rule-set Production (GARP), Bioclim and Maximum entroys(MaxEnt). Maximum entropy has predicted wider potential distribution (10.36%) compared to GARP (4.63%) and Bioclim (2.44%). Validation confirms that these outputs are comparable to the present distribution pattern of the B. atistata. This exercise highlights that this species favours Western Himalaya. However, GARP and MaxEnt's prediction of Eastern Himalayan states (i.e. Arunachal Pradesh, Nagaland and Manipur) are also identified as potential occurrence places require further exploration.
Resumo:
Users can rarely reveal their information need in full detail to a search engine within 1--2 words, so search engines need to "hedge their bets" and present diverse results within the precious 10 response slots. Diversity in ranking is of much recent interest. Most existing solutions estimate the marginal utility of an item given a set of items already in the response, and then use variants of greedy set cover. Others design graphs with the items as nodes and choose diverse items based on visit rates (PageRank). Here we introduce a radically new and natural formulation of diversity as finding centers in resistive graphs. Unlike in PageRank, we do not specify the edge resistances (equivalently, conductances) and ask for node visit rates. Instead, we look for a sparse set of center nodes so that the effective conductance from the center to the rest of the graph has maximum entropy. We give a cogent semantic justification for turning PageRank thus on its head. In marked deviation from prior work, our edge resistances are learnt from training data. Inference and learning are NP-hard, but we give practical solutions. In extensive experiments with subtopic retrieval, social network search, and document summarization, our approach convincingly surpasses recently-published diversity algorithms like subtopic cover, max-marginal relevance (MMR), Grasshopper, DivRank, and SVMdiv.
Resumo:
In a quantum system, there may be many density matrices associated with a state on an algebra of observables. For each density matrix, one can compute its entropy. These are, in general, different. Therefore, one reaches the remarkable possibility that there may be many entropies for a given state R. Sorkin (private communication)]. This ambiguity in entropy can often be traced to a gauge symmetry emergent from the nontrivial topological character of the configuration space of the underlying system. It can also happen in finite-dimensional matrix models. In the present work, we discuss this entropy ambiguity and its consequences for an ethylene molecule. This is a very simple and well-known system, where these notions can be put to tests. Of particular interest in this discussion is the fact that the change of the density matrix with the corresponding entropy increase drives the system towards the maximally disordered state with maximum entropy, where Boltzman's formula applies. Besides its intrinsic conceptual interest, the simplicity of this model can serve as an introduction to a similar discussion of systems such as colored monopoles and the breaking of color symmetry.
Resumo:
Rising global temperatures threaten the survival of many plant and animal species. Having already risen at an unprecedented rate in the past century, temperatures are predicted to rise between 0.3 and 7.5C in North America over the next 100 years (Hawkes et al. 2007). Studies have documented the effects of climate warming on phenology (timing of seasonal activities), with observations of early arrival at breeding grounds, earlier ends to the reproductive season, and delayed autumnal migrations (Pike et al. 2006). In addition, for species not suited to the physiological demands of cold winter temperatures, increasing temperatures could shift tolerable habitats to higher latitudes (Hawkes et al. 2007). More directly, climate warming will impact thermally sensitive species like sea turtles, who exhibit temperature-dependent sexual determination. Temperatures in the middle third of the incubation period determine the sex of sea turtle offspring, with higher temperatures resulting in a greater abundance of female offspring. Consequently, increasing temperatures from climate warming would drastically change the offspring sex ratio (Hawkes et al. 2007). Of the seven extant species of sea turtles, three (leatherback, Kemp’s ridley, and hawksbill) are critically endangered, two (olive ridley and green) are endangered, and one (loggerhead) is threatened. Considering the predicted scenarios of climate warming and the already tenuous status of sea turtle populations, it is essential that efforts are made to understand how increasing temperatures may affect sea turtle populations and how these species might adapt in the face of such changes. In this analysis, I seek to identify the impact of changing climate conditions over the next 50 years on the availability of sea turtle nesting habitat in Florida given predicted changes in temperature and precipitation. I predict that future conditions in Florida will be less suitable for sea turtle nesting during the historic nesting season. This may imply that sea turtles will nest at a different time of year, in more northern latitudes, to a lesser extent, or possibly not at all. It seems likely that changes in temperature and precipitation patterns will alter the distribution of sea turtle nesting locations worldwide, provided that beaches where the conditions are suitable for nesting still exist. Hijmans and Graham (2006) evaluate a range of climate envelope models in terms of their ability to predict species distributions under climate change scenarios. Their results suggested that the choice of species distribution model is dependent on the specifics of each individual study. Fuller et al. (2008) used a maximum entropy approach to model the potential distribution of 11 species in the Arctic Coastal Plain of Alaska under a series of projected climate scenarios. Recently, Pike (in press) developed Maxent models to investigate the impacts of climate change on green sea turtle nest distribution and timing. In each of these studies, a set of environmental predictor variables (including climate variables), for which ‘current’ conditions are available and ‘future’ conditions have been projected, is used in conjunction with species occurrence data to map potential species distribution under the projected conditions. In this study, I will take a similar approach in mapping the potential sea turtle nesting habitat in Florida by developing a Maxent model based on environmental and climate data and projecting the model for future climate data. (PDF contains 5 pages)
Resumo:
This thesis aims at a simple one-parameter macroscopic model of distributed damage and fracture of polymers that is amenable to a straightforward and efficient numerical implementation. The failure model is motivated by post-mortem fractographic observations of void nucleation, growth and coalescence in polyurea stretched to failure, and accounts for the specific fracture energy per unit area attendant to rupture of the material.
Furthermore, it is shown that the macroscopic model can be rigorously derived, in the sense of optimal scaling, from a micromechanical model of chain elasticity and failure regularized by means of fractional strain-gradient elasticity. Optimal scaling laws that supply a link between the single parameter of the macroscopic model, namely the critical energy-release rate of the material, and micromechanical parameters pertaining to the elasticity and strength of the polymer chains, and to the strain-gradient elasticity regularization, are derived. Based on optimal scaling laws, it is shown how the critical energy-release rate of specific materials can be determined from test data. In addition, the scope and fidelity of the model is demonstrated by means of an example of application, namely Taylor-impact experiments of polyurea rods. Hereby, optimal transportation meshfree approximation schemes using maximum-entropy interpolation functions are employed.
Finally, a different crazing model using full derivatives of the deformation gradient and a core cut-off is presented, along with a numerical non-local regularization model. The numerical model takes into account higher-order deformation gradients in a finite element framework. It is shown how the introduction of non-locality into the model stabilizes the effect of strain localization to small volumes in materials undergoing softening. From an investigation of craze formation in the limit of large deformations, convergence studies verifying scaling properties of both local- and non-local energy contributions are presented.