529 resultados para Bootstrap
Resumo:
1. Distance sampling is a widely used technique for estimating the size or density of biological populations. Many distance sampling designs and most analyses use the software Distance. 2. We briefly review distance sampling and its assumptions, outline the history, structure and capabilities of Distance, and provide hints on its use. 3. Good survey design is a crucial prerequisite for obtaining reliable results. Distance has a survey design engine, with a built-in geographic information system, that allows properties of different proposed designs to be examined via simulation, and survey plans to be generated. 4. A first step in analysis of distance sampling data is modeling the probability of detection. Distance contains three increasingly sophisticated analysis engines for this: conventional distance sampling, which models detection probability as a function of distance from the transect and assumes all objects at zero distance are detected; multiple-covariate distance sampling, which allows covariates in addition to distance; and mark–recapture distance sampling, which relaxes the assumption of certain detection at zero distance. 5. All three engines allow estimation of density or abundance, stratified if required, with associated measures of precision calculated either analytically or via the bootstrap. 6. Advanced analysis topics covered include the use of multipliers to allow analysis of indirect surveys (such as dung or nest surveys), the density surface modeling analysis engine for spatial and habitat-modeling, and information about accessing the analysis engines directly from other software. 7. Synthesis and applications. Distance sampling is a key method for producing abundance and density estimates in challenging field conditions. The theory underlying the methods continues to expand to cope with realistic estimation situations. In step with theoretical developments, state-of- the-art software that implements these methods is described that makes the methods accessible to practicing ecologists.
Resumo:
The Gumbel distribution is perhaps the most widely applied statistical distribution for problems in engineering. We propose a generalization-referred to as the Kumaraswamy Gumbel distribution-and provide a comprehensive treatment of its structural properties. We obtain the analytical shapes of the density and hazard rate functions. We calculate explicit expressions for the moments and generating function. The variation of the skewness and kurtosis measures is examined and the asymptotic distribution of the extreme values is investigated. Explicit expressions are also derived for the moments of order statistics. The methods of maximum likelihood and parametric bootstrap and a Bayesian procedure are proposed for estimating the model parameters. We obtain the expected information matrix. An application of the new model to a real dataset illustrates the potentiality of the proposed model. Two bivariate generalizations of the model are proposed.
Resumo:
Assessment of the suitability of anthropogenic landscapes for wildlife species is crucial for setting priorities for biodiversity conservation. This study aimed to analyse the environmental suitability of a highly fragmented region of the Brazilian Atlantic Forest, one of the world's 25 recognized biodiversity hotspots, for forest bird species. Eight forest bird species were selected for the analyses, based on point counts (n = 122) conducted in April-September 2006 and January-March 2009. Six additional variables (landscape diversity, distance from forest and streams, aspect, elevation and slope) were modelled in Maxent for (1) actual and (2) simulated land cover, based on the forest expansion required by existing Brazilian forest legislation. Models were evaluated by bootstrap or jackknife methods and their performance was assessed by AUC, omission error, binomial probability or p value. All predictive models were statistically significant, with high AUC values and low omission errors. A small proportion of the actual landscape (24.41 +/- 6.31%) was suitable for forest bird species. The simulated landscapes lead to an increase of c. 30% in total suitable areas. In average, models predicted a small increase (23.69 +/- 6.95%) in the area of suitable native forest for bird species. Being close to forest increased the environmental suitability of landscapes for all bird species; landscape diversity was also a significant factor for some species. In conclusion, this study demonstrates that species distribution modelling (SDM) successfully predicted bird distribution across a heterogeneous landscape at fine spatial resolution, as all models were biologically relevant and statistically significant. The use of landscape variables as predictors contributed significantly to the results, particularly for species distributions over small extents and at fine scales. This is the first study to evaluate the environmental suitability of the remaining Brazilian Atlantic Forest for bird species in an agricultural landscape, and provides important additional data for regional environmental planning.
Resumo:
Bovine coronavirus has been associated with diarrhoea in newborn calves, winter dysentery in adult cattle and respiratory tract infections in calves and feedlot cattle. In Cuba, the presence of BCoV was first reported in 2006. Since then, sporadic outbreaks have continued to occur. This study was aimed at deepening the knowledge of the evolution, molecular markers of virulence and epidemiology of BCoV in Cuba. A total of 30 samples collected between 2009 and 2011 were used for PCR amplification and direct sequencing of partial or full S gene. Sequence comparison and phylogenetic studies were conducted using partial or complete S gene sequences as phylogenetic markers. All Cuban bovine coronavirus sequences were located in a single cluster supported by 100% bootstrap and 1.00 posterior probability values. The Cuban bovine coronavirus sequences were also clustered with the USA BCoV strains corresponding to the GenBank accession numbers EF424621 and EF424623, suggesting a common origin for these viruses. This phylogenetic cluster was also the only group of sequences in which no recombination events were detected. Of the 45 amino acid changes found in the Cuban strains, four were unique. (C) 2012 Elsevier B.V. All rights reserved.
Resumo:
Long-term survival models have historically been considered for analyzing time-to-event data with long-term survivors fraction. However, situations in which a fraction (1 - p) of systems is subject to failure from independent competing causes of failure, while the remaining proportion p is cured or has not presented the event of interest during the time period of the study, have not been fully considered in the literature. In order to accommodate such situations, we present in this paper a new long-term survival model. Maximum likelihood estimation procedure is discussed as well as interval estimation and hypothesis tests. A real dataset illustrates the methodology.
Resumo:
During 2008D2010, ticks were collected from road-killed wild animals within the Serra dos Orgaos National Park area in the state of Rio de Janeiro, Brazil. In total, 193 tick specimens were collected, including Amblyomma dubitatum Neumann and Amblyomma cajennense (F.) from four Hydrochoerus hydrochaeris (L.), Amblyomma calcaratum Neumann and A. cajennense from four Tamandua tetradactyla (L.), Amblyomma aureolatum (Pallas) and A. cajennense from five Cerdocyon thous L., Amblyomma longirostre (Koch) from one Sphiggurus villosus (Cuvier), Amblyomma varium Koch from three Bradypus variegatus Schinz, and A. cajennense from one Buteogallus meridionalis (Latham). Molecular analyses based on polymerase chain reaction targeting two rickettsial genes (gltA and ompA) on tick DNA extracts showed that 70.6% (12/17) of the A. dubitatum adult ticks, and all Amblyomma sp. nymphal pools collected from capybaras were shown to contain rickettsial DNA, which after DNA sequencing, revealed to be 100% identical to the recently identified Rickettsia sp. strain Pampulha from A. dubitatum ticks collected in the state of Minas Gerais, Brazil. Phylogenetic analysis with concatenated sequences (gltA-ompA) showed that our sequence from A. dubitatum ticks, referred to Rickettsia sp. strain Serra dos Orgaos, segregated under 99% bootstrap support in a same cluster with Old World rickettsiae, namely R. tamurae, R. monacensis, and Rickettsia sp. strain 774e. Because A. dubitatum is known to bite humans, the potential role of Rickettsia sp. strain Serra dos Orgaos as human pathogen must be taken into account, because both R. tamurae and R. monacencis have been reported infecting human beings.
Resumo:
Background Statistical methods for estimating usual intake require at least two short-term dietary measurements in a subsample of the target population. However, the percentage of individuals with a second dietary measurement (replication rate) may influence the precision of estimates, such as percentiles and proportions of individuals below cut-offs of intake. Objective To investigate the precision of the usual food intake estimates using different replication rates and different sample sizes. Participants/setting Adolescents participating in the continuous National Health and Nutrition Examination Survey 2007-2008 (n=1,304) who completed two 24-hour recalls. Statistical analyses performed The National Cancer Institute method was used to estimate the usual intake of dark green vegetables in the original sample comprising 1,304 adolescents with a replication rate of 100%. A bootstrap with 100 replications was performed to estimate CIs for percentiles and proportions of individuals below cut-offs of intake. Using the same bootstrap replications, four sets of data sets were sampled with different replication rates (80%, 60%, 40%, and 20%). For each data set created, the National Cancer Institute method was performed and percentiles, Cl, and proportions of individuals below cut-offs were calculated. Precision estimates were checked by comparing each Cl obtained from data sets with different replication rates with the Cl obtained from original data set. Further, we sampled 1,000, 750, 500, and 250 individuals from the original data set, and performed the same analytical procedures. Results Percentiles of intake and percentage of individuals below the cut-off points were similar throughout the replication rates and sample sizes, but the Cl increased as the replication rate decreased. Wider CIs were observed at 40% and 20% of replication rate. Conclusions The precision of the usual intake estimates decreased when low replication rates were used. However, even with different sample sizes, replication rates >40% may not lead to an important loss of precision. J Acad Nutr Diet. 2012;112:1015-1020.
Resumo:
The development of new statistical and computational methods is increasingly making it possible to bridge the gap between hard sciences and humanities. In this study, we propose an approach based on a quantitative evaluation of attributes of objects in fields of humanities, from which concepts such as dialectics and opposition are formally defined mathematically. As case studies, we analyzed the temporal evolution of classical music and philosophy by obtaining data for 8 features characterizing the corresponding fields for 7 well-known composers and philosophers, which were treated with multivariate statistics and pattern recognition methods. A bootstrap method was applied to avoid statistical bias caused by the small sample data set, with which hundreds of artificial composers and philosophers were generated, influenced by the 7 names originally chosen. Upon defining indices for opposition, skewness and counter-dialectics, we confirmed the intuitive analysis of historians in that classical music evolved according to a master apprentice tradition, while in philosophy changes were driven by opposition. Though these case studies were meant only to show the possibility of treating phenomena in humanities quantitatively, including a quantitative measure of concepts such as dialectics and opposition, the results are encouraging for further application of the approach presented here to many other areas, since it is entirely generic.
Resumo:
The self-consistency of a thermodynamical theory for hadronic systems based on the non-extensive statistics is investigated. We show that it is possible to obtain a self-consistent theory according to the asymptotic bootstrap principle if the mass spectrum and the energy density increase q-exponentially. A direct consequence is the existence of a limiting effective temperature for the hadronic system. We show that this result is in agreement with experiments. (C) 2012 Elsevier B.V. All rights reserved.
Resumo:
Fraud is a global problem that has required more attention due to an accentuated expansion of modern technology and communication. When statistical techniques are used to detect fraud, whether a fraud detection model is accurate enough in order to provide correct classification of the case as a fraudulent or legitimate is a critical factor. In this context, the concept of bootstrap aggregating (bagging) arises. The basic idea is to generate multiple classifiers by obtaining the predicted values from the adjusted models to several replicated datasets and then combining them into a single predictive classification in order to improve the classification accuracy. In this paper, for the first time, we aim to present a pioneer study of the performance of the discrete and continuous k-dependence probabilistic networks within the context of bagging predictors classification. Via a large simulation study and various real datasets, we discovered that the probabilistic networks are a strong modeling option with high predictive capacity and with a high increment using the bagging procedure when compared to traditional techniques. (C) 2012 Elsevier Ltd. All rights reserved.
Resumo:
Abstract Background Vampire bat related rabies harms both livestock industry and public health sector in central Brazil. The geographical distributions of vampire bat-transmitted rabies virus variants are delimited by mountain chains. These findings were elucidated by analyzing a high conserved nucleoprotein gene. This study aims to elucidate the detailed epidemiological characters of vampire bat-transmitted rabies virus by phylogenetic methods based on 619-nt sequence including unconserved G-L intergenic region. Findings The vampire bat-transmitted rabies virus isolates divided into 8 phylogenetic lineages in the previous nucleoprotein gene analysis were divided into 10 phylogenetic lineages with significant bootstrap values. The distributions of most variants were reconfirmed to be delimited by mountain chains. Furthermore, variants in undulating areas have narrow distributions and are apparently separated by mountain ridges. Conclusions This study demonstrates that the 619-nt sequence including G-L intergenic region is more useful for a state-level phylogenetic analysis of rabies virus than the partial nucleoprotein gene, and simultaneously that the distribution of vampire bat-transmitted RABV variants tends to be separated not only by mountain chains but also by mountain ridges, thus suggesting that the diversity of vampire bat-transmitted RABV variants was delimited by geographical undulations.
Resumo:
[ES] Este Trabajo de Fin de Grado ha tenido como objetivo el desarrollo de un gestor de menús de restaurantes como aplicación web para una empresa que ofrece hostings de menús y publicidad mediante la publicación de dichos menús en pantallas y portales web. Las empresas asociadas (bares y restaurantes) podrán elaborar menús compuestos de dos platos (primero y segundo), postre y bebidas para ser ‘enviados’ al servicio de publicación. La aplicación proporciona un sistema de gestión de dichos menús facilitando la reutilización de platos entre menús, la personalización de la imagen representativa de cada plato, así como diversas operaciones de copia, visualización y modificación de los menús y de los platos. Los usuarios registrados tendrán la posibilidad de recuperar su contraseña de forma automática en caso de que la misma sea olvidada. La información relacionada con los platos, menús y usuarios registrados será almacenada automáticamente sobre una base de datos diseñada al efecto. Por otro lado, la aplicación web dispone de una página accesible únicamente para el administrador para la gestión de los usuarios, por ejemplo, editar, alta, baja, habilitar y deshabilitar cuentas de usuarios. Por último, las tecnologías y herramientas utilizadas en la elaboración de este trabajo incluyen Php, Mysql, jQuery, CSS, HTML y sobre todo el framework Twitter Bootstrap que ha sido de gran ayuda en el desarrollo del proyecto.
Resumo:
[ES] Juego Stacker para HTML5 propone una aplicación web con dos modalidades de juegos basados en el clásico Stacker. La modalidad classic stacker pretende simular dicho juego, en la cual el jugador ha de apilar una fila horizontal de cuadrados que se desplazan a velocidad constante horizontalmente sobre otra fila horizontal de cuadrados que se encuentran en la parte inferior sin describir ningún movimiento. La velocidad de movimiento de la fila que ha de apilar irá aumentando conforme se vayan superando los niveles. El juego acaba cuando no dispone de más cuadrados en la fila, que se perderán si no se consigue apilar de manera exacta. La otra modalidad de juego se le conoce como super stacker. En esta modalidad, el jugador ha de apilar una serie de figuras con formas distintas sobre otras figuras estáticas que forman parte de un mundo generado. Las figuras que ha de apilar el jugador son sensibles a fuerzas tales como la gravedad, colisión entre objetos, fricción, etc. Si alguna de estas figuras entra en contacto con alguno de los límites del mundo, el jugador ha perdido. Ganará cuando la estructura final formada aguanta un número de segundos determinados, pasando así a otro nivel (escenario) de mayor complejidad. Para esta modalidad de juego se ha necesitado un motor físico portado a Javascript que simule las fuerzas mencionadas anteriormente. Resaltar también que se ha optado por realizar un diseño adaptable utilizando frameworks como bootstrap 3 debido al gran auge de los dispositivos móviles con dimensiones de pantalla variables.
Resumo:
La valutazione dell’intensità secondo una procedura formale trasparente, obiettiva e che permetta di ottenere valori numerici attraverso scelte e criteri rigorosi, rappresenta un passo ed un obiettivo per la trattazione e l’impiego delle informazioni macrosismiche. I dati macrosismici possono infatti avere importanti applicazioni per analisi sismotettoniche e per la stima della pericolosità sismica. Questa tesi ha affrontato il problema del formalismo della stima dell’intensità migliorando aspetti sia teorici che pratici attraverso tre passaggi fondamentali sviluppati in ambiente MS-Excel e Matlab: i) la raccolta e l’archiviazione del dataset macrosismico; ii), l’associazione (funzione di appartenenza o membership function) tra effetti e gradi di intensità della scala macrosismica attraverso i principi della logica dei fuzzy sets; iii) l’applicazione di algoritmi decisionali rigorosi ed obiettivi per la stima dell’intensità finale. L’intera procedura è stata applicata a sette terremoti italiani sfruttando varie possibilità, anche metodologiche, come la costruzione di funzioni di appartenenza combinando le informazioni macrosismiche di più terremoti: Monte Baldo (1876), Valle d’Illasi (1891), Marsica (1915), Santa Sofia (1918), Mugello (1919), Garfagnana (1920) e Irpinia (1930). I risultati ottenuti hanno fornito un buon accordo statistico con le intensità di un catalogo macrosismico di riferimento confermando la validità dell’intera metodologia. Le intensità ricavate sono state poi utilizzate per analisi sismotettoniche nelle aree dei terremoti studiati. I metodi di analisi statistica sui piani quotati (distribuzione geografica delle intensità assegnate) si sono rivelate in passato uno strumento potente per analisi e caratterizzazione sismotettonica, determinando i principali parametri (localizzazione epicentrale, lunghezza, larghezza, orientazione) della possibile sorgente sismogenica. Questa tesi ha implementato alcuni aspetti delle metodologie di analisi grazie a specifiche applicazioni sviluppate in Matlab che hanno permesso anche di stimare le incertezze associate ai parametri di sorgente, grazie a tecniche di ricampionamento statistico. Un’analisi sistematica per i terremoti studiati è stata portata avanti combinando i vari metodi per la stima dei parametri di sorgente con i piani quotati originali e ricalcolati attraverso le procedure decisionali fuzzy. I risultati ottenuti hanno consentito di valutare le caratteristiche delle possibili sorgenti e formulare ipotesi di natura sismotettonica che hanno avuto alcuni riscontri indiziali con dati di tipo geologico e geologico-strutturale. Alcuni eventi (1915, 1918, 1920) presentano una forte stabilità dei parametri calcolati (localizzazione epicentrale e geometria della possibile sorgente) con piccole incertezze associate. Altri eventi (1891, 1919 e 1930) hanno invece mostrato una maggiore variabilità sia nella localizzazione dell’epicentro che nella geometria delle box: per il primo evento ciò è probabilmente da mettere in relazione con la ridotta consistenza del dataset di intensità mentre per gli altri con la possibile molteplicità delle sorgenti sismogenetiche. Anche l’analisi bootstrap ha messo in evidenza, in alcuni casi, le possibili asimmetrie nelle distribuzioni di alcuni parametri (ad es. l’azimut della possibile struttura), che potrebbero suggerire meccanismi di rottura su più faglie distinte.
Resumo:
We use data from about 700 GPS stations in the EuroMediterranen region to investigate the present-day behavior of the the Calabrian subduction zone within the Mediterranean-scale plates kinematics and to perform local scale studies about the strain accumulation on active structures. We focus attenction on the Messina Straits and Crati Valley faults where GPS data show extentional velocity gradients of ∼3 mm/yr and ∼2 mm/yr, respectively. We use dislocation model and a non-linear constrained optimization algorithm to invert for fault geometric parameters and slip-rates and evaluate the associated uncertainties adopting a bootstrap approach. Our analysis suggest the presence of two partially locked normal faults. To investigate the impact of elastic strain contributes from other nearby active faults onto the observed velocity gradient we use a block modeling approach. Our models show that the inferred slip-rates on the two analyzed structures are strongly impacted by the assumed locking width of the Calabrian subduction thrust. In order to frame the observed local deformation features within the present- day central Mediterranean kinematics we realyze a statistical analysis testing the indipendent motion (w.r.t. the African and Eurasias plates) of the Adriatic, Cal- abrian and Sicilian blocks. Our preferred model confirms a microplate like behaviour for all the investigated blocks. Within these kinematic boundary conditions we fur- ther investigate the Calabrian Slab interface geometry using a combined approach of block modeling and χ2ν statistic. Almost no information is obtained using only the horizontal GPS velocities that prove to be a not sufficient dataset for a multi-parametric inversion approach. Trying to stronger constrain the slab geometry we estimate the predicted vertical velocities performing suites of forward models of elastic dislocations varying the fault locking depth. Comparison with the observed field suggest a maximum resolved locking depth of 25 km.