982 resultados para Statistics of extremes
Resumo:
2000 Mathematics Subject Classification: 62G32, 62G20.
Resumo:
Background: Genome wide association studies (GWAS) are becoming the approach of choice to identify genetic determinants of complex phenotypes and common diseases. The astonishing amount of generated data and the use of distinct genotyping platforms with variable genomic coverage are still analytical challenges. Imputation algorithms combine directly genotyped markers information with haplotypic structure for the population of interest for the inference of a badly genotyped or missing marker and are considered a near zero cost approach to allow the comparison and combination of data generated in different studies. Several reports stated that imputed markers have an overall acceptable accuracy but no published report has performed a pair wise comparison of imputed and empiric association statistics of a complete set of GWAS markers. Results: In this report we identified a total of 73 imputed markers that yielded a nominally statistically significant association at P < 10(-5) for type 2 Diabetes Mellitus and compared them with results obtained based on empirical allelic frequencies. Interestingly, despite their overall high correlation, association statistics based on imputed frequencies were discordant in 35 of the 73 (47%) associated markers, considerably inflating the type I error rate of imputed markers. We comprehensively tested several quality thresholds, the haplotypic structure underlying imputed markers and the use of flanking markers as predictors of inaccurate association statistics derived from imputed markers. Conclusions: Our results suggest that association statistics from imputed markers showing specific MAF (Minor Allele Frequencies) range, located in weak linkage disequilibrium blocks or strongly deviating from local patterns of association are prone to have inflated false positive association signals. The present study highlights the potential of imputation procedures and proposes simple procedures for selecting the best imputed markers for follow-up genotyping studies.
Resumo:
Within the framework of a retrospective study of the incidence of hip fractures in the canton of Vaud (Switzerland), all cases of hip fracture occurring among the resident population in 1986 and treated in the hospitals of the canton were identified from among five different information sources. Relevant data were then extracted from the medical records. At least two sources of information were used to identify cases in each hospital, among them the statistics of the Swiss Hospital Association (VESKA). These statistics were available for 9 of the 18 hospitals in the canton that participated in the study. The number of cases identified from the VESKA statistics was compared to the total number of cases for each hospital. For the 9 hospitals the number of cases in the VESKA statistics was 407, whereas, after having excluded diagnoses that were actually "status after fracture" and double entries, the total for these hospitals was 392, that is 4% less than the VESKA statistics indicate. It is concluded that the VESKA statistics provide a good approximation of the actual number of cases treated in these hospitals, with a tendency to overestimate this number. In order to use these statistics for calculating incidence figures, however, it is imperative that a greater proportion of all hospitals (50% presently in the canton, 35% nationwide) participate in these statistics.
Resumo:
Statistics of causes of death remain an important source of epidemiological data for the evaluation of various medical and health problems. The improvement of analytical techniques and, above all, the transformation of demographic and morbid structures of populations have prompted researchers in the field to give more importance to the quality of death certificates. After describing the data collection system presently used in Switzerland, the paper discusses various indirect estimations of the quality of Swiss data and reviews the corresponding international literature.
Resumo:
Understanding niche evolution, dynamics, and the response of species to climate change requires knowledge of the determinants of the environmental niche and species range limits. Mean values of climatic variables are often used in such analyses. In contrast, the increasing frequency of climate extremes suggests the importance of understanding their additional influence on range limits. Here, we assess how measures representing climate extremes (i.e., interannual variability in climate parameters) explain and predict spatial patterns of 11 tree species in Switzerland. We find clear, although comparably small, improvement (+20% in adjusted D(2), +8% and +3% in cross-validated True Skill Statistic and area under the receiver operating characteristics curve values) in models that use measures of extremes in addition to means. The primary effect of including information on climate extremes is a correction of local overprediction and underprediction. Our results demonstrate that measures of climate extremes are important for understanding the climatic limits of tree species and assessing species niche characteristics. The inclusion of climate variability likely will improve models of species range limits under future conditions, where changes in mean climate and increased variability are expected.
Resumo:
This paper proposes a nonparametric test in order to establish the level of accuracy of theforeign trade statistics of 17 Latin American countries when contrasted with the trade statistics of the main partners in 1925. The Wilcoxon Matched-Pairs Ranks test is used to determine whether the differences between the data registered by exporters and importers are meaningful, and if so, whether the differences are systematic in any direction. The paper tests for the reliability of the data registered for two homogeneous products, petroleum and coal, both in volume and value. The conclusion of the several exercises performed is that we cannot accept the existence of statistically significant differences between the data provided by the exporters and the registered by the importing countries in most cases. The qualitative historiography of Latin American describes its foreign trade statistics as mostly unusable. Our quantitative results contest this view.
Resumo:
Genome-wide association studies (GWASs) have identified many genetic variants underlying complex traits. Many detected genetic loci harbor variants that associate with multiple-even distinct-traits. Most current analysis approaches focus on single traits, even though the final results from multiple traits are evaluated together. Such approaches miss the opportunity to systemically integrate the phenome-wide data available for genetic association analysis. In this study, we propose a general approach that can integrate association evidence from summary statistics of multiple traits, either correlated, independent, continuous, or binary traits, which might come from the same or different studies. We allow for trait heterogeneity effects. Population structure and cryptic relatedness can also be controlled. Our simulations suggest that the proposed method has improved statistical power over single-trait analysis in most of the cases we studied. We applied our method to the Continental Origins and Genetic Epidemiology Network (COGENT) African ancestry samples for three blood pressure traits and identified four loci (CHIC2, HOXA-EVX1, IGFBP1/IGFBP3, and CDH17; p < 5.0 × 10(-8)) associated with hypertension-related traits that were missed by a single-trait analysis in the original report. Six additional loci with suggestive association evidence (p < 5.0 × 10(-7)) were also observed, including CACNA1D and WNT3. Our study strongly suggests that analyzing multiple phenotypes can improve statistical power and that such analysis can be executed with the summary statistics from GWASs. Our method also provides a way to study a cross phenotype (CP) association by using summary statistics from GWASs of multiple phenotypes.
Resumo:
The intensity-duration-frequency occurrence ratio (IDF) is a tool commonly used for precipitation-runoff data transformation, which is established from observations of intense precipitations over a period sufficiently long as to allow the occurrence of extremes at the observation site. This study focused on verifying the existence or absence of new data, in terms of IDF ratio, by using partial duration records produced from data on maximum daily disaggregated rainfall for pre determined durations. The partial duration records considered a base value of 55 mm, totaling 279 values. After the rainfall series were established, their independence and seasonality were assessed. Using the Student's t-test statistics, it was established that no new data, as IDF ratio, emerged from the analysis of the partial duration series with the recommended base value of precipitation, as compared to the historical records.
Resumo:
The importance of temperature in the determination of the yield of an annual crop (groundnut; Arachis hypogaea L. in India) was assessed. Simulations from a regional climate model (PRECIS) were used with a crop model (GLAM) to examine crop growth under simulated current (1961-1990) and future (2071-2100) climates. Two processes were examined: the response of crop duration to mean temperature and the response of seed-set to extremes of temperature. The relative importance of, and interaction between, these two processes was examined for a number of genotypic characteristics, which were represented by using different values of crop model parameters derived from experiments. The impact of mean and extreme temperatures varied geographically, and depended upon the simulated genotypic properties. High temperature stress was not a major determinant of simulated yields in the current climate, but affected the mean and variability of yield under climate change in two regions which had contrasting statistics of daily maximum temperature. Changes in mean temperature had a similar impact on mean yield to that of high temperature stress in some locations and its effects were more widespread. Where the optimal temperature for development was exceeded, the resulting increase in duration in some simulations fully mitigated the negative impacts of extreme temperatures when sufficient water was available for the extended growing period. For some simulations the reduction in mean yield between the current and future climates was as large as 70%, indicating the importance of genotypic adaptation to changes in both means and extremes of temperature under climate change. (c) 2006 Elsevier B.V. All rights reserved.
Resumo:
Hourly sea level records from 1954 to 2012 at 20 tide gauges at and adjacent to the Chinese coasts are used to analyze extremes in sea level and in tidal residual. Tides and tropical cyclones determine the spatial distribution of sea level maxima. Tidal residual maxima are predominantly determined by tropical cyclones. The 50 year return level is found to be sensitive to the number of extreme events used in the estimation. This is caused by the small number of tropical cyclone events happening each year which lead to other local storm events included thus significantly affecting the estimates. Significant increase in sea level extremes is found with trends in the range between 2.0 and 14.1 mm yr−1. The trends are primarily driven by changes in median sea level but also linked with increases in tidal amplitudes at three stations. Tropical cyclones cause significant interannual variations in the extremes. The interannual variability in the sea level extremes is also influenced by the changes in median sea level at the north and by the 18.6 year nodal cycle at the South China Sea. Neither of PDO and ENSO is found to be an indicator of changes in the size of extremes, but ENSO appears to regulate the number of tropical cyclones that reach the Chinese coasts. Global mean atmospheric temperature appears to be a good descriptor of the interannual variability of tidal residual extremes induced by tropical cyclones but the trend in global temperature is inconsistent with the lack of trend in the residuals.
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Includes bibliography
Resumo:
The Spin-Statistics theorem states that the statistics of a system of identical particles is determined by their spin: Particles of integer spin are Bosons (i.e. obey Bose-Einstein statistics), whereas particles of half-integer spin are Fermions (i.e. obey Fermi-Dirac statistics). Since the original proof by Fierz and Pauli, it has been known that the connection between Spin and Statistics follows from the general principles of relativistic Quantum Field Theory. In spite of this, there are different approaches to Spin-Statistics and it is not clear whether the theorem holds under assumptions that are different, and even less restrictive, than the usual ones (e.g. Lorentz-covariance). Additionally, in Quantum Mechanics there is a deep relation between indistinguishabilty and the geometry of the configuration space. This is clearly illustrated by Gibbs' paradox. Therefore, for many years efforts have been made in order to find a geometric proof of the connection between Spin and Statistics. Recently, various proposals have been put forward, in which an attempt is made to derive the Spin-Statistics connection from assumptions different from the ones used in the relativistic, quantum field theoretic proofs. Among these, there is the one due to Berry and Robbins (BR), based on the postulation of a certain single-valuedness condition, that has caused a renewed interest in the problem. In the present thesis, we consider the problem of indistinguishability in Quantum Mechanics from a geometric-algebraic point of view. An approach is developed to study configuration spaces Q having a finite fundamental group, that allows us to describe different geometric structures of Q in terms of spaces of functions on the universal cover of Q. In particular, it is shown that the space of complex continuous functions over the universal cover of Q admits a decomposition into C(Q)-submodules, labelled by the irreducible representations of the fundamental group of Q, that can be interpreted as the spaces of sections of certain flat vector bundles over Q. With this technique, various results pertaining to the problem of quantum indistinguishability are reproduced in a clear and systematic way. Our method is also used in order to give a global formulation of the BR construction. As a result of this analysis, it is found that the single-valuedness condition of BR is inconsistent. Additionally, a proposal aiming at establishing the Fermi-Bose alternative, within our approach, is made.
Resumo:
The Universidad Politécnica of Madrid (UPM) includes schools and faculties that were for engineering degrees, architecture and computer science, that are now in a quick EEES Bolonia Plan metamorphosis getting into degrees, masters and doctorate structures. They are focused towards action in machines, constructions, enterprises, that are subjected to machines, human and environment created risks. These are present in actions such as use loads, wind, snow, waves, flows, earthquakes, forces and effects in machines, vehicles behavior, chemical effects, and other environmental factors including effects of crops, cattle and beasts, forests, and varied essential economic and social disturbances. Emphasis is for authors in this session more about risks of natural origin, such as for hail, winds, snow or waves that are not exactly known a priori, but that are often considered with statistical expected distributions giving extreme values for convenient return periods. These distributions are known from measures in time, statistic of extremes and models about hazard scenarios and about responses of man made constructions or devices. In each engineering field theories were built about hazards scenarios and how to cover for important risks. Engineers must get that the systems they handle, such as vehicles, machines, firms or agro lands or forests, obtain production with enough safety for persons and with decent economic results in spite of risks. For that risks must be considered in planning, in realization and in operation, and safety margins must be taken but at a reasonable cost. That is a small level of risks will often remain, due to limitations in costs or because of due to strange hazards, and maybe they will be covered by insurance in cases such as in transport with cars, ships or aircrafts, in agro for hail, or for fire in houses or in forests. These and other decisions about quality, security for men or about business financial risks are sometimes considered with Decision Theories models, using often tools from Statistics or operational Research. The authors have done and are following field surveys about risk consideration in the careers in UPM, making deep analysis of curricula taking into account the new structures of degrees in the EEES Bolonia Plan, and they have considered the risk structures offered by diverse schools of Decision theories. That gives an aspect of the needs and uses, and recommendations about improving in the teaching about risk, that may include special subjects especially oriented for each career, school or faculty, so as to be recommended to be included into the curricula, including an elaboration and presentation format using a multi-criteria decision model.
Resumo:
In this work, we use large eddy simulations (LES) and Lagrangian tracking to study the influence of gravity on particle statistics in a fully developed turbulent upward/downward flow in a vertical channel and pipe at matched Kàrmàn number. Only drag and gravity are considered in the equation of motion for solid particles, which are assumed to have no influence on the flow field. Particle interactions with the wall are fully elastic. Our findings obtained from the particle statistics confirm that: (i) the gravity seems to modify both the quantitative and qualitative behavior of the particle distribution and statistics of the particle velocity in wall normal direction; (ii) however, only the quantitative behavior of velocity particle in streamwise direction and the root mean square of velocity components is modified; (iii) the statistics of fluid and particles coincide very well near the wall in channel and pipe flow with equal Kàrmàn number; (iv) pipe curvature seems to have quantitative and qualitative influence on the particle velocity and on the particle concentration in wall normal direction.