924 resultados para spatial data analysis
Resumo:
Riparian zones are dynamic, transitional ecosystems between aquatic and terrestrial ecosystems with well defined vegetation and soil characteristics. Development of an all-encompassing definition for riparian ecotones, because of their high variability, is challenging. However, there are two primary factors that all riparian ecotones are dependent on: the watercourse and its associated floodplain. Previous approaches to riparian boundary delineation have utilized fixed width buffers, but this methodology has proven to be inadequate as it only takes the watercourse into consideration and ignores critical geomorphology, associated vegetation and soil characteristics. Our approach offers advantages over other previously used methods by utilizing: the geospatial modeling capabilities of ArcMap GIS; a better sampling technique along the water course that can distinguish the 50-year flood plain, which is the optimal hydrologic descriptor of riparian ecotones; the Soil Survey Database (SSURGO) and National Wetland Inventory (NWI) databases to distinguish contiguous areas beyond the 50-year plain; and land use/cover characteristics associated with the delineated riparian zones. The model utilizes spatial data readily available from Federal and State agencies and geospatial clearinghouses. An accuracy assessment was performed to assess the impact of varying the 50-year flood height, changing the DEM spatial resolution (1, 3, 5 and 10m), and positional inaccuracies with the National Hydrography Dataset (NHD) streams layer on the boundary placement of the delineated variable width riparian ecotones area. The result of this study is a robust and automated GIS based model attached to ESRI ArcMap software to delineate and classify variable-width riparian ecotones.
Resumo:
Exposimeters are increasingly applied in bioelectromagnetic research to determine personal radiofrequency electromagnetic field (RF-EMF) exposure. The main advantages of exposimeter measurements are their convenient handling for study participants and the large amount of personal exposure data, which can be obtained for several RF-EMF sources. However, the large proportion of measurements below the detection limit is a challenge for data analysis. With the robust ROS (regression on order statistics) method, summary statistics can be calculated by fitting an assumed distribution to the observed data. We used a preliminary sample of 109 weekly exposimeter measurements from the QUALIFEX study to compare summary statistics computed by robust ROS with a naïve approach, where values below the detection limit were replaced by the value of the detection limit. For the total RF-EMF exposure, differences between the naïve approach and the robust ROS were moderate for the 90th percentile and the arithmetic mean. However, exposure contributions from minor RF-EMF sources were considerably overestimated with the naïve approach. This results in an underestimation of the exposure range in the population, which may bias the evaluation of potential exposure-response associations. We conclude from our analyses that summary statistics of exposimeter data calculated by robust ROS are more reliable and more informative than estimates based on a naïve approach. Nevertheless, estimates of source-specific medians or even lower percentiles depend on the assumed data distribution and should be considered with caution.
Resumo:
Turrialba is one of the largest and most active stratovolcanoes in the Central Cordillera of Costa Rica and an excellent target for validation of satellite data using ground based measurements due to its high elevation, relative ease of access, and persistent elevated SO2 degassing. The Ozone Monitoring Instrument (OMI) aboard the Aura satellite makes daily global observations of atmospheric trace gases and it is used in this investigation to obtain volcanic SO2 retrievals in the Turrialba volcanic plume. We present and evaluate the relative accuracy of two OMI SO2 data analysis procedures, the automatic Band Residual Index (BRI) technique and the manual Normalized Cloud-mass (NCM) method. We find a linear correlation and good quantitative agreement between SO2 burdens derived from the BRI and NCM techniques, with an improved correlation when wet season data are excluded. We also present the first comparisons between volcanic SO2 emission rates obtained from ground-based mini-DOAS measurements at Turrialba and three new OMI SO2 data analysis techniques: the MODIS smoke estimation, OMI SO2 lifetime, and OMI SO2 transect techniques. A robust validation of OMI SO2 retrievals was made, with both qualitative and quantitative agreements under specific atmospheric conditions, proving the utility of satellite measurements for estimating accurate SO2 emission rates and monitoring passively degassing volcanoes.
DIMENSION REDUCTION FOR POWER SYSTEM MODELING USING PCA METHODS CONSIDERING INCOMPLETE DATA READINGS
Resumo:
Principal Component Analysis (PCA) is a popular method for dimension reduction that can be used in many fields including data compression, image processing, exploratory data analysis, etc. However, traditional PCA method has several drawbacks, since the traditional PCA method is not efficient for dealing with high dimensional data and cannot be effectively applied to compute accurate enough principal components when handling relatively large portion of missing data. In this report, we propose to use EM-PCA method for dimension reduction of power system measurement with missing data, and provide a comparative study of traditional PCA and EM-PCA methods. Our extensive experimental results show that EM-PCA method is more effective and more accurate for dimension reduction of power system measurement data than traditional PCA method when dealing with large portion of missing data set.
Resumo:
Dr. Rossi discusses the common errors that are made when fitting statistical models to data. Focuses on the planning, data analysis, and interpretation phases of a statistical analysis, and highlights the errors that are commonly made by researchers of these phases. The implications of these commonly made errors are discussed along with a discussion of the methods that can be used to prevent these errors from occurring. A prescription for carrying out a correct statistical analysis will be discussed.
Resumo:
This study presents a proxy-based, quantitative reconstruction of cold-season (mean October to May, TOct–May) air temperatures covering nearly the entire last millennium (AD 1060–2003, some hiatuses). The reconstruction was based on subfossil chrysophyte stomatocyst remains in the varved sediments of high-Alpine Lake Silvaplana, eastern Swiss Alps (46°27’N, 9°48′W, 1791 m a.s.l.). Previous studies have demonstrated the reliability of this proxy by comparison to meteorological data. Cold-season air temperatures could therefore be reconstructed quantitatively, at a high resolution (5-yr) and with high chronological accuracy. Spatial correlation analysis suggests that the reconstruction reflects cold season climate variability over the high- Alpine region and substantial parts of central and western Europe. Cold-season temperatures were characterized by a relatively stable first part of the millennium until AD 1440 (2σ of 5-yr mean values = 0.7 °C) and highly variable TOct–May after that (AD 1440–1900, 2σ of 5-yr mean values = 1.3 °C). Recent decades (AD, 1991-present) were unusually warm in the context of the last millennium (exceeding the 2σ-range of the mean decadal TOct–May) but this warmth was not unprecedented. The coolest decades occurred from AD 1510–1520 and AD 1880–1890. The timing of extremely warm and cold decades is generally in good agreement with documentary data representing Switzerland and central European lowlands. The transition from relatively stable to highly variable TOct–May coincided with large changes in atmospheric circulation patterns in the North Atlantic region. Comparison of reconstructed cold season temperatures to the North Atlantic Oscillation index (NAO) during the past 1000 years showed that the relatively stable and warm conditions at the study site until AD 1440 coincided with a persistent positive mode of the NAO. We propose that the transition to large TOct–May variability around AD 1440 was linked to the subsequent absence of this persistent zonal flow pattern, which would allow other climatic drivers to gain importance in the study area. From AD 1440–1900, the similarity of reconstructed TOct–May to reconstructed air pressure in the Siberian High suggests a relatively strong influence of continental anticyclonic systems on Alpine cold season climate parameters during periods when westerly airflow was subdued. A more continental type of atmospheric circulation thus seems to be characteristic for the Little Ice Age in Europe. Comparison of Toct–May to summer temperature reconstructions from the same study site shows that, as expected, summer and cold season temperature trends and variability differed completely throughout nearly the entire last 1000 years. Since AD 1980, however, summer and cold season temperatures show a simultaneous, strong increase, which is unprecedented in the context of the last millennium. We suggest that the most likely explanation for this recent trend is anthropogenic greenhouse gas (GHG) forcing.
Resumo:
Frequency-transformed EEG resting data has been widely used to describe normal and abnormal brain functional states as function of the spectral power in different frequency bands. This has yielded a series of clinically relevant findings. However, by transforming the EEG into the frequency domain, the initially excellent time resolution of time-domain EEG is lost. The topographic time-frequency decomposition is a novel computerized EEG analysis method that combines previously available techniques from time-domain spatial EEG analysis and time-frequency decomposition of single-channel time series. It yields a new, physiologically and statistically plausible topographic time-frequency representation of human multichannel EEG. The original EEG is accounted by the coefficients of a large set of user defined EEG like time-series, which are optimized for maximal spatial smoothness and minimal norm. These coefficients are then reduced to a small number of model scalp field configurations, which vary in intensity as a function of time and frequency. The result is thus a small number of EEG field configurations, each with a corresponding time-frequency (Wigner) plot. The method has several advantages: It does not assume that the data is composed of orthogonal elements, it does not assume stationarity, it produces topographical maps and it allows to include user-defined, specific EEG elements, such as spike and wave patterns. After a formal introduction of the method, several examples are given, which include artificial data and multichannel EEG during different physiological and pathological conditions.
Resumo:
Cluster randomized trials (CRTs) use as the unit of randomization clusters, which are usually defined as a collection of individuals sharing some common characteristics. Common examples of clusters include entire dental practices, hospitals, schools, school classes, villages, and towns. Additionally, several measurements (repeated measurements) taken on the same individual at different time points are also considered to be clusters. In dentistry, CRTs are applicable as patients may be treated as clusters containing several individual teeth. CRTs require certain methodological procedures during sample calculation, randomization, data analysis, and reporting, which are often ignored in dental research publications. In general, due to similarity of the observations within clusters, each individual within a cluster provides less information compared with an individual in a non-clustered trial. Therefore, clustered designs require larger sample sizes compared with non-clustered randomized designs, and special statistical analyses that account for the fact that observations within clusters are correlated. It is the purpose of this article to highlight with relevant examples the important methodological characteristics of cluster randomized designs as they may be applied in orthodontics and to explain the problems that may arise if clustered observations are erroneously treated and analysed as independent (non-clustered).
Resumo:
This paper presents an overview of the Mobile Data Challenge (MDC), a large-scale research initiative aimed at generating innovations around smartphone-based research, as well as community-based evaluation of mobile data analysis methodologies. First, we review the Lausanne Data Collection Campaign (LDCC), an initiative to collect unique longitudinal smartphone dataset for the MDC. Then, we introduce the Open and Dedicated Tracks of the MDC, describe the specific datasets used in each of them, discuss the key design and implementation aspects introduced in order to generate privacy-preserving and scientifically relevant mobile data resources for wider use by the research community, and summarize the main research trends found among the 100+ challenge submissions. We finalize by discussing the main lessons learned from the participation of several hundred researchers worldwide in the MDC Tracks.
Resumo:
Brain tumor is one of the most aggressive types of cancer in humans, with an estimated median survival time of 12 months and only 4% of the patients surviving more than 5 years after disease diagnosis. Until recently, brain tumor prognosis has been based only on clinical information such as tumor grade and patient age, but there are reports indicating that molecular profiling of gliomas can reveal subgroups of patients with distinct survival rates. We hypothesize that coupling molecular profiling of brain tumors with clinical information might improve predictions of patient survival time and, consequently, better guide future treatment decisions. In order to evaluate this hypothesis, the general goal of this research is to build models for survival prediction of glioma patients using DNA molecular profiles (U133 Affymetrix gene expression microarrays) along with clinical information. First, a predictive Random Forest model is built for binary outcomes (i.e. short vs. long-term survival) and a small subset of genes whose expression values can be used to predict survival time is selected. Following, a new statistical methodology is developed for predicting time-to-death outcomes using Bayesian ensemble trees. Due to a large heterogeneity observed within prognostic classes obtained by the Random Forest model, prediction can be improved by relating time-to-death with gene expression profile directly. We propose a Bayesian ensemble model for survival prediction which is appropriate for high-dimensional data such as gene expression data. Our approach is based on the ensemble "sum-of-trees" model which is flexible to incorporate additive and interaction effects between genes. We specify a fully Bayesian hierarchical approach and illustrate our methodology for the CPH, Weibull, and AFT survival models. We overcome the lack of conjugacy using a latent variable formulation to model the covariate effects which decreases computation time for model fitting. Also, our proposed models provides a model-free way to select important predictive prognostic markers based on controlling false discovery rates. We compare the performance of our methods with baseline reference survival methods and apply our methodology to an unpublished data set of brain tumor survival times and gene expression data, selecting genes potentially related to the development of the disease under study. A closing discussion compares results obtained by Random Forest and Bayesian ensemble methods under the biological/clinical perspectives and highlights the statistical advantages and disadvantages of the new methodology in the context of DNA microarray data analysis.
Resumo:
In the present study the challenge of analyzing complex micro X-ray diffraction (microXRD) patterns from cement–clay interfaces has been addressed. In order to extract the maximum information concerning both the spatial distribution and the crystal structure type associated with each of the many diffracting grains in heterogeneous, polycrystalline samples, an approach has been developed in which microXRD was applied to thin sections which were rotated in the X-ray beam. The data analysis, performed on microXRD patterns collected from a filled vein of a cement–clay interface from the natural analogue in Maqarin (Jordan), and a sample from a two-year-old altered interface between cement and argillaceous rock, demonstrate the potential of this method.
Resumo:
Water stable isotope ratios and net snow accumulation in ice cores are commonly interpreted as temperature or precipitation proxies. However, only in a few cases has a direct calibration with instrumental data been attempted. In this study we took advantage of the dense network of observations in the European Alpine region to rigorously test the relationship of the annual and seasonal resolved proxy data from two highly resolved ice cores with local temperature and precipitation. We focused on the time period 1961–2001 with the highest amount and quality of meteorological data and the minimal uncertainty in ice core dating (±1 year). The two ice cores were retrieved from the Fiescherhorn glacier (northern Alps, 3900 m a.s.l.), and Grenzgletscher (southern Alps, 4200 m a.s.l.). A parallel core from the Fiescherhorn glacier allowed assessing the reproducibility of the ice core proxy data. Due to the orographic barrier, the two flanks of the Alpine chain are affected by distinct patterns of precipitation. The different location of the two glaciers therefore offers a unique opportunity to test whether such a specific setting is reflected in the proxy data. On a seasonal scale a high fraction of δ18O variability was explained by the seasonal cycle of temperature (~60% for the ice cores, ~70% for the nearby stations of the Global Network of Isotopes in Precipitation – GNIP). When the seasonality is removed, the correlations decrease for all sites, indicating that factors other than temperature such as changing moisture sources and/or precipitation regimes affect the isotopic signal on this timescale. Post-depositional phenomena may additionally modify the ice core data. On an annual scale, the δ18O/temperature relationship was significant at the Fiescherhorn, whereas for Grenzgletscher this was the case only when weighting the temperature with precipitation. In both cases the fraction of interannual temperature variability explained was ~20%, comparable to the values obtained from the GNIP stations data. Consistently with previous studies, we found an altitude effect for the δ18O of −0.17‰/100 m for an extended elevation range combining data of the two ice core sites and four GNIP stations. Significant correlations between net accumulation and precipitation were observed for Grenzgletscher during the entire period of investigation, whereas for Fiescherhorn this was the case only for the less recent period (1961–1977). Local phenomena, probably related to wind, seem to partly disturb the Fiescherhorn accumulation record. Spatial correlation analysis shows the two glaciers to be influenced by different precipitation regimes, with the Grenzgletscher reflecting the characteristic precipitation regime south of the Alps and the Fiescherhorn accumulation showing a pattern more closely linked to northern Alpine stations.
Resumo:
OBJECTIVES To identify the timing of significant arch dimensional increases during orthodontic alignment involving round and rectangular nickel-titanium (NiTi) wires and rectangular stainless steel (SS). A secondary aim was to compare the timing of changes occurring with conventional and self-ligating fixed appliance systems. METHODS In this non-primary publication, additional data from a multicenter randomised trial initially involving 96 patients, aged 16 years and above, were analysed. The main pre-specified outcome measures were the magnitude and timing of maxillary intercanine, interpremolar, and intermolar dimensions. Each participant underwent alignment with a standard Damon (Ormco, Orange, CA) wire sequence for a minimum of 34 weeks. Blinding of clinicians and patients was not possible; however, outcome assessors and data analysts were kept blind to the appliance type during data analysis. RESULTS Complete data were obtained from 71 subjects. Significant arch dimensional changes were observed relatively early in treatment. In particular, changes in maxillary inter-first and second premolar dimensions occurred after alignment with an 0.014in. NiTi wire (P<0.05). No statistical differences in transverse dimensions were found between rectangular NiTi and working SS wires for each transverse dimension (P>0.05). Bracket type had no significant effect on the timing of the transverse dimensional changes. CONCLUSIONS Arch dimensional changes were found to occur relatively early in treatment, irrespective of the appliance type. Nickel-titanium wires may have a more profound effect on transverse dimensions than previously believed. CLINICAL SIGNIFICANCE On the basis of this research orthodontic expansion may occur relatively early in treatment. Nickel-titanium wires may have a more profound effect on transverse dimensions than previously believed.