957 resultados para Multivariate statistical methods
Resumo:
Background: The inherent complexity of statistical methods and clinical phenomena compel researchers with diverse domains of expertise to work in interdisciplinary teams, where none of them have a complete knowledge in their counterpart's field. As a result, knowledge exchange may often be characterized by miscommunication leading to misinterpretation, ultimately resulting in errors in research and even clinical practice. Though communication has a central role in interdisciplinary collaboration and since miscommunication can have a negative impact on research processes, to the best of our knowledge, no study has yet explored how data analysis specialists and clinical researchers communicate over time. Methods/Principal Findings: We conducted qualitative analysis of encounters between clinical researchers and data analysis specialists (epidemiologist, clinical epidemiologist, and data mining specialist). These encounters were recorded and systematically analyzed using a grounded theory methodology for extraction of emerging themes, followed by data triangulation and analysis of negative cases for validation. A policy analysis was then performed using a system dynamics methodology looking for potential interventions to improve this process. Four major emerging themes were found. Definitions using lay language were frequently employed as a way to bridge the language gap between the specialties. Thought experiments presented a series of ""what if'' situations that helped clarify how the method or information from the other field would behave, if exposed to alternative situations, ultimately aiding in explaining their main objective. Metaphors and analogies were used to translate concepts across fields, from the unfamiliar to the familiar. Prolepsis was used to anticipate study outcomes, thus helping specialists understand the current context based on an understanding of their final goal. Conclusion/Significance: The communication between clinical researchers and data analysis specialists presents multiple challenges that can lead to errors.
Resumo:
Creation of cold dark matter (CCDM) can macroscopically be described by a negative pressure, and, therefore, the mechanism is capable to accelerate the Universe, without the need of an additional dark energy component. In this framework, we discuss the evolution of perturbations by considering a Neo-Newtonian approach where, unlike in the standard Newtonian cosmology, the fluid pressure is taken into account even in the homogeneous and isotropic background equations (Lima, Zanchin, and Brandenberger, MNRAS 291, L1, 1997). The evolution of the density contrast is calculated in the linear approximation and compared to the one predicted by the Lambda CDM model. The difference between the CCDM and Lambda CDM predictions at the perturbative level is quantified by using three different statistical methods, namely: a simple chi(2)-analysis in the relevant space parameter, a Bayesian statistical inference, and, finally, a Kolmogorov-Smirnov test. We find that under certain circumstances, the CCDM scenario analyzed here predicts an overall dynamics (including Hubble flow and matter fluctuation field) which fully recovers that of the traditional cosmic concordance model. Our basic conclusion is that such a reduction of the dark sector provides a viable alternative description to the accelerating Lambda CDM cosmology.
Resumo:
Aims. In this work, we describe the pipeline for the fast supervised classification of light curves observed by the CoRoT exoplanet CCDs. We present the classification results obtained for the first four measured fields, which represent a one-year in-orbit operation. Methods. The basis of the adopted supervised classification methodology has been described in detail in a previous paper, as is its application to the OGLE database. Here, we present the modifications of the algorithms and of the training set to optimize the performance when applied to the CoRoT data. Results. Classification results are presented for the observed fields IRa01, SRc01, LRc01, and LRa01 of the CoRoT mission. Statistics on the number of variables and the number of objects per class are given and typical light curves of high-probability candidates are shown. We also report on new stellar variability types discovered in the CoRoT data. The full classification results are publicly available.
Resumo:
Context. The detailed chemical abundances of extremely metal-poor (EMP) stars are key guides to understanding the early chemical evolution of the Galaxy. Most existing data, however, treat giant stars that may have experienced internal mixing later. Aims. We aim to compare the results for giants with new, accurate abundances for all observable elements in 18 EMP turno. stars. Methods. VLT/UVES spectra at R similar to 45 000 and S/N similar to 130 per pixel (lambda lambda 330-1000 nm) are analysed with OSMARCS model atmospheres and the TURBOSPECTRUM code to derive abundances for C, Mg, Si, Ca, Sc, Ti, Cr, Mn, Co, Ni, Zn, Sr, and Ba. Results. For Ca, Ni, Sr, and Ba, we find excellent consistency with our earlier sample of EMP giants, at all metallicities. However, our abundances of C, Sc, Ti, Cr, Mn and Co are similar to 0.2 dex larger than in giants of similar metallicity. Mg and Si abundances are similar to 0.2 dex lower (the giant [Mg/Fe] values are slightly revised), while Zn is again similar to 0.4 dex higher than in giants of similar [Fe/H] (6 stars only). Conclusions. For C, the dwarf/giant discrepancy could possibly have an astrophysical cause, but for the other elements it must arise from shortcomings in the analysis. Approximate computations of granulation (3D) effects yield smaller corrections for giants than for dwarfs, but suggest that this is an unlikely explanation, except perhaps for C, Cr, and Mn. NLTE computations for Na and Al provide consistent abundances between dwarfs and giants, unlike the LTE results, and would be highly desirable for the other discrepant elements as well. Meanwhile, we recommend using the giant abundances as reference data for Galactic chemical evolution models.
Resumo:
Catalytic ozonation has been recognized in the scientific community as an efficient technique, reaching elevated rates of recalcitrant organic material mineralization, even at the presence of scavenger species of hydroxyl free radicals. This study presents the most significant factors involving the leachate treatment stabilized by the municipal landfill of the city of Guaratingueta, State of Sao Paulo, Brazil, by using a catalytic ozonation activated by metallic ions Fe(3+), Zn(2+), Mn(2+), Ni(2+) and Cr(3+). The Taguchi L(16) orthogonal array and its associated statistical methods were also used in this study. Among the researched ions, the most notable catalysis was obtained with ferric ion, statistically significant in the reduction of COD with a confidence level of 99.5%.
Resumo:
In a sample of censored survival times, the presence of an immune proportion of individuals who are not subject to death, failure or relapse, may be indicated by a relatively high number of individuals with large censored survival times. In this paper the generalized log-gamma model is modified for the possibility that long-term survivors may be present in the data. The model attempts to separately estimate the effects of covariates on the surviving fraction, that is, the proportion of the population for which the event never occurs. The logistic function is used for the regression model of the surviving fraction. Inference for the model parameters is considered via maximum likelihood. Some influence methods, such as the local influence and total local influence of an individual are derived, analyzed and discussed. Finally, a data set from the medical area is analyzed under the log-gamma generalized mixture model. A residual analysis is performed in order to select an appropriate model.
Resumo:
In this paper, we present various diagnostic methods for polyhazard models. Polyhazard models are a flexible family for fitting lifetime data. Their main advantage over the single hazard models, such as the Weibull and the log-logistic models, is to include a large amount of nonmonotone hazard shapes, as bathtub and multimodal curves. Some influence methods, such as the local influence and total local influence of an individual are derived, analyzed and discussed. A discussion of the computation of the likelihood displacement as well as the normal curvature in the local influence method are presented. Finally, an example with real data is given for illustration.
Resumo:
In this study, 73 South American red wines (Vitis vinifera) from 5 varietals were classified based on sensory quality, retail price and antioxidant activity and characterised in relation to their phenolic composition. ORAC and DPPH assays were assessed to determine the antioxidant activity, and sensory analysis was conducted by seven professional tasters using the Wine Spirits Education Trust`s structured scales. The use of multivariate statistical techniques allowed the identification of wines with the best combination of sensory characteristics, price and antioxidant activity. The most favourable varieties were Malbec, Cabernet Sauvignon, and Syrah produced in Chile and Argentina. Conversely, Pinot Noir wines displayed the lowest sensory characteristics and antioxidant activity. These results suggest that the volatile compounds may be the main substances responsible for differentiating red wines on the basis of sensory evaluation. (C) 2011 Elsevier Ltd. All rights reserved.
Resumo:
Small area health statistics has assumed increasing importance as the focus of population and public health moves to a more individualised approach of smaller area populations. Small populations and low event occurrence produce difficulties in interpretation and require appropriate statistical methods, including for age adjustment. There are also statistical questions related to multiple comparisons. Privacy and confidentiality issues include the possibility of revealing information on individuals or health care providers by fine cross-tabulations. Interpretation of small area population differences in health status requires consideration of migrant and Indigenous composition, socio-economic status and rural-urban geography before assessment of the effects of physical environmental exposure and services and interventions. Burden of disease studies produce a single measure for morbidity and mortality - disability adjusted life year (DALY) - which is the sum of the years of life lost (YLL) from premature mortality and the years lived with disability (YLD) for particular diseases (or all conditions). Calculation of YLD requires estimates of disease incidence (and complications) and duration, and weighting by severity. These procedures often mean problematic assumptions, as does future discounting and age weighting of both YLL and YLD. Evaluation of the Victorian small area population disease burden study presents important cross-disciplinary challenges as it relies heavily on synthetic approaches of demography and economics rather than on the empirical methods of epidemiology. Both empirical and synthetic methods are used to compute small area mortality and morbidity, disease burden, and then attribution to risk factors. Readers need to examine the methodology and assumptions carefully before accepting the results.
Resumo:
Recent advances in the control of molecular engineering architectures have allowed unprecedented ability of molecular recognition in biosensing, with a promising impact for clinical diagnosis and environment control. The availability of large amounts of data from electrical, optical, or electrochemical measurements requires, however, sophisticated data treatment in order to optimize sensing performance. In this study, we show how an information visualization system based on projections, referred to as Projection Explorer (PEx), can be used to achieve high performance for biosensors made with nanostructured films containing immobilized antigens. As a proof of concept, various visualizations were obtained with impedance spectroscopy data from an array of sensors whose electrical response could be specific toward a given antibody (analyte) owing to molecular recognition processes. In addition to discussing the distinct methods for projection and normalization of the data, we demonstrate that an excellent distinction can be made between real samples tested positive for Chagas disease and Leishmaniasis, which could not be achieved with conventional statistical methods. Such high performance probably arose from the possibility of treating the data in the whole frequency range. Through a systematic analysis, it was inferred that Sammon`s mapping with standardization to normalize the data gives the best results, where distinction could be made of blood serum samples containing 10(-7) mg/mL of the antibody. The method inherent in PEx and the procedures for analyzing the impedance data are entirely generic and can be extended to optimize any type of sensor or biosensor.
Resumo:
Background: Urban air pollutants are associated with cardiovascular events. Traffic controllers are at high risk for pollution exposure during outdoor work shifts. Objective: The purpose of this study was to evaluate the relationship between air pollution and systemic blood pressure in traffic controllers during their work shifts. Methods: This cross-sectional study enrolled 19 male traffic controllers from Santo Andre city (Sao Paulo, Brazil) who were 30-60 years old and exposed to ambient air during outdoor work shifts. Systolic and diastolic blood pressure readings were measured every 15 min by an Ambulatory Arterial Blood Pressure Monitoring device. Hourly measurements (lags of 0-5 h) and the moving averages (2-5 h) of particulate matter (PM(10)), ozone (O(3)) ambient concentrations and the acquired daily minimum temperature and humidity means from the Sao Paulo State Environmental Agency were correlated with both systolic and diastolic blood pressures. Statistical methods included descriptive analysis and linear mixed effect models adjusted for temperature, humidity, work periods and time of day. Results: Interquartile increases of PM(10) (33 mu g/m(3)) and O(3) (49 mu g/m(3)) levels were associated with increases in all arterial pressure parameters, ranging from 1.06 to 2.53 mmHg. PM(10) concentration was associated with early effects (lag 0), mainly on systolic blood pressure. However, O(3) was weakly associated most consistently with diastolic blood pressure and with late cumulative effects. Conclusions: Santo Andre traffic controllers presented higher blood pressure readings while working their outdoor shifts during periods of exposure to ambient pollutant fluctuations. However, PM(10) and O(3) induced cardiovascular effects demonstrated different time courses and end-point behaviors and probably acted through different mechanisms. (C) 2011 Elsevier Inc. All rights reserved.
Resumo:
Objectives: We sought to compare long-term outcomes after coronary bypass surgery with and without an internal thoracic artery graft. Methods: We analyzed clinical outcomes over a median follow-up of 6.7 years among 3,087 patients who received coronary bypass surgery as participants in one of 8 clinical trials comparing surgical intervention with angioplasty. We used 2 statistical methods (covariate adjustment and propensity score matching) to adjust for the nonrandomized selection of internal thoracic artery grafts. Results: Internal thoracic artery grafting was associated with lower mortality, with hazard ratios of 0.77 (confidence interval, 0.62-0.97; P = .02) for covariate adjustment and 0.77 (confidence interval, 0.57-1.05; P = .10) for propensity score matching. The composite end point of death or myocardial infarction was reduced to a similar extent, with hazard ratios of 0.83 (confidence interval, 0.69-1.00; P = .05) for covariate adjustment to 0.78 (confidence interval, 0.61-1.00; P = .05) for propensity score matching. There was a trend toward less angina at 1 year, with odds ratios of 0.81 (confidence interval, 0.61-1.09; P = .16) in the covariate-adjusted model and 0.81 (confidence interval, 0.55-1.19; P = .28) in the propensity score-adjusted model. Conclusions: Use of an internal thoracic artery graft during coronary bypass surgery seems to improve long-term clinical outcomes. (J Thorac Cardiovasc Surg 2011; 142: 829-35)
Resumo:
The identification, modeling, and analysis of interactions between nodes of neural systems in the human brain have become the aim of interest of many studies in neuroscience. The complex neural network structure and its correlations with brain functions have played a role in all areas of neuroscience, including the comprehension of cognitive and emotional processing. Indeed, understanding how information is stored, retrieved, processed, and transmitted is one of the ultimate challenges in brain research. In this context, in functional neuroimaging, connectivity analysis is a major tool for the exploration and characterization of the information flow between specialized brain regions. In most functional magnetic resonance imaging (fMRI) studies, connectivity analysis is carried out by first selecting regions of interest (ROI) and then calculating an average BOLD time series (across the voxels in each cluster). Some studies have shown that the average may not be a good choice and have suggested, as an alternative, the use of principal component analysis (PCA) to extract the principal eigen-time series from the ROI(s). In this paper, we introduce a novel approach called cluster Granger analysis (CGA) to study connectivity between ROIs. The main aim of this method was to employ multiple eigen-time series in each ROI to avoid temporal information loss during identification of Granger causality. Such information loss is inherent in averaging (e.g., to yield a single ""representative"" time series per ROI). This, in turn, may lead to a lack of power in detecting connections. The proposed approach is based on multivariate statistical analysis and integrates PCA and partial canonical correlation in a framework of Granger causality for clusters (sets) of time series. We also describe an algorithm for statistical significance testing based on bootstrapping. By using Monte Carlo simulations, we show that the proposed approach outperforms conventional Granger causality analysis (i.e., using representative time series extracted by signal averaging or first principal components estimation from ROIs). The usefulness of the CGA approach in real fMRI data is illustrated in an experiment using human faces expressing emotions. With this data set, the proposed approach suggested the presence of significantly more connections between the ROIs than were detected using a single representative time series in each ROI. (c) 2010 Elsevier Inc. All rights reserved.
Resumo:
OBJECTIVE. The objective of our study was to describe the T1 and T2 signal intensity characteristics of papillary renal cell carcinoma (RCC) and clear cell RCC with pathologic correlation. MATERIALS AND METHODS. Of 539 RCCs, 49 tumors (21 papillary RCCs and 28 clear cell RCCs) in 45 patients were examined with MRI. Two radiologists retrospectively and independently assessed each tumor`s T1 and T2 signal intensity qualitatively and quantitatively (i.e., the signal intensity [SI] ratio [tumor SI/renal cortex SI]). Of the 49 tumors, 37 (76%) were assessed for pathology features including tumor architecture and the presence of hemosiderin, ferritin, necrosis, and fibrosis. MRI findings and pathology features were correlated. Statistical methods included summary statistics and Wilcoxon`s rank sum test for signal intensity, contingency tables for assessing reader agreement, concordance rate between the two readers with 95% CIs, and Fisher`s exact test for independence, all stratified by RCC type. RESULTS. Papillary RCCs and clear cell RCCs had a similar appearance and signal intensity ratio on T1-weighted images. On T2-weighted images, most papillary RCCs were hypointense (reader 1, 13/21; reader 2, 14/21), with an average mean signal intensity ratio for both readers of 0.67 +/- 0.2, and none was hyperintense, whereas most clear cell RCCs were hyperintense (reader 1, 21/28; reader 2, 17/28), with an average mean signal intensity ratio for both readers of 1.41 +/- 0.4 (p < 0.05). A tumor T2 signal intensity ratio of <= 0.66 had a specificity of 100% and sensitivity of 54% for papillary RCC. Most T2 hypointense tumors exhibited predominant papillary architecture; most T2 hyperintense tumors had a predominant nested architecture (p < 0.05). CONCLUSION. On T2-weighted images, most papillary RCCs are hypointense and clear cell RCCs, hyperintense. The T2 hypointense appearance of papillary RCCs correlated with a predominant papillary architecture at pathology.
Resumo:
Objective To assess MHC I and II expressions in muscle fibres of juvenile dermatomyositis (JDM) and compare with the expression in polymyositis (PM), dermatomyositis (DM) and dystrophy. Patients and methods Forty-eight JDM patients and 17 controls (8 PM, 5 DM and 4 dystrophy) were studied. The mean age at disease onset was 7.1 +/- 3.0 years and the mean duration of weakness before biopsy was 9.4 +/- 12.9 months. Routine histochemistry and immunohistochemistry (StreptABComplex/HRP) for MHC I and II (Dakopatts) were performed on serial frozen muscle sections in all patients. Mann-Whitney, Kruskal Wallis, chi-square and Fisher`s exact statistical methods were used. Results MHC I expression was positive in 47 (97.9%) JDM cases. This expression was observed independent of time of disease corticotherapy previous to muscle biopsy and to the grading of inflammation observed in clinical, laboratorial and histological parameters. The expression of MHC I was similar on JDM, PM and DM, and lower in dystrophy. On the other hand, MHC II expression was positive in just 28.2% of JDM cases was correlated to histological features as inflammatory infiltrate, increased connective tissue and VAS for global degree of abnormality (p < 0.05). MCH II expression was similar in DM/PM and lower in JDM and dystrophy, and it was based on the frequency of positive staining rather than to the degree of the MCH II expression. Conclusions MHC I expression in muscle fibres is a premature and late marker of JDM patient independent to corticotherapy, and MHC II expression was lower in JDM than in PM and DM.