946 resultados para LINEAR-ANALYSIS


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Objective: In this secondary data analysis, three statistical methodologies were implemented to handle cases with missing data in a motivational interviewing and feedback study. The aim was to evaluate the impact that these methodologies have on the data analysis. ^ Methods: We first evaluated whether the assumption of missing completely at random held for this study. We then proceeded to conduct a secondary data analysis using a mixed linear model to handle missing data with three methodologies (a) complete case analysis, (b) multiple imputation with explicit model containing outcome variables, time, and the interaction of time and treatment, and (c) multiple imputation with explicit model containing outcome variables, time, the interaction of time and treatment, and additional covariates (e.g., age, gender, smoke, years in school, marital status, housing, race/ethnicity, and if participants play on athletic team). Several comparisons were conducted including the following ones: 1) the motivation interviewing with feedback group (MIF) vs. the assessment only group (AO), the motivation interviewing group (MIO) vs. AO, and the intervention of the feedback only group (FBO) vs. AO, 2) MIF vs. FBO, and 3) MIF vs. MIO.^ Results: We first evaluated the patterns of missingness in this study, which indicated that about 13% of participants showed monotone missing patterns, and about 3.5% showed non-monotone missing patterns. Then we evaluated the assumption of missing completely at random by Little's missing completely at random (MCAR) test, in which the Chi-Square test statistic was 167.8 with 125 degrees of freedom, and its associated p-value was p=0.006, which indicated that the data could not be assumed to be missing completely at random. After that, we compared if the three different strategies reached the same results. For the comparison between MIF and AO as well as the comparison between MIF and FBO, only the multiple imputation with additional covariates by uncongenial and congenial models reached different results. For the comparison between MIF and MIO, all the methodologies for handling missing values obtained different results. ^ Discussions: The study indicated that, first, missingness was crucial in this study. Second, to understand the assumptions of the model was important since we could not identify if the data were missing at random or missing not at random. Therefore, future researches should focus on exploring more sensitivity analyses under missing not at random assumption.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Genome-wide association studies (GWAS) have successfully identified several genetic loci associated with inherited predisposition to primary biliary cirrhosis (PBC), the most common autoimmune disease of the liver. Pathway-based tests constitute a novel paradigm for GWAS analysis. By evaluating genetic variation across a biological pathway (gene set), these tests have the potential to determine the collective impact of variants with subtle effects that are individually too weak to be detected in traditional single variant GWAS analysis. To identify biological pathways associated with the risk of development of PBC, GWAS of PBC from Italy (449 cases and 940 controls) and Canada (530 cases and 398 controls) were independently analyzed. The linear combination test (LCT), a recently developed pathway-level statistical method was used for this analysis. For additional validation, pathways that were replicated at the P <0.05 level of significance in both GWAS on LCT analysis were also tested for association with PBC in each dataset using two complementary GWAS pathway approaches. The complementary approaches included a modification of the gene set enrichment analysis algorithm (i-GSEA4GWAS) and Fisher's exact test for pathway enrichment ratios. Twenty-five pathways were associated with PBC risk on LCT analysis in the Italian dataset at P<0.05, of which eight had an FDR<0.25. The top pathway in the Italian dataset was the TNF/stress related signaling pathway (p=7.38×10 -4, FDR=0.18). Twenty-six pathways were associated with PBC at the P<0.05 level using the LCT in the Canadian dataset with the regulation and function of ChREBP in liver pathway (p=5.68×10-4, FDR=0.285) emerging as the most significant pathway. Two pathways, phosphatidylinositol signaling system (Italian: p=0.016, FDR=0.436; Canadian: p=0.034, FDR=0.693) and hedgehog signaling (Italian: p=0.044, FDR=0.636; Canadian: p=0.041, FDR=0.693), were replicated at LCT P<0.05 in both datasets. Statistically significant association of both pathways with PBC genetic susceptibility was confirmed in the Italian dataset on i-GSEA4GWAS. Results for the phosphatidylinositol signaling system were also significant in both datasets on applying Fisher's exact test for pathway enrichment ratios. This study identified a combination of known and novel pathway-level associations with PBC risk. If functionally validated, the findings may yield fresh insights into the etiology of this complex autoimmune disease with possible preventive and therapeutic application.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Complex diseases, such as cancer, are caused by various genetic and environmental factors, and their interactions. Joint analysis of these factors and their interactions would increase the power to detect risk factors but is statistically. Bayesian generalized linear models using student-t prior distributions on coefficients, is a novel method to simultaneously analyze genetic factors, environmental factors, and interactions. I performed simulation studies using three different disease models and demonstrated that the variable selection performance of Bayesian generalized linear models is comparable to that of Bayesian stochastic search variable selection, an improved method for variable selection when compared to standard methods. I further evaluated the variable selection performance of Bayesian generalized linear models using different numbers of candidate covariates and different sample sizes, and provided a guideline for required sample size to achieve a high power of variable selection using Bayesian generalize linear models, considering different scales of number of candidate covariates. ^ Polymorphisms in folate metabolism genes and nutritional factors have been previously associated with lung cancer risk. In this study, I simultaneously analyzed 115 tag SNPs in folate metabolism genes, 14 nutritional factors, and all possible genetic-nutritional interactions from 1239 lung cancer cases and 1692 controls using Bayesian generalized linear models stratified by never, former, and current smoking status. SNPs in MTRR were significantly associated with lung cancer risk across never, former, and current smokers. In never smokers, three SNPs in TYMS and three gene-nutrient interactions, including an interaction between SHMT1 and vitamin B12, an interaction between MTRR and total fat intake, and an interaction between MTR and alcohol use, were also identified as associated with lung cancer risk. These lung cancer risk factors are worthy of further investigation.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Schizophrenia (SZ) is a complex disorder with high heritability and variable phenotypes that has limited success in finding causal genes associated with the disease development. Pathway-based analysis is an effective approach in investigating the molecular mechanism of susceptible genes associated with complex diseases. The etiology of complex diseases could be a network of genetic factors and within the genes, interaction may occur. In this work we argue that some genes might be of small effect that by itself are neither sufficient nor necessary to cause the disease however, their effect may induce slight changes to the gene expression or affect the protein function, therefore, analyzing the gene-gene interaction mechanism within the disease pathway would play crucial role in dissecting the genetic architecture of complex diseases, making the pathway-based analysis a complementary approach to GWAS technique. ^ In this study, we implemented three novel linkage disequilibrium based statistics, the linear combination, the quadratic, and the decorrelation test statistics, to investigate the interaction between linked and unlinked genes in two independent case-control GWAS datasets for SZ including participants of European (EA) and African (AA) ancestries. The EA population included 1,173 cases and 1,378 controls with 729,454 genotyped SNPs, while the AA population included 219 cases and 288 controls with 845,814 genotyped SNPs. We identified 17,186 interacting gene-sets at significant level in EA dataset, and 12,691 gene-sets in AA dataset using the gene-gene interaction method. We also identified 18,846 genes in EA dataset and 19,431 genes in AA dataset that were in the disease pathways. However, few genes were reported of significant association to SZ. ^ Our research determined the pathways characteristics for schizophrenia through the gene-gene interaction and gene-pathway based approaches. Our findings suggest insightful inferences of our methods in studying the molecular mechanisms of common complex diseases.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Quantitative real-time polymerase chain reaction (qPCR) is a sensitive gene quantitation method that has been widely used in the biological and biomedical fields. The currently used methods for PCR data analysis, including the threshold cycle (CT) method, linear and non-linear model fitting methods, all require subtracting background fluorescence. However, the removal of background fluorescence is usually inaccurate, and therefore can distort results. Here, we propose a new method, the taking-difference linear regression method, to overcome this limitation. Briefly, for each two consecutive PCR cycles, we subtracted the fluorescence in the former cycle from that in the later cycle, transforming the n cycle raw data into n-1 cycle data. Then linear regression was applied to the natural logarithm of the transformed data. Finally, amplification efficiencies and the initial DNA molecular numbers were calculated for each PCR run. To evaluate this new method, we compared it in terms of accuracy and precision with the original linear regression method with three background corrections, being the mean of cycles 1-3, the mean of cycles 3-7, and the minimum. Three criteria, including threshold identification, max R2, and max slope, were employed to search for target data points. Considering that PCR data are time series data, we also applied linear mixed models. Collectively, when the threshold identification criterion was applied and when the linear mixed model was adopted, the taking-difference linear regression method was superior as it gave an accurate estimation of initial DNA amount and a reasonable estimation of PCR amplification efficiencies. When the criteria of max R2 and max slope were used, the original linear regression method gave an accurate estimation of initial DNA amount. Overall, the taking-difference linear regression method avoids the error in subtracting an unknown background and thus it is theoretically more accurate and reliable. This method is easy to perform and the taking-difference strategy can be extended to all current methods for qPCR data analysis.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With most clinical trials, missing data presents a statistical problem in evaluating a treatment's efficacy. There are many methods commonly used to assess missing data; however, these methods leave room for bias to enter the study. This thesis was a secondary analysis on data taken from TIME, a phase 2 randomized clinical trial conducted to evaluate the safety and effect of the administration timing of bone marrow mononuclear cells (BMMNC) for subjects with acute myocardial infarction (AMI).^ We evaluated the effect of missing data by comparing the variance inflation factor (VIF) of the effect of therapy between all subjects and only subjects with complete data. Through the general linear model, an unbiased solution was made for the VIF of the treatment's efficacy using the weighted least squares method to incorporate missing data. Two groups were identified from the TIME data: 1) all subjects and 2) subjects with complete data (baseline and follow-up measurements). After the general solution was found for the VIF, it was migrated Excel 2010 to evaluate data from TIME. The resulting numerical value from the two groups was compared to assess the effect of missing data.^ The VIF values from the TIME study were considerably less in the group with missing data. By design, we varied the correlation factor in order to evaluate the VIFs of both groups. As the correlation factor increased, the VIF values increased at a faster rate in the group with only complete data. Furthermore, while varying the correlation factor, the number of subjects with missing data was also varied to see how missing data affects the VIF. When subjects with only baseline data was increased, we saw a significant rate increase in VIF values in the group with only complete data while the group with missing data saw a steady and consistent increase in the VIF. The same was seen when we varied the group with follow-up only data. This essentially showed that the VIFs steadily increased when missing data is not ignored. When missing data is ignored as with our comparison group, the VIF values sharply increase as correlation increases.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Hierarchical linear growth model (HLGM), as a flexible and powerful analytic method, has played an increased important role in psychology, public health and medical sciences in recent decades. Mostly, researchers who conduct HLGM are interested in the treatment effect on individual trajectories, which can be indicated by the cross-level interaction effects. However, the statistical hypothesis test for the effect of cross-level interaction in HLGM only show us whether there is a significant group difference in the average rate of change, rate of acceleration or higher polynomial effect; it fails to convey information about the magnitude of the difference between the group trajectories at specific time point. Thus, reporting and interpreting effect sizes have been increased emphases in HLGM in recent years, due to the limitations and increased criticisms for statistical hypothesis testing. However, most researchers fail to report these model-implied effect sizes for group trajectories comparison and their corresponding confidence intervals in HLGM analysis, since lack of appropriate and standard functions to estimate effect sizes associated with the model-implied difference between grouping trajectories in HLGM, and also lack of computing packages in the popular statistical software to automatically calculate them. ^ The present project is the first to establish the appropriate computing functions to assess the standard difference between grouping trajectories in HLGM. We proposed the two functions to estimate effect sizes on model-based grouping trajectories difference at specific time, we also suggested the robust effect sizes to reduce the bias of estimated effect sizes. Then, we applied the proposed functions to estimate the population effect sizes (d ) and robust effect sizes (du) on the cross-level interaction in HLGM by using the three simulated datasets, and also we compared the three methods of constructing confidence intervals around d and du recommended the best one for application. At the end, we constructed 95% confidence intervals with the suitable method for the effect sizes what we obtained with the three simulated datasets. ^ The effect sizes between grouping trajectories for the three simulated longitudinal datasets indicated that even though the statistical hypothesis test shows no significant difference between grouping trajectories, effect sizes between these grouping trajectories can still be large at some time points. Therefore, effect sizes between grouping trajectories in HLGM analysis provide us additional and meaningful information to assess group effect on individual trajectories. In addition, we also compared the three methods to construct 95% confident intervals around corresponding effect sizes in this project, which handled with the uncertainty of effect sizes to population parameter. We suggested the noncentral t-distribution based method when the assumptions held, and the bootstrap bias-corrected and accelerated method when the assumptions are not met.^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A 560-meter-thick sequence of Cenomanian through Pleistocene sediments cored at DSDP Site 462 in the Nauru Basin overlies a 500-meter-thick complex unit of altered basalt flows, diabase sills, and thin intercalated volcaniclastic sediments. The Upper Cretaceous and Cenozoic sediments contain a high proportion of calcareous fossils, although the site has apparently been below the calcite compensation depth (CCD) from the late Mesozoic to the Pleistocene. This fact and the contemporaneous fluctuations of the calcite and opal accumulation rates suggest an irregular influx of displaced pelagic sediments from the shallow margins of the basin to its center, resulting in unusually high overall sedimentation rates for such a deep (5190 m) site. Shallow-water benthic fossils and planktonic foraminifers both occur as reworked materials, but usually are not found in the same intervals of the sediment section. We interpret this as recording separate erosional interludes in the shallow-water and intermediate-water regimes. Lower and upper Cenozoic hiatuses also are believed to have resulted from mid-water events. High accumulation rates of volcanogenic material during Santonian time suggest a corresponding significant volcanic episode. The coincidence of increased carbonate accumulation rates during the Campanian and displacement of shallow-water fossils during the late Campanian-early Maestrichtian with the volcanic event implies that this early event resulted in formation of the island chains around the Nauru Basin, which then served as platforms for initial carbonate deposition.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Several studies indicate that the 2011 Tohoku-Oki earthquake (Mw 9.0) off the Pacific coast of Japan has induced slip to the trench and triggered landslides in the Japan Trench. In order to better understand these processes, detailed mapping and shallow-coring landslides at the trench as well as Integrated Ocean Drilling Program (IODP) deep drilling to recover the plate boundary décollement (Japan Trench Fast Earthquake Drilling Project, JFAST) have been conducted. In this study we report sediment core data from the rapid response R/V SONNE cruise (SO219A) to the Japan Trench, evidencing a Mass Transport Deposit (MTD) in the uppermost section later drilled at this JFAST-site during IODP Expedition 343. A 8.7 m long gravity core (GeoB16423-1) recovered from ~7,000 m water depth reveals a 8 m sequence of semi-consolidated mud clast breccias embedded in a distorted chaotic sediment matrix. The MTD is covered by a thin veneer of 50 cm hemipelagic, bioturbated diatomaceous mud. This stratigraphic boundary can be clearly distinguished by using physical properties data from Multi Sensor Core Logging and from fall-cone penetrometer shear strength measurements. The geochemical analysis of the pore-water shows undisturbed linear profiles measured from the seafloor downcore across the stratigraphic contact between overlying younger background-sediment and MTD below. This indicates that the investigated section has not been affected by a recent sediment destabilization in the course of the giant Tohoku-Oki earthquake event. Instead, we report an older landslide which occurred between 700 and 10,000 years ago, implying that submarine mass movements are dominant processes along the Japan Trench. However, they occur on local sites and not during each megathrust earthquake.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Thecosome pteropods (shelled pelagic molluscs) can play an important role in the food web of various ecosystems and play a key role in the cycling of carbon and carbonate. Since they harbor an aragonitic shell, they could be very sensitive to ocean acidification driven by the increase of anthropogenic CO2 emissions. The impact of changes in the carbonate chemistry was investigated on Limacina helicina, a key species of Arctic ecosystems. Pteropods were kept in culture under controlled pH conditions corresponding to pCO2 levels of 350 and 760 µatm. Calcification was estimated using a fluorochrome and the radioisotope 45Ca. It exhibits a 28 % decrease at the pH value expected for 2100 compared to the present pH value. This result supports the concern for the future of pteropods in a high-CO2 world, as well as of those species dependent upon them as a food resource. A decline of their populations would likely cause dramatic changes to the structure, function and services of polar ecosystems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

North Atlantic climate variations are reflected in sedimentary records from the northern Indian Ocean in which two basins, the Arabian Sea and the Bay of Bengal, are strongly affected by the monsoon. Contrary to the Bay of Bengal the Arabian Sea plays an important role in the global marine nitrogen cycle. In its mid-water oxygen minimum zone (OMZ) bioavailable fixed nitrogen is reduced to nitrogen gas (NO3- - > N2), whereas oxygen concentrations are slightly above the threshold of nitrate reduction in the OMZ of the Bay of Bengal. A coral colony (Porites lutea) growing south of Port Blair on the Andaman Islands in the Bay of Bengal was studied for its response to changes in the monsoon system and its link to temperature changes in the North Atlantic Ocean, between 1975 and 2006. Its linear extension rates, d13C and d18O values measured within the coral skeleton reveal a strong seasonality, which seems to be caused by the monsoon-driven reversal of the surface ocean circulation. The sampling site appears to be influenced by low salinity Bay of Bengal Water during the NE monsoon (boreal winter) and by the high salinity Arabian Sea Water during the SW monsoon in summer. The high salinity Arabian Sea Water circulates along with the Summer Monsoon Current (S-MC) from the Arabia Sea into the Bay of Bengal. Decreasing d18O and reconstructed salinity values correlate to the increasing SSTs in the North Atlantic Ocean indicating a reduced influence of the S-MC at the sampling site in the course of northern hemispheric warming. During such periods oxygen-depletion became stronger in the OMZ of the Arabian Sea as indicated by the sedimentary records. A reduced propagation of oxygen-depleted high salinity Arabian Sea Water into the Bay of Bengal could be a mechanism maintaining oxygen concentration above the threshold of nitrate reduction in the OMZ of the Bay of Bengal in times of global warming.