8 resultados para ElGamal, CZK, Multiple discrete logarithm assumption, Extended linear algebra
em DigitalCommons@The Texas Medical Center
Resumo:
Extremes of electrocardiographic QT interval are associated with increased risk for sudden cardiac death (SCD); thus, identification and characterization of genetic variants that modulate QT interval may elucidate the underlying etiology of SCD. Previous studies have revealed an association between a common genetic variant in NOS1AP and QT interval in populations of European ancestry, but this finding has not been extended to other ethnic populations. We sought to characterize the effects of NOS1AP genetic variants on QT interval in the multi-ethnic population-based Dallas Heart Study (DHS, n = 3,072). The SNP most strongly associated with QT interval in previous samples of European ancestry, rs16847548, was the most strongly associated in White (P = 0.005) and Black (P = 3.6 x 10(-5)) participants, with the same direction of effect in Hispanics (P = 0.17), and further showed a significant SNP x sex-interaction (P = 0.03). A second SNP, rs16856785, uncorrelated with rs16847548, was also associated with QT interval in Blacks (P = 0.01), with qualitatively similar results in Whites and Hispanics. In a previously genotyped cohort of 14,107 White individuals drawn from the combined Atherosclerotic Risk in Communities (ARIC) and Cardiovascular Health Study (CHS) cohorts, we validated both the second locus at rs16856785 (P = 7.63 x 10(-8)), as well as the sex-interaction with rs16847548 (P = 8.68 x 10(-6)). These data extend the association of genetic variants in NOS1AP with QT interval to a Black population, with similar trends, though not statistically significant at P<0.05, in Hispanics. In addition, we identify a strong sex-interaction and the presence of a second independent site within NOS1AP associated with the QT interval. These results highlight the consistent and complex role of NOS1AP genetic variants in modulating QT interval.
Resumo:
Objective: In this secondary data analysis, three statistical methodologies were implemented to handle cases with missing data in a motivational interviewing and feedback study. The aim was to evaluate the impact that these methodologies have on the data analysis. ^ Methods: We first evaluated whether the assumption of missing completely at random held for this study. We then proceeded to conduct a secondary data analysis using a mixed linear model to handle missing data with three methodologies (a) complete case analysis, (b) multiple imputation with explicit model containing outcome variables, time, and the interaction of time and treatment, and (c) multiple imputation with explicit model containing outcome variables, time, the interaction of time and treatment, and additional covariates (e.g., age, gender, smoke, years in school, marital status, housing, race/ethnicity, and if participants play on athletic team). Several comparisons were conducted including the following ones: 1) the motivation interviewing with feedback group (MIF) vs. the assessment only group (AO), the motivation interviewing group (MIO) vs. AO, and the intervention of the feedback only group (FBO) vs. AO, 2) MIF vs. FBO, and 3) MIF vs. MIO.^ Results: We first evaluated the patterns of missingness in this study, which indicated that about 13% of participants showed monotone missing patterns, and about 3.5% showed non-monotone missing patterns. Then we evaluated the assumption of missing completely at random by Little's missing completely at random (MCAR) test, in which the Chi-Square test statistic was 167.8 with 125 degrees of freedom, and its associated p-value was p=0.006, which indicated that the data could not be assumed to be missing completely at random. After that, we compared if the three different strategies reached the same results. For the comparison between MIF and AO as well as the comparison between MIF and FBO, only the multiple imputation with additional covariates by uncongenial and congenial models reached different results. For the comparison between MIF and MIO, all the methodologies for handling missing values obtained different results. ^ Discussions: The study indicated that, first, missingness was crucial in this study. Second, to understand the assumptions of the model was important since we could not identify if the data were missing at random or missing not at random. Therefore, future researches should focus on exploring more sensitivity analyses under missing not at random assumption.^
Resumo:
It is system dynamics that determines the function of cells, tissues and organisms. To develop mathematical models and estimate their parameters are an essential issue for studying dynamic behaviors of biological systems which include metabolic networks, genetic regulatory networks and signal transduction pathways, under perturbation of external stimuli. In general, biological dynamic systems are partially observed. Therefore, a natural way to model dynamic biological systems is to employ nonlinear state-space equations. Although statistical methods for parameter estimation of linear models in biological dynamic systems have been developed intensively in the recent years, the estimation of both states and parameters of nonlinear dynamic systems remains a challenging task. In this report, we apply extended Kalman Filter (EKF) to the estimation of both states and parameters of nonlinear state-space models. To evaluate the performance of the EKF for parameter estimation, we apply the EKF to a simulation dataset and two real datasets: JAK-STAT signal transduction pathway and Ras/Raf/MEK/ERK signaling transduction pathways datasets. The preliminary results show that EKF can accurately estimate the parameters and predict states in nonlinear state-space equations for modeling dynamic biochemical networks.
Resumo:
Interaction effect is an important scientific interest for many areas of research. Common approach for investigating the interaction effect of two continuous covariates on a response variable is through a cross-product term in multiple linear regression. In epidemiological studies, the two-way analysis of variance (ANOVA) type of method has also been utilized to examine the interaction effect by replacing the continuous covariates with their discretized levels. However, the implications of model assumptions of either approach have not been examined and the statistical validation has only focused on the general method, not specifically for the interaction effect.^ In this dissertation, we investigated the validity of both approaches based on the mathematical assumptions for non-skewed data. We showed that linear regression may not be an appropriate model when the interaction effect exists because it implies a highly skewed distribution for the response variable. We also showed that the normality and constant variance assumptions required by ANOVA are not satisfied in the model where the continuous covariates are replaced with their discretized levels. Therefore, naïve application of ANOVA method may lead to an incorrect conclusion. ^ Given the problems identified above, we proposed a novel method modifying from the traditional ANOVA approach to rigorously evaluate the interaction effect. The analytical expression of the interaction effect was derived based on the conditional distribution of the response variable given the discretized continuous covariates. A testing procedure that combines the p-values from each level of the discretized covariates was developed to test the overall significance of the interaction effect. According to the simulation study, the proposed method is more powerful then the least squares regression and the ANOVA method in detecting the interaction effect when data comes from a trivariate normal distribution. The proposed method was applied to a dataset from the National Institute of Neurological Disorders and Stroke (NINDS) tissue plasminogen activator (t-PA) stroke trial, and baseline age-by-weight interaction effect was found significant in predicting the change from baseline in NIHSS at Month-3 among patients received t-PA therapy.^
Resumo:
Current statistical methods for estimation of parametric effect sizes from a series of experiments are generally restricted to univariate comparisons of standardized mean differences between two treatments. Multivariate methods are presented for the case in which effect size is a vector of standardized multivariate mean differences and the number of treatment groups is two or more. The proposed methods employ a vector of independent sample means for each response variable that leads to a covariance structure which depends only on correlations among the $p$ responses on each subject. Using weighted least squares theory and the assumption that the observations are from normally distributed populations, multivariate hypotheses analogous to common hypotheses used for testing effect sizes were formulated and tested for treatment effects which are correlated through a common control group, through multiple response variables observed on each subject, or both conditions.^ The asymptotic multivariate distribution for correlated effect sizes is obtained by extending univariate methods for estimating effect sizes which are correlated through common control groups. The joint distribution of vectors of effect sizes (from $p$ responses on each subject) from one treatment and one control group and from several treatment groups sharing a common control group are derived. Methods are given for estimation of linear combinations of effect sizes when certain homogeneity conditions are met, and for estimation of vectors of effect sizes and confidence intervals from $p$ responses on each subject. Computational illustrations are provided using data from studies of effects of electric field exposure on small laboratory animals. ^
Resumo:
Atherosclerosis is widely accepted as a complex genetic phenotype and is the usual cause of cardiovascular disease, the world’s leading killer. Genetic factors have been proven to be important risk contributors for atherosclerosis and much work has been done to identify promising candidates that might play a role in the development of atherosclerosis. It is well known that many independent replications are needed to unequivocally establish a valid genotype-phenotype association across different populations before the findings are extended to clinical settings and to the expensive follow-up studies designed to identify causal genetic variants. Aiming to replicate the association with atherosclerosis in the Pathobiological Determinants of Atherosclerosis in Youth (PDAY) study, we assessed the relationship of 32 atherosclerosis candidate SNPs to atherosclerosis in the PDAY cohort, consisting of AA and EA young people aged 15-34 years who died of non-medical causes. Two association studies, a whole sample study and a 1:1 matched case control study were performed by use of multiple linear regression and logistic regression analyses, respectively. For the whole sample association study, 32 SNPs among 2,650 individuals (1,369 AA and 1,281 EA) were tested for the association with six early atherosclerosis phenotypes: abdominal aorta fatty streaks, abdominal aorta raised lesions, right coronary artery fatty streaks, right coronary artery raised lesions, thoracic aorta fatty streaks, and thoracic aorta raised lesions. For the matched case-control association study, 337 case-control paired samples were included; cases were chosen with the highest total raised lesion scores from the studied population, while controls were randomly selected from individuals that had no raised lesions and matched to cases by age, gender and race. Sixteen SNPs in 13 genes were found to be significantly associated with atherosclerosis in at least one of the PDAY association studies. Among these 16 findings: eight SNPs (rs9579646, rs6053733, rs3849150, rs10499903, rs2148079, rs5073691, rs10116277, and rs17228212) successfully replicated previous results, six SNPs (rs17222814, rs10811661, rs7028570, rs7291467, rs16996148 and rs10401969) were reported as new findings exclusive to our study, the last two of the 16 SNPs, rs501120 and rs6922269, showed either intriguing or conflicting result. SNP rs17222814 in ALOX5AP and SNP rs3849150 in LRRC18 were consistently associated with atherosclerosis in both prior and the two PDAY association studies. SNP rs3849150 was also identified to be highly correlated with a non-synonymous coding SNP, rs17772611, which may damage the protein (polyphen score = 0.996), suggesting that SNP rs17772611 may be the causal functional variant.^ In conclusion, our study added more support for the association of these candidate genes with atherosclerosis. SNPs rs3849150 and rs17772611 of LRRC18, as well as SNP rs17222814 of ALOX5AP, were the most significant findings from our study, and may be ranked among the best for further study.^
Resumo:
Treating patients with combined agents is a growing trend in cancer clinical trials. Evaluating the synergism of multiple drugs is often the primary motivation for such drug-combination studies. Focusing on the drug combination study in the early phase clinical trials, our research is composed of three parts: (1) We conduct a comprehensive comparison of four dose-finding designs in the two-dimensional toxicity probability space and propose using the Bayesian model averaging method to overcome the arbitrariness of the model specification and enhance the robustness of the design; (2) Motivated by a recent drug-combination trial at MD Anderson Cancer Center with a continuous-dose standard of care agent and a discrete-dose investigational agent, we propose a two-stage Bayesian adaptive dose-finding design based on an extended continual reassessment method; (3) By combining phase I and phase II clinical trials, we propose an extension of a single agent dose-finding design. We model the time-to-event toxicity and efficacy to direct dose finding in two-dimensional drug-combination studies. We conduct extensive simulation studies to examine the operating characteristics of the aforementioned designs and demonstrate the designs' good performances in various practical scenarios.^
New methods for quantification and analysis of quantitative real-time polymerase chain reaction data
Resumo:
Quantitative real-time polymerase chain reaction (qPCR) is a sensitive gene quantitation method that has been widely used in the biological and biomedical fields. The currently used methods for PCR data analysis, including the threshold cycle (CT) method, linear and non-linear model fitting methods, all require subtracting background fluorescence. However, the removal of background fluorescence is usually inaccurate, and therefore can distort results. Here, we propose a new method, the taking-difference linear regression method, to overcome this limitation. Briefly, for each two consecutive PCR cycles, we subtracted the fluorescence in the former cycle from that in the later cycle, transforming the n cycle raw data into n-1 cycle data. Then linear regression was applied to the natural logarithm of the transformed data. Finally, amplification efficiencies and the initial DNA molecular numbers were calculated for each PCR run. To evaluate this new method, we compared it in terms of accuracy and precision with the original linear regression method with three background corrections, being the mean of cycles 1-3, the mean of cycles 3-7, and the minimum. Three criteria, including threshold identification, max R2, and max slope, were employed to search for target data points. Considering that PCR data are time series data, we also applied linear mixed models. Collectively, when the threshold identification criterion was applied and when the linear mixed model was adopted, the taking-difference linear regression method was superior as it gave an accurate estimation of initial DNA amount and a reasonable estimation of PCR amplification efficiencies. When the criteria of max R2 and max slope were used, the original linear regression method gave an accurate estimation of initial DNA amount. Overall, the taking-difference linear regression method avoids the error in subtracting an unknown background and thus it is theoretically more accurate and reliable. This method is easy to perform and the taking-difference strategy can be extended to all current methods for qPCR data analysis.^