997 resultados para Biology, Biostatistics|Statistics|Health Sciences, Epidemiology
Resumo:
Ordinal outcomes are frequently employed in diagnosis and clinical trials. Clinical trials of Alzheimer's disease (AD) treatments are a case in point using the status of mild, moderate or severe disease as outcome measures. As in many other outcome oriented studies, the disease status may be misclassified. This study estimates the extent of misclassification in an ordinal outcome such as disease status. Also, this study estimates the extent of misclassification of a predictor variable such as genotype status. An ordinal logistic regression model is commonly used to model the relationship between disease status, the effect of treatment, and other predictive factors. A simulation study was done. First, data based on a set of hypothetical parameters and hypothetical rates of misclassification was created. Next, the maximum likelihood method was employed to generate likelihood equations accounting for misclassification. The Nelder-Mead Simplex method was used to solve for the misclassification and model parameters. Finally, this method was applied to an AD dataset to detect the amount of misclassification present. The estimates of the ordinal regression model parameters were close to the hypothetical parameters. β1 was hypothesized at 0.50 and the mean estimate was 0.488, β2 was hypothesized at 0.04 and the mean of the estimates was 0.04. Although the estimates for the rates of misclassification of X1 were not as close as β1 and β2, they validate this method. X 1 0-1 misclassification was hypothesized as 2.98% and the mean of the simulated estimates was 1.54% and, in the best case, the misclassification of k from high to medium was hypothesized at 4.87% and had a sample mean of 3.62%. In the AD dataset, the estimate for the odds ratio of X 1 of having both copies of the APOE 4 allele changed from an estimate of 1.377 to an estimate 1.418, demonstrating that the estimates of the odds ratio changed when the analysis includes adjustment for misclassification. ^
Resumo:
The factorial validity of the SF-36 was evaluated using confirmatory factor analysis (CFA) methods, structural equation modeling (SEM), and multigroup structural equation modeling (MSEM). First, the measurement and structural model of the hypothesized SF-36 was explicated. Second, the model was tested for the validity of a second-order factorial structure, upon evidence of model misfit, determined the best-fitting model, and tested the validity of the best-fitting model on a second random sample from the same population. Third, the best-fitting model was tested for invariance of the factorial structure across race, age, and educational subgroups using MSEM.^ The findings support the second-order factorial structure of the SF-36 as proposed by Ware and Sherbourne (1992). However, the results suggest that: (a) Mental Health and Physical Health covary; (b) general mental health cross-loads onto Physical Health; (c) general health perception loads onto Mental Health instead of Physical Health; (d) many of the error terms are correlated; and (e) the physical function scale is not reliable across these two samples. This hierarchical factor pattern was replicated across both samples of health care workers, suggesting that the post hoc model fitting was not data specific. Subgroup analysis suggests that the physical function scale is not reliable across the "age" or "education" subgroups and that the general mental health scale path from Mental Health is not reliable across the "white/nonwhite" or "education" subgroups.^ The importance of this study is in the use of SEM and MSEM in evaluating sample data from the use of the SF-36. These methods are uniquely suited to the analysis of latent variable structures and are widely used in other fields. The use of latent variable models for self reported outcome measures has become widespread, and should now be applied to medical outcomes research. Invariance testing is superior to mean scores or summary scores when evaluating differences between groups. From a practical, as well as, psychometric perspective, it seems imperative that construct validity research related to the SF-36 establish whether this same hierarchical structure and invariance holds for other populations.^ This project is presented as three articles to be submitted for publication. ^
Resumo:
The application of Markov processes is very useful to health-care problems. The objective of this study is to provide a structured methodology of forecasting cost based upon combining a stochastic model of utilization (Markov Chain) and deterministic cost function. The perspective of the cost in this study is the reimbursement for the services rendered. The data to be used is the OneCare database of claim records of their enrollees over a two-year period of January 1, 1996–December 31, 1997. The model combines a Markov Chain that describes the utilization pattern and its variability where the use of resources by risk groups (age, gender, and diagnosis) will be considered in the process and a cost function determined from a fixed schedule based on real costs or charges for those in the OneCare claims database. The cost function is a secondary application to the model. Goodness-of-fit will be used checked for the model against the traditional method of cost forecasting. ^
Resumo:
The main objective of this study was to develop and validate a computer-based statistical algorithm based on a multivariable logistic model that can be translated into a simple scoring system in order to ascertain stroke cases using hospital admission medical records data. This algorithm, the Risk Index Score (RISc), was developed using data collected prospectively by the Brain Attack Surveillance in Corpus Christ (BASIC) project. The validity of the RISc was evaluated by estimating the concordance of scoring system stroke ascertainment to stroke ascertainment accomplished by physician review of hospital admission records. The goal of this study was to develop a rapid, simple, efficient, and accurate method to ascertain the incidence of stroke from routine hospital admission hospital admission records for epidemiologic investigations. ^ The main objectives of this study were to develop and validate a computer-based statistical algorithm based on a multivariable logistic model that could be translated into a simple scoring system to ascertain stroke cases using hospital admission medical records data. (Abstract shortened by UMI.)^
Resumo:
Background. Obesity is a major health problem throughout the industrialized world. Despite numerous attempts to curtail the rapid growth of obesity, its incidence continues to rise. Therefore, it is crucial to better understand the etiology of obesity beyond the concept of energy balance.^ Aims. The first aim of this study was to first investigate the relationship between eating behaviors and body size. The second goal was to identify genetic variation associated with eating behaviors. Thirdly, this study aimed to examine the joint relationships between eating behavior, body size and genetic variation.^ Methods. This study utilized baseline data ascertained in young adults from the Training Interventions and Genetics of Exercise (TIGER) Study. Variables assessed included eating behavior (Emotional Eating Scale, Eating Attitudes Test-26, and the Block98 Food Frequency Questionnaire), body size (body mass index, waist and hip circumference, waist/hip ratio, and percent body fat), genetic variation in genes implicated related to the hypothalamic control of energy balance, and appropriate covariates (age, gender, race/ethnicity, smoking status, and physical activity. For the genetic association analyses, genotypes were collapsed by minor allele frequency, and haplotypes were estimated for each gene. Additionally, Bayesian networks were constructed in order to determine the relationships between genetic variation, eating behavior and body size.^ Results. We report that the EAT-26 score, Caloric intake, percent fat, fiber intake, HEAT index, and daily servings of vegetables, meats, grains, and fats were significantly associated with at least one body size measure. Multiple SNPs in 17 genes and haplotypes from 12 genes were tested for their association with body size. Variation within both DRD4 and HTR2A was found to be associated with EAT-26 score. In addition, variation in the ghrelin gene (GHRL) was significantly associated with daily Caloric intake. A significant interaction between daily servings of grains and the HEAT index and variation within the leptin receptor gene (LEPR) was shown to influence body size.^ Conclusion. This study has shown that there is a substantial genetic component to eating behavior and that genetic variation interacts with eating behavior to influence body size.^
Resumo:
Current statistical methods for estimation of parametric effect sizes from a series of experiments are generally restricted to univariate comparisons of standardized mean differences between two treatments. Multivariate methods are presented for the case in which effect size is a vector of standardized multivariate mean differences and the number of treatment groups is two or more. The proposed methods employ a vector of independent sample means for each response variable that leads to a covariance structure which depends only on correlations among the $p$ responses on each subject. Using weighted least squares theory and the assumption that the observations are from normally distributed populations, multivariate hypotheses analogous to common hypotheses used for testing effect sizes were formulated and tested for treatment effects which are correlated through a common control group, through multiple response variables observed on each subject, or both conditions.^ The asymptotic multivariate distribution for correlated effect sizes is obtained by extending univariate methods for estimating effect sizes which are correlated through common control groups. The joint distribution of vectors of effect sizes (from $p$ responses on each subject) from one treatment and one control group and from several treatment groups sharing a common control group are derived. Methods are given for estimation of linear combinations of effect sizes when certain homogeneity conditions are met, and for estimation of vectors of effect sizes and confidence intervals from $p$ responses on each subject. Computational illustrations are provided using data from studies of effects of electric field exposure on small laboratory animals. ^
Resumo:
A non-parametric method was developed and tested to compare the partial areas under two correlated Receiver Operating Characteristic curves. Based on the theory of generalized U-statistics the mathematical formulas have been derived for computing ROC area, and the variance and covariance between the portions of two ROC curves. A practical SAS application also has been developed to facilitate the calculations. The accuracy of the non-parametric method was evaluated by comparing it to other methods. By applying our method to the data from a published ROC analysis of CT image, our results are very close to theirs. A hypothetical example was used to demonstrate the effects of two crossed ROC curves. The two ROC areas are the same. However each portion of the area between two ROC curves were found to be significantly different by the partial ROC curve analysis. For computation of ROC curves with large scales, such as a logistic regression model, we applied our method to the breast cancer study with Medicare claims data. It yielded the same ROC area computation as the SAS Logistic procedure. Our method also provides an alternative to the global summary of ROC area comparison by directly comparing the true-positive rates for two regression models and by determining the range of false-positive values where the models differ. ^
Resumo:
The objective of this study is to identify the relationship between population density and the initial stages of the spread of disease in a local population. This study proposes to concentrate on the question of how population density affects the distribution of the susceptible individuals in a local population and thus affects the spread of the disease, measles. Population density is measured by the average of the number of contacts with susceptible individuals by each individual in the population during a fixed-length time period. The term “contact with susceptible individuals” means sufficient contact between two people for the disease to pass from an infectious person to a susceptible person. The fixed-length time period is taken to be the average length of time an infected person is infectious without symptoms of the disease. For this study of measles, the time period will be seven days. ^ While much attention has been given to modeling the entire epidemic process of measles, attempts have not been made to study the characteristics of contact rates required to initiate an epidemic. This study explores the relationship between population density, given a specific herd immunity rate in the population, and initial rate of the spread of the disease by considering the underlying distribution of contacts with susceptibles by the individuals in the population. ^ This study does not seek to model an entire measles epidemic, but to model the above stated relationship for the local population within which the first infective person is introduced. This study describes the mathematical relationship between population density parameters and contact distribution parameters. ^ The results are displayed in graphs that show the effects of different population densities on the spread of disease. The results support the idea that the number of new infectives is strongly related to the distribution of susceptible contacts. The results also show large differences in the epidemic measures between populations with densities equal to four versus three. ^
Resumo:
This study applies the multilevel analysis technique to longitudinal data of a large clinical trial. The technique accounts for the correlation at different levels when modeling repeated blood pressure measurements taken throughout the trial. This modeling allows for closer inspection of the remaining correlation and non-homogeneity of variance in the data. Three methods of modeling the correlation were compared. ^
Resumo:
It is estimated that 50% of all lung cancer patients continue to smoke after diagnosis. Many of these lung cancer patients who are current smokers often experience tremendous guilt and responsibility for their disease, and feel it might be too late for them to quit smoking. In addition, many oncologists may be heard to say that it is 'too late', 'it doesn't matter', 'it is too difficult', 'it is too stressful' for their patients to stop smoking, or they never identify the smoking status of the patient. Many oncologists feel unprepared to address smoking cessation as part of their clinical practice. In reality, physicians can have tremendous effects on motivating patients, particularly when patients are initially being diagnosed with cancer. More information is needed to convince patients to quit smoking and to encourage clinicians to assist patients with their smoking cessation. ^ In this current study, smoking status at time of lung cancer diagnosis was assessed to examine its impact on complications and survival, after exploring the reliability of smoking data that is self-reported. Logistic Regression was used to determine the risks of smoking prior to lung resection. In addition, survival analysis was performed to examine the impact of smoking on survival. ^ The reliability of how patients report their smoking status was high, but there was some discordance between current smokers and recent quitters. In addition, we found that cigarette pack-year history and duration of smoking cessation were directly related to the rate of a pulmonary complication. In regards to survival, we found that current smoking at time of lung cancer diagnosis was an independent predictor of early stage lung cancer. This evidence supports the idea that it is "never too late" for patients to quit smoking and health care providers should incorporate smoking status regularly into their clinical practice.^
Resumo:
Studies on the relationship between psychosocial determinants and HIV risk behaviors have produced little evidence to support hypotheses based on theoretical relationships. One limitation inherent in many articles in the literature is the method of measurement of the determinants and the analytic approach selected. ^ To reduce the misclassification associated with unit scaling of measures specific to internalized homonegativity, I evaluated the psychometric properties of the Reactions to Homosexuality scale in a confirmatory factor analytic framework. In addition, I assessed the measurement invariance of the scale across racial/ethnic classifications in a sample of men who have sex with men. The resulting measure contained eight items loading on three first-order factors. Invariance assessment identified metric and partial strong invariance between racial/ethnic groups in the sample. ^ Application of the updated measure to a structural model allowed for the exploration of direct and indirect effects of internalized homonegativity on unprotected anal intercourse. Pathways identified in the model show that drug and alcohol use at last sexual encounter, the number of sexual partners in the previous three months and sexual compulsivity all contribute directly to risk behavior. Internalized homonegativity reduced the likelihood of exposure to drugs, alcohol or higher numbers of partners. For men who developed compulsive sexual behavior as a coping strategy for internalized homonegativity, there was an increase in the prevalence odds of risk behavior. ^ In the final stage of the analysis, I conducted a latent profile analysis of the items in the updated Reactions to Homosexuality scale. This analysis identified five distinct profiles, which suggested that the construct was not homogeneous in samples of men who have sex with men. Lack of prior consideration of these distinct manifestations of internalized homonegativity may have contributed to the analytic difficulty in identifying a relationship between the trait and high-risk sexual practices. ^
Resumo:
Background. Previous research shows inconsistent results as to the association between part-time employment and sexual behavior among younger teens. Studies of older teens cannot be generalized to younger teens because of the wide differences in types of work performed, nature of work environments, and work intensity. Objective. Examine the relationship between part-time employment and sexual behavior in a cross-sectional sample of public middle school students in Houston, Texas. Methods . The study presents a secondary analysis of data from the It’s Your Game…Keep it Real baseline data collection (11/2004–1/2005). It’s Your Game… is an intervention program for middle school students designed to prevent Sexually Transmitted Infections. Statistical analysis. Univariate and multivariate logistic regression analyses were conducted to examine the association between part-time employment and vaginal intercourse: (a) ever had sex; and (b) current sexual activity. Results. Overall, 13.2% of students worked for pay; male students were 1.5 times as likely as females to be working. Of all the students, 11.0% had had sexual intercourse; students who worked were 3 times more likely to be sexually experienced than those who did not. Among students who were sexually experienced, 67.0% were currently sexually active. After adjusting for the other covariates, Hispanic students were almost 3.6 times more likely to report current sexual activity compared to students in other racial/ethnic groups. In univariate analysis, students who worked 1-5 hrs/week were more likely to be sexually experienced than those not currently employed, and the likelihood increased with number of hours worked. There is a similar pattern in the multivariate model, but the odds ratios are too close for the evidence to be more than suggestive. Of sexually experienced students, students working 1-5 hrs/week were 2.7 times more likely to report current sexual intercourse than those not working; those working >5 hrs/week were 4.7 times more likely. The multivariate model showed a similar increase in likelihood, and adjustment for covariates increased these associations: students who worked 1-5 hrs/week were 3.6 times more likely to report current sexual intercourse, and students who worked >5 hrs/week were 4.5 times more likely, than students not currently employed.^
Resumo:
Generalized linear Poisson and logistic regression models were utilized to examine the relationship between temperature and precipitation and cases of Saint Louis encephalitis virus spread in the Houston metropolitan area. The models were investigated with and without repeated measures, with a first order autoregressive (AR1) correlation structure used for the repeated measures model. The two types of Poisson regression models, with and without correlation structure, showed that a unit increase in temperature measured in degrees Fahrenheit increases the occurrence of the virus 1.7 times and a unit increase in precipitation measured in inches increases the occurrence of the virus 1.5 times. Logistic regression did not show these covariates to be significant as predictors for encephalitis activity in Houston for either correlation structure. This discrepancy for the logistic model could be attributed to the small data set.^ Keywords: Saint Louis Encephalitis; Generalized Linear Model; Poisson; Logistic; First Order Autoregressive; Temperature; Precipitation. ^
Resumo:
In Part One, the foundations of Bayesian inference are reviewed, and the technicalities of the Bayesian method are illustrated. Part Two applies the Bayesian meta-analysis program, the Confidence Profile Method (CPM), to clinical trial data and evaluates the merits of using Bayesian meta-analysis for overviews of clinical trials.^ The Bayesian method of meta-analysis produced similar results to the classical results because of the large sample size, along with the input of a non-preferential prior probability distribution. These results were anticipated through explanations in Part One of the mechanics of the Bayesian approach. ^
Resumo:
Mixed longitudinal designs are important study designs for many areas of medical research. Mixed longitudinal studies have several advantages over cross-sectional or pure longitudinal studies, including shorter study completion time and ability to separate time and age effects, thus are an attractive choice. Statistical methodology used in general longitudinal studies has been rapidly developing within the last few decades. Common approaches for statistical modeling in studies with mixed longitudinal designs have been the linear mixed-effects model incorporating an age or time effect. The general linear mixed-effects model is considered an appropriate choice to analyze repeated measurements data in longitudinal studies. However, common use of linear mixed-effects model on mixed longitudinal studies often incorporates age as the only random-effect but fails to take into consideration the cohort effect in conducting statistical inferences on age-related trajectories of outcome measurements. We believe special attention should be paid to cohort effects when analyzing data in mixed longitudinal designs with multiple overlapping cohorts. Thus, this has become an important statistical issue to address. ^ This research aims to address statistical issues related to mixed longitudinal studies. The proposed study examined the existing statistical analysis methods for the mixed longitudinal designs and developed an alternative analytic method to incorporate effects from multiple overlapping cohorts as well as from different aged subjects. The proposed study used simulation to evaluate the performance of the proposed analytic method by comparing it with the commonly-used model. Finally, the study applied the proposed analytic method to the data collected by an existing study Project HeartBeat!, which had been evaluated using traditional analytic techniques. Project HeartBeat! is a longitudinal study of cardiovascular disease (CVD) risk factors in childhood and adolescence using a mixed longitudinal design. The proposed model was used to evaluate four blood lipids adjusting for age, gender, race/ethnicity, and endocrine hormones. The result of this dissertation suggest the proposed analytic model could be a more flexible and reliable choice than the traditional model in terms of fitting data to provide more accurate estimates in mixed longitudinal studies. Conceptually, the proposed model described in this study has useful features, including consideration of effects from multiple overlapping cohorts, and is an attractive approach for analyzing data in mixed longitudinal design studies.^