872 resultados para interval censoring
Resumo:
This paper introduces a novel approach to making inference about the regression parameters in the accelerated failure time (AFT) model for current status and interval censored data. The estimator is constructed by inverting a Wald type test for testing a null proportional hazards model. A numerically efficient Markov chain Monte Carlo (MCMC) based resampling method is proposed to simultaneously obtain the point estimator and a consistent estimator of its variance-covariance matrix. We illustrate our approach with interval censored data sets from two clinical studies. Extensive numerical studies are conducted to evaluate the finite sample performance of the new estimators.
Resumo:
In this paper, we derive score test statistics to discriminate between proportional hazards and proportional odds models for grouped survival data. These models are embedded within a power family transformation in order to obtain the score tests. In simple cases, some small-sample results are obtained for the score statistics using Monte Carlo simulations. Score statistics have distributions well approximated by the chi-squared distribution. Real examples illustrate the proposed tests.
Resumo:
In this article, proportional hazards and logistic models for grouped survival data were extended to incorporate time-dependent covariates. The extension was motivated by a forestry experiment designed to compare five different water stresses in Eucalyptus grandis seedlings. The response was the seedling lifetime. The data set was grouped since there were just three occasions in which the seedlings was visited by the researcher. In each of these occasions also the shoot height was measured and therefore it is a time-dependent covariate. Both extended models were used in this example, and the results were very similar.
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
A method is given for proving efficiency of NPMLE directly linked to empirical process theory. The conditions in general are appropriate consistency of the NPMLE, differentiability of the model, differentiability of the parameter of interest, local convexity of the parameter space, and a Donsker class condition for the class of efficient influence functions obtained by varying the parameters. For the case that the model is linear in the parameter and the parameter space is convex, as with most nonparametric missing data models, we show that the method leads to an identity for the NPMLE which almost says that the NPMLE is efficient and provides us straightforwardly with a consistency and efficiency proof. This identify is extended to an almost linear class of models which contain biased sampling models. To illustrate, the method is applied to the univariate censoring model, random truncation models, interval censoring case I model, the class of parametric models and to a class of semiparametric models.
Resumo:
In interval-censored survival data, the event of interest is not observed exactly but is only known to occur within some time interval. Such data appear very frequently. In this paper, we are concerned only with parametric forms, and so a location-scale regression model based on the exponentiated Weibull distribution is proposed for modeling interval-censored data. We show that the proposed log-exponentiated Weibull regression model for interval-censored data represents a parametric family of models that include other regression models that are broadly used in lifetime data analysis. Assuming the use of interval-censored data, we employ a frequentist analysis, a jackknife estimator, a parametric bootstrap and a Bayesian analysis for the parameters of the proposed model. We derive the appropriate matrices for assessing local influences on the parameter estimates under different perturbation schemes and present some ways to assess global influences. Furthermore, for different parameter settings, sample sizes and censoring percentages, various simulations are performed; in addition, the empirical distribution of some modified residuals are displayed and compared with the standard normal distribution. These studies suggest that the residual analysis usually performed in normal linear regression models can be straightforwardly extended to a modified deviance residual in log-exponentiated Weibull regression models for interval-censored data. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
This work develops a new methodology in order to discriminate models for interval-censored data based on bootstrap residual simulation by observing the deviance difference from one model in relation to another, according to Hinde (1992). Generally, this sort of data can generate a large number of tied observations and, in this case, survival time can be regarded as discrete. Therefore, the Cox proportional hazards model for grouped data (Prentice & Gloeckler, 1978) and the logistic model (Lawless, 1982) can befitted by means of generalized linear models. Whitehead (1989) considered censoring to be an indicative variable with a binomial distribution and fitted the Cox proportional hazards model using complementary log-log as a link function. In addition, a logistic model can be fitted using logit as a link function. The proposed methodology arises as an alternative to the score tests developed by Colosimo et al. (2000), where such models can be obtained for discrete binary data as particular cases from the Aranda-Ordaz distribution asymmetric family. These tests are thus developed with a basis on link functions to generate such a fit. The example that motivates this study was the dataset from an experiment carried out on a flax cultivar planted on four substrata susceptible to the pathogen Fusarium oxysoprum. The response variable, which is the time until blighting, was observed in intervals during 52 days. The results were compared with the model fit and the AIC values.
Resumo:
In biostatistical applications interest often focuses on the estimation of the distribution of a time-until-event variable T. If one observes whether or not T exceeds an observed monitoring time at a random number of monitoring times, then the data structure is called interval censored data. We extend this data structure by allowing the presence of a possibly time-dependent covariate process that is observed until end of follow up. If one only assumes that the censoring mechanism satisfies coarsening at random, then, by the curve of dimensionality, typically no regular estimators will exist. To fight the curse of dimensionality we follow the approach of Robins and Rotnitzky (1992) by modeling parameters of the censoring mechanism. We model the right-censoring mechanism by modeling the hazard of the follow up time, conditional on T and the covariate process. For the monitoring mechanism we avoid modeling the joint distribution of the monitoring times by only modeling a univariate hazard of the pooled monitoring times, conditional on the follow up time, T, and the covariates process, which can be estimated by treating the pooled sample of monitoring times as i.i.d. In particular, it is assumed that the monitoring times and the right-censoring times only depend on T through the observed covariate process. We introduce inverse probability of censoring weighted (IPCW) estimator of the distribution of T and of smooth functionals thereof which are guaranteed to be consistent and asymptotically normal if we have available correctly specified semiparametric models for the two hazards of the censoring process. Furthermore, given such correctly specified models for these hazards of the censoring process, we propose a one-step estimator which will improve on the IPCW estimator if we correctly specify a lower-dimensional working model for the conditional distribution of T, given the covariate process, that remains consistent and asymptotically normal if this latter working model is misspecified. It is shown that the one-step estimator is efficient if each subject is at most monitored once and the working model contains the truth. In general, it is shown that the one-step estimator optimally uses the surrogate information if the working model contains the truth. It is not optimal in using the interval information provided by the current status indicators at the monitoring times, but simulations in Peterson, van der Laan (1997) show that the efficiency loss is small.
Resumo:
In medical follow-up studies, ordered bivariate survival data are frequently encountered when bivariate failure events are used as the outcomes to identify the progression of a disease. In cancer studies interest could be focused on bivariate failure times, for example, time from birth to cancer onset and time from cancer onset to death. This paper considers a sampling scheme where the first failure event (cancer onset) is identified within a calendar time interval, the time of the initiating event (birth) can be retrospectively confirmed, and the occurrence of the second event (death) is observed sub ject to right censoring. To analyze this type of bivariate failure time data, it is important to recognize the presence of bias arising due to interval sampling. In this paper, nonparametric and semiparametric methods are developed to analyze the bivariate survival data with interval sampling under stationary and semi-stationary conditions. Numerical studies demonstrate the proposed estimating approaches perform well with practical sample sizes in different simulated models. We apply the proposed methods to SEER ovarian cancer registry data for illustration of the methods and theory.
Resumo:
This paper examines performances that defy established representations of disease, deformity and bodily difference. Historically, the ‘deformed’ body has been cast – onstage and in sideshows – as flawed, an object of pity, or an example of the human capacity to overcome. Such representations define the boundaries of the ‘normal’ body by displaying its Other. They bracket the ‘abnormal’ body off as an example of deviance from the ‘norm’, thus, paradoxically, decreasing the social and symbolic visibility (and agency) of disabled people. Yet, in contemporary theory and culture, these representations are reappropriated – by disabled artists, certainly, but also as what Carrie Sandahl has called a ‘master trope’ for representing a range of bodily differences. In this paper, I investigate this phenomenon. I analyse French Canadian choreographer Marie Chouinard’s bODY rEMIX/gOLDBERG vARIATIONS, in which 10 able-bodied dancers are reborn as bizarre biotechnical mutants via the use of crutches, walkers, ballet shoes and barres as prosthetic pseudo-organs. These bodies defy boundaries, defy expectations, develop new modes of expression, and celebrate bodily difference. The self-inflicted pain dancers experience during training is cast as a ‘disablement’ that is ultimately ‘enabling’. I ask what effect encountering able bodies celebrating ‘dis’ or ‘diff’ ability has on audiences. Do we see the emergence of a once-repressed Other, no longer silenced, censored or negated? Or does using ‘disability’ to express the dancers’ difference and self-determination usurp a ‘trope’ by which disabled people themselves might speak back to the dominant culture, creating further censorship?
Resumo:
This paper investigates the robust H∞ control for Takagi-Sugeno (T-S) fuzzy systems with interval time-varying delay. By employing a new and tighter integral inequality and constructing an appropriate type of Lyapunov functional, delay-dependent stability criteria are derived for the control problem. Because neither any model transformation nor free weighting matrices are employed in our theoretical derivation, the developed stability criteria significantly improve and simplify the existing stability conditions. Also, the maximum allowable upper delay bound and controller feedback gains can be obtained simultaneously from the developed approach by solving a constrained convex optimization problem. Numerical examples are given to demonstrate the effectiveness of the proposed methods.
Resumo:
Loss of the short arm of chromosome 1 is frequently observed in many tumor types, including melanoma. We recently localized a third melanoma susceptibility locus to chromosome band 1p22. Critical recombinants in linked families localized the gene to a 15-Mb region between D1S430 and D1S2664. To map the locus more finely we have performed studies to assess allelic loss across the region in a panel of melanomas from 1p22-linked families, sporadic melanomas, and melanoma cell lines. Eighty percent of familial melanomas exhibited loss of heterozygosity (LOH) within the region, with a smallest region of overlapping deletions (SRO) of 9 Mb between D1S207 and D1S435. This high frequency of LOH makes it very likely that the susceptibility locus is a tumor suppressor. In sporadic tumors, four SROs were defined. SRO1 and SRO2 map within the critical recombinant and familial tumor region, indicating that one or the other is likely to harbor the susceptibility gene. However, SRO3 may also be significant because it overlaps with the markers with the highest 2-point LOD score (D1S2776), part of the linkage recombinant region, and the critical region defined in mesothelioma. The candidate genes PRKCL2 and GTF2B, within SRO2, and TGFBR3, CDC7, and EVI5, in a broad region encompassing SRO3, were screened in 1p22-linked melanoma kindreds, but no coding mutations were detected. Allelic loss in melanoma cell lines was significantly less frequent than in fresh tumors, indicating that this gene may not be involved late in progression, such as in overriding cellular senescence, necessary for the propagation of melanoma cells in culture.