956 resultados para censored translations
Resumo:
This work develops a new methodology in order to discriminate models for interval-censored data based on bootstrap residual simulation by observing the deviance difference from one model in relation to another, according to Hinde (1992). Generally, this sort of data can generate a large number of tied observations and, in this case, survival time can be regarded as discrete. Therefore, the Cox proportional hazards model for grouped data (Prentice & Gloeckler, 1978) and the logistic model (Lawless, 1982) can befitted by means of generalized linear models. Whitehead (1989) considered censoring to be an indicative variable with a binomial distribution and fitted the Cox proportional hazards model using complementary log-log as a link function. In addition, a logistic model can be fitted using logit as a link function. The proposed methodology arises as an alternative to the score tests developed by Colosimo et al. (2000), where such models can be obtained for discrete binary data as particular cases from the Aranda-Ordaz distribution asymmetric family. These tests are thus developed with a basis on link functions to generate such a fit. The example that motivates this study was the dataset from an experiment carried out on a flax cultivar planted on four substrata susceptible to the pathogen Fusarium oxysoprum. The response variable, which is the time until blighting, was observed in intervals during 52 days. The results were compared with the model fit and the AIC values.
Resumo:
In this paper we introduce the concept of conservative translation between logics. We present some necessary and sufficient conditions for a translation to be conservative and study some general properties of logical systems, these properties being characterized by the existence of conservative translations between the systems. We prove that the class constituted by logics and conservative translations between them determines a co-complete subcategory of the bi-complex category constituted by logics and translations. (C) 2001 Elsevier B.V. B.V. All rights reserved.
Resumo:
One common problem in all basic techniques of knowledge representation is the handling of the trade-off between precision of inferences and resource constraints, such as time and memory. Michalski and Winston (1986) suggested the Censored Production Rule (CPR) as an underlying representation and computational mechanism to enable logic based systems to exhibit variable precision in which certainty varies while specificity stays constant. As an extension of CPR, the Hierarchical Censored Production Rules (HCPRs) system of knowledge representation, proposed by Bharadwaj & Jain (1992), exhibits both variable certainty as well as variable specificity and offers mechanisms for handling the trade-off between the two. An HCPR has the form: Decision If(preconditions) Unless(censor) Generality(general_information) Specificity(specific_information). As an attempt towards evolving a generalized knowledge representation, an Extended Hierarchical Censored Production Rules (EHCPRs) system is suggested in this paper. With the inclusion of new operators, an Extended Hierarchical Censored Production Rule (EHCPR) takes the general form: Concept If (Preconditions) Unless (Exceptions) Generality (General-Concept) Specificity (Specific Concepts) Has_part (default: structural-parts) Has_property (default:characteristic-properties) Has_instance (instances). How semantic networks and frames are represented in terms of an EHCPRs is shown. Multiple inheritance, inheritance with and without cancellation, recognition with partial match, and a few default logic problems are shown to be tackled efficiently in the proposed system.
Resumo:
Pós-graduação em Letras - IBILCE
Resumo:
Evaluations of measurement invariance provide essential construct validity evidence. However, the quality of such evidence is partly dependent upon the validity of the resulting statistical conclusions. The presence of Type I or Type II errors can render measurement invariance conclusions meaningless. The purpose of this study was to determine the effects of categorization and censoring on the behavior of the chi-square/likelihood ratio test statistic and two alternative fit indices (CFI and RMSEA) under the context of evaluating measurement invariance. Monte Carlo simulation was used to examine Type I error and power rates for the (a) overall test statistic/fit indices, and (b) change in test statistic/fit indices. Data were generated according to a multiple-group single-factor CFA model across 40 conditions that varied by sample size, strength of item factor loadings, and categorization thresholds. Seven different combinations of model estimators (ML, Yuan-Bentler scaled ML, and WLSMV) and specified measurement scales (continuous, censored, and categorical) were used to analyze each of the simulation conditions. As hypothesized, non-normality increased Type I error rates for the continuous scale of measurement and did not affect error rates for the categorical scale of measurement. Maximum likelihood estimation combined with a categorical scale of measurement resulted in more correct statistical conclusions than the other analysis combinations. For the continuous and censored scales of measurement, the Yuan-Bentler scaled ML resulted in more correct conclusions than normal-theory ML. The censored measurement scale did not offer any advantages over the continuous measurement scale. Comparing across fit statistics and indices, the chi-square-based test statistics were preferred over the alternative fit indices, and ΔRMSEA was preferred over ΔCFI. Results from this study should be used to inform the modeling decisions of applied researchers. However, no single analysis combination can be recommended for all situations. Therefore, it is essential that researchers consider the context and purpose of their analyses.
Resumo:
For the first time, we introduce a generalized form of the exponentiated generalized gamma distribution [Cordeiro et al. The exponentiated generalized gamma distribution with application to lifetime data, J. Statist. Comput. Simul. 81 (2011), pp. 827-842.] that is the baseline for the log-exponentiated generalized gamma regression model. The new distribution can accommodate increasing, decreasing, bathtub- and unimodal-shaped hazard functions. A second advantage is that it includes classical distributions reported in the lifetime literature as special cases. We obtain explicit expressions for the moments of the baseline distribution of the new regression model. The proposed model can be applied to censored data since it includes as sub-models several widely known regression models. It therefore can be used more effectively in the analysis of survival data. We obtain maximum likelihood estimates for the model parameters by considering censored data. We show that our extended regression model is very useful by means of two applications to real data.
Resumo:
The beta-Birnbaum-Saunders (Cordeiro and Lemonte, 2011) and Birnbaum-Saunders (Birnbaum and Saunders, 1969a) distributions have been used quite effectively to model failure times for materials subject to fatigue and lifetime data. We define the log-beta-Birnbaum-Saunders distribution by the logarithm of the beta-Birnbaum-Saunders distribution. Explicit expressions for its generating function and moments are derived. We propose a new log-beta-Birnbaum-Saunders regression model that can be applied to censored data and be used more effectively in survival analysis. We obtain the maximum likelihood estimates of the model parameters for censored data and investigate influence diagnostics. The new location-scale regression model is modified for the possibility that long-term survivors may be presented in the data. Its usefulness is illustrated by means of two real data sets. (C) 2011 Elsevier B.V. All rights reserved.
Resumo:
Development aid involves a complex network of numerous and extremely heterogeneous actors. Nevertheless, all actors seem to speak the same ‘development jargon’ and to display a congruence that extends from the donor over the professional consultant to the village chief. And although the ideas about what counts as ‘good’ and ‘bad’ aid have constantly changed over time —with new paradigms and policies sprouting every few years— the apparent congruence between actors more or less remains unchanged. How can this be explained? Is it a strategy of all actors to get into the pocket of the donor, or are the social dynamics in development aid more complex? When a new development paradigm appears, where does it come from and how does it gain support? Is this support really homogeneous? To answer the questions, a multi-sited ethnography was conducted in the sector of water-related development aid, with a focus on 3 paradigms that are currently hegemonic in this sector: Integrated Water Resources Management, Capacity Building, and Adaptation to Climate Change. The sites of inquiry were: the headquarters of a multilateral organization, the headquarters of a development NGO, and the Inner Niger Delta in Mali. The research shows that paradigm shifts do not happen overnight but that new paradigms have long lines of descent. Moreover, they require a lot of work from actors in order to become hegemonic; the actors need to create a tight network of support. Each actor, however, interprets the paradigms in a slightly different way, depending on the position in the network. They implant their own interests in their interpretation of the paradigm (the actors ‘translate’ their interests), regardless of whether they constitute the donor, a mediator, or the aid recipient. These translations are necessary to cement and reproduce the network.
Resumo:
Gleno-humeral joint (GHJ) is the most mobile joint of the human body. This is related to theincongr uence between the large humeral head articulating with the much smaller glenoid (ratio 3:1). The GHJ laxity is the ability of the humeral head to be passively translated on the glenoid fossa and, when physiological, it guarantees the normal range of motion of the joint. Three-dimensional GHJ linear displacements have been measured, both in vivo and in vitro by means of different instrumental techniques. In vivo gleno-humeral displacements have been assessed by means of stereophotogrammetry, electromagnetic tracking sensors, and bio-imaging techniques. Both stereophotogrammetric systems and electromagnetic tracking devices, due to the deformation of the soft tissues surrounding the bones, are not capable to accurately assess small displacements, such as gleno-humeral joint translations. The bio-imaging techniques can ensure for an accurate joint kinematic (linear and angular displacement) description, but, due to the radiation exposure, most of these techniques, such as computer tomography or fluoroscopy, are invasive for patients. Among the bioimaging techniques, an alternative which could provide an acceptable level of accuracy and that is innocuous for patients is represented by magnetic resonance imaging (MRI). Unfortunately, only few studies have been conducted for three-dimensional analysis and very limited data is available in situations where preset loads are being applied. The general aim of this doctoral thesis is to develop a non-invasive methodology based on open-MRI for in-vivo evaluation of the gleno-humeral translation components in healthy subjects under the application of external loads.
Resumo:
In biostatistical applications interest often focuses on the estimation of the distribution of a time-until-event variable T. If one observes whether or not T exceeds an observed monitoring time at a random number of monitoring times, then the data structure is called interval censored data. We extend this data structure by allowing the presence of a possibly time-dependent covariate process that is observed until end of follow up. If one only assumes that the censoring mechanism satisfies coarsening at random, then, by the curve of dimensionality, typically no regular estimators will exist. To fight the curse of dimensionality we follow the approach of Robins and Rotnitzky (1992) by modeling parameters of the censoring mechanism. We model the right-censoring mechanism by modeling the hazard of the follow up time, conditional on T and the covariate process. For the monitoring mechanism we avoid modeling the joint distribution of the monitoring times by only modeling a univariate hazard of the pooled monitoring times, conditional on the follow up time, T, and the covariates process, which can be estimated by treating the pooled sample of monitoring times as i.i.d. In particular, it is assumed that the monitoring times and the right-censoring times only depend on T through the observed covariate process. We introduce inverse probability of censoring weighted (IPCW) estimator of the distribution of T and of smooth functionals thereof which are guaranteed to be consistent and asymptotically normal if we have available correctly specified semiparametric models for the two hazards of the censoring process. Furthermore, given such correctly specified models for these hazards of the censoring process, we propose a one-step estimator which will improve on the IPCW estimator if we correctly specify a lower-dimensional working model for the conditional distribution of T, given the covariate process, that remains consistent and asymptotically normal if this latter working model is misspecified. It is shown that the one-step estimator is efficient if each subject is at most monitored once and the working model contains the truth. In general, it is shown that the one-step estimator optimally uses the surrogate information if the working model contains the truth. It is not optimal in using the interval information provided by the current status indicators at the monitoring times, but simulations in Peterson, van der Laan (1997) show that the efficiency loss is small.