963 resultados para COUNT DATA MODELS
Resumo:
OBJECTIVE: To better understand the structure of the Patient Assessment of Chronic Illness Care (PACIC) instrument. More specifically to test all published validation models, using one single data set and appropriate statistical tools. DESIGN: Validation study using data from cross-sectional survey. PARTICIPANTS: A population-based sample of non-institutionalized adults with diabetes residing in Switzerland (canton of Vaud). MAIN OUTCOME MEASURE: French version of the 20-items PACIC instrument (5-point response scale). We conducted validation analyses using confirmatory factor analysis (CFA). The original five-dimension model and other published models were tested with three types of CFA: based on (i) a Pearson estimator of variance-covariance matrix, (ii) a polychoric correlation matrix and (iii) a likelihood estimation with a multinomial distribution for the manifest variables. All models were assessed using loadings and goodness-of-fit measures. RESULTS: The analytical sample included 406 patients. Mean age was 64.4 years and 59% were men. Median of item responses varied between 1 and 4 (range 1-5), and range of missing values was between 5.7 and 12.3%. Strong floor and ceiling effects were present. Even though loadings of the tested models were relatively high, the only model showing acceptable fit was the 11-item single-dimension model. PACIC was associated with the expected variables of the field. CONCLUSIONS: Our results showed that the model considering 11 items in a single dimension exhibited the best fit for our data. A single score, in complement to the consideration of single-item results, might be used instead of the five dimensions usually described.
Resumo:
Epidemiological processes leave a fingerprint in the pattern of genetic structure of virus populations. Here, we provide a new method to infer epidemiological parameters directly from viral sequence data. The method is based on phylogenetic analysis using a birth-death model (BDM) rather than the commonly used coalescent as the model for the epidemiological transmission of the pathogen. Using the BDM has the advantage that transmission and death rates are estimated independently and therefore enables for the first time the estimation of the basic reproductive number of the pathogen using only sequence data, without further assumptions like the average duration of infection. We apply the method to genetic data of the HIV-1 epidemic in Switzerland.
Resumo:
Diagnosis Related Groups (DRG) are frequently used to standardize the comparison of consumption variables, such as length of stay (LOS). In order to be reliable, this comparison must control for the presence of outliers, i.e. values far removed from the pattern set by the majority of the data. Indeed, outliers can distort the usual statistical summaries, such as means and variances. A common practice is to trim LOS values according to various empirical rules, but there is little theoretical support for choosing between alternative procedures. This pilot study explores the possibility of describing LOS distributions with parametric models which provide the necessary framework for the use of robust methods.
Resumo:
Geoelectrical techniques are widely used to monitor groundwater processes, while surprisingly few studies have considered audio (AMT) and radio (RMT) magnetotellurics for such purposes. In this numerical investigation, we analyze to what extent inversion results based on AMT and RMT monitoring data can be improved by (1) time-lapse difference inversion; (2) incorporation of statistical information about the expected model update (i.e., the model regularization is based on a geostatistical model); (3) using alternative model norms to quantify temporal changes (i.e., approximations of l(1) and Cauchy norms using iteratively reweighted least-squares), (4) constraining model updates to predefined ranges (i.e., using Lagrange Multipliers to only allow either increases or decreases of electrical resistivity with respect to background conditions). To do so, we consider a simple illustrative model and a more realistic test case related to seawater intrusion. The results are encouraging and show significant improvements when using time-lapse difference inversion with non l(2) model norms. Artifacts that may arise when imposing compactness of regions with temporal changes can be suppressed through inequality constraints to yield models without oscillations outside the true region of temporal changes. Based on these results, we recommend approximate l(1)-norm solutions as they can resolve both sharp and smooth interfaces within the same model. (C) 2012 Elsevier B.V. All rights reserved.
Resumo:
When continuous data are coded to categorical variables, two types of coding are possible: crisp coding in the form of indicator, or dummy, variables with values either 0 or 1; or fuzzy coding where each observation is transformed to a set of "degrees of membership" between 0 and 1, using co-called membership functions. It is well known that the correspondence analysis of crisp coded data, namely multiple correspondence analysis, yields principal inertias (eigenvalues) that considerably underestimate the quality of the solution in a low-dimensional space. Since the crisp data only code the categories to which each individual case belongs, an alternative measure of fit is simply to count how well these categories are predicted by the solution. Another approach is to consider multiple correspondence analysis equivalently as the analysis of the Burt matrix (i.e., the matrix of all two-way cross-tabulations of the categorical variables), and then perform a joint correspondence analysis to fit just the off-diagonal tables of the Burt matrix - the measure of fit is then computed as the quality of explaining these tables only. The correspondence analysis of fuzzy coded data, called "fuzzy multiple correspondence analysis", suffers from the same problem, albeit attenuated. Again, one can count how many correct predictions are made of the categories which have highest degree of membership. But here one can also defuzzify the results of the analysis to obtain estimated values of the original data, and then calculate a measure of fit in the familiar percentage form, thanks to the resultant orthogonal decomposition of variance. Furthermore, if one thinks of fuzzy multiple correspondence analysis as explaining the two-way associations between variables, a fuzzy Burt matrix can be computed and the same strategy as in the crisp case can be applied to analyse the off-diagonal part of this matrix. In this paper these alternative measures of fit are defined and applied to a data set of continuous meteorological variables, which are coded crisply and fuzzily into three categories. Measuring the fit is further discussed when the data set consists of a mixture of discrete and continuous variables.
Resumo:
This article reviews the methodology of the studies on drug utilization with particular emphasis on primary care. Population based studies of drug inappropriateness can be done with microdata from Health Electronic Records and e-prescriptions. Multilevel models estimate the influence of factors affecting the appropriateness of drug prescription at different hierarchical levels: patient, doctor, health care organization and regulatory environment. Work by the GIUMAP suggest that patient characteristics are the most important factor in the appropriateness of prescriptions with significant effects at the general practicioner level.
Resumo:
Experimental animal models offer possibilities of physiology knowledge, pathogenesis of disease and action of drugs that are directly related to quality nursing care. This integrative review describes the current state of the instrumental and ethical aspects of experimental research with animal models, including the main recommendations of ethics committees that focus on animal welfare and raises questions about the impact of their findings in nursing care. Data show that, in Brazil, the progress in ethics for the use of animals for scientific purposes was consolidated with Law No. 11.794/2008 establishing ethical procedures, attending health, genetic and experimental parameters. The application of ethics in handling of animals for scientific and educational purposes and obtaining consistent and quality data brings unquestionable contributions to the nurse, as they offer subsidies to relate pathophysiological mechanisms and the clinical aspect on the patient.
Resumo:
A Comparison of the Management Models of Protected Areas between China and the African South Region allows reading and evaluating the similarities and differences in the use of management model as a management tool for protected areas in China and South African Region. Specifically, some positive and negative features of the management approaches for the two regions. Secondary data was collected from various related literature such as policy documents, students‟ dissertations/thesis, scientific articles and magazines. Based on the method above, the study found out that China's first nature reserve was the Dingus Mountain Nature Reserve in Zhaoqing, Guangdong province established in 1956. By the end of 2005, about 2,349 nature reserves of various kinds were set up throughout the country, covering a total area of 149.95 million ha and accounting for 15 percent of the total land territory. The study further found that Southern Africa has approximately 4,390 protected areas out of 11487920 total land areas and Eastern Africa has approximately 1838144 protected areas, which is equivalent to 15.0% of the total land areas. South Africa in this region had its first declared natural park in 1926 after Paul Kruger (a war hero) had alerted the authorities of the extinguishing threat of some animal species of region.
Resumo:
This paper presents a review of methodology for semi-supervised modeling with kernel methods, when the manifold assumption is guaranteed to be satisfied. It concerns environmental data modeling on natural manifolds, such as complex topographies of the mountainous regions, where environmental processes are highly influenced by the relief. These relations, possibly regionalized and nonlinear, can be modeled from data with machine learning using the digital elevation models in semi-supervised kernel methods. The range of the tools and methodological issues discussed in the study includes feature selection and semisupervised Support Vector algorithms. The real case study devoted to data-driven modeling of meteorological fields illustrates the discussed approach.
Resumo:
During adolescence, cognitive abilities increase robustly. To search for possible related structural alterations of the cerebral cortex, we measured neuronal soma dimension (NSD = width times height), cortical thickness and neuronal densities in different types of neocortex in post-mortem brains of five 12-16 and five 17-24 year-olds (each 2F, 3M). Using a generalized mixed model analysis, mean normalized NSD comparing the age groups shows layer-specific change for layer 2 (p < .0001) and age-related differences between categorized type of cortex: primary/primary association cortex (BA 1, 3, 4, and 44) shows a generalized increase; higher-order regions (BA 9, 21, 39, and 45) also show increase in layers 2 and 5 but decrease in layers 3, 4, and 6 while limbic/orbital cortex (BA 23, 24, and 47) undergoes minor decrease (BA 1, 3, 4, and 44 vs. BA 9, 21, 39, and 45: p = .036 and BA 1, 3, 4, and 44 vs. BA 23, 24, and 47: p = .004). These data imply the operation of cortical layer- and type-specific processes of growth and regression adding new evidence that the human brain matures during adolescence not only functionally but also structurally.
Resumo:
We consider the application of normal theory methods to the estimation and testing of a general type of multivariate regressionmodels with errors--in--variables, in the case where various data setsare merged into a single analysis and the observable variables deviatepossibly from normality. The various samples to be merged can differ on the set of observable variables available. We show that there is a convenient way to parameterize the model so that, despite the possiblenon--normality of the data, normal--theory methods yield correct inferencesfor the parameters of interest and for the goodness--of--fit test. Thetheory described encompasses both the functional and structural modelcases, and can be implemented using standard software for structuralequations models, such as LISREL, EQS, LISCOMP, among others. An illustration with Monte Carlo data is presented.
Resumo:
This paper investigates the role of learning by private agents and the central bank(two-sided learning) in a New Keynesian framework in which both sides of the economyhave asymmetric and imperfect knowledge about the true data generating process. Weassume that all agents employ the data that they observe (which may be distinct fordifferent sets of agents) to form beliefs about unknown aspects of the true model ofthe economy, use their beliefs to decide on actions, and revise these beliefs througha statistical learning algorithm as new information becomes available. We study theshort-run dynamics of our model and derive its policy recommendations, particularlywith respect to central bank communications. We demonstrate that two-sided learningcan generate substantial increases in volatility and persistence, and alter the behaviorof the variables in the model in a significant way. Our simulations do not convergeto a symmetric rational expectations equilibrium and we highlight one source thatinvalidates the convergence results of Marcet and Sargent (1989). Finally, we identifya novel aspect of central bank communication in models of learning: communicationcan be harmful if the central bank's model is substantially mis-specified.
Resumo:
The growth of pharmaceutical expenditure and its prediction is a major concern for policy makers and health care managers. This paper explores different predictive models to estimate future drug expenses, using demographic and morbidity individual information from an integrated healthcare delivery organization in Catalonia for years 2002 and 2003. The morbidity information consists of codified health encounters grouped through the Clinical Risk Groups (CRGs). We estimate pharmaceutical costs using several model specifications, and CRGs as risk adjusters, providing an alternative way of obtaining high predictive power comparable to other estimations of drug expenditures in the literature. These results have clear implications for the use of risk adjustment and CRGs in setting the premiums for pharmaceutical benefits.
Resumo:
We evaluate conditional predictive densities for U.S. output growth and inflationusing a number of commonly used forecasting models that rely on a large number ofmacroeconomic predictors. More specifically, we evaluate how well conditional predictive densities based on the commonly used normality assumption fit actual realizationsout-of-sample. Our focus on predictive densities acknowledges the possibility that, although some predictors can improve or deteriorate point forecasts, they might have theopposite effect on higher moments. We find that normality is rejected for most modelsin some dimension according to at least one of the tests we use. Interestingly, however,combinations of predictive densities appear to be correctly approximated by a normaldensity: the simple, equal average when predicting output growth and Bayesian modelaverage when predicting inflation.
Resumo:
Low corporate taxes can help attract new firms. This is the main mechanism underpinning the standard 'race-to-the-bottom'view of tax competition. A recent theoretical literature has qualified this view by formalizing the argument that agglomeration forces can reduce firms' sensitivity to tax differentials across locations. We test this proposition using data on firm startups across Swiss municipalities. We find that, on average, high corporate income taxes do deter new firms, but that this relationship is significantly weaker in the most spatially concentrated sectors. Location choices of firms in sectors with an agglomeration intensity at the twentieth percentile of the sample distribution are estimated to be twice as responsive to a given difference in local corporate tax burdens as firms in sectors with an agglomeration intensity at the eightieth percentile. Hence, our analysis confirms the theoretical prediction: agglomeration economies can neutralize the impact of tax differentials on firms' location choices.