1000 resultados para Habeas Data
Resumo:
Background Multiple logistic regression is precluded from many practical applications in ecology that aim to predict the geographic distributions of species because it requires absence data, which are rarely available or are unreliable. In order to use multiple logistic regression, many studies have simulated "pseudo-absences" through a number of strategies, but it is unknown how the choice of strategy influences models and their geographic predictions of species. In this paper we evaluate the effect of several prevailing pseudo-absence strategies on the predictions of the geographic distribution of a virtual species whose "true" distribution and relationship to three environmental predictors was predefined. We evaluated the effect of using a) real absences b) pseudo-absences selected randomly from the background and c) two-step approaches: pseudo-absences selected from low suitability areas predicted by either Ecological Niche Factor Analysis: (ENFA) or BIOCLIM. We compared how the choice of pseudo-absence strategy affected model fit, predictive power, and information-theoretic model selection results. Results Models built with true absences had the best predictive power, best discriminatory power, and the "true" model (the one that contained the correct predictors) was supported by the data according to AIC, as expected. Models based on random pseudo-absences had among the lowest fit, but yielded the second highest AUC value (0.97), and the "true" model was also supported by the data. Models based on two-step approaches had intermediate fit, the lowest predictive power, and the "true" model was not supported by the data. Conclusion If ecologists wish to build parsimonious GLM models that will allow them to make robust predictions, a reasonable approach is to use a large number of randomly selected pseudo-absences, and perform model selection based on an information theoretic approach. However, the resulting models can be expected to have limited fit.
Resumo:
It has been demonstrated in earlier studies that patients with a cochlear implant have increased abilities for audio-visual integration because the crude information transmitted by the cochlear implant requires the persistent use of the complementary speech information from the visual channel. The brain network for these abilities needs to be clarified. We used an independent components analysis (ICA) of the activation (H2 (15) O) positron emission tomography data to explore occipito-temporal brain activity in post-lingually deaf patients with unilaterally implanted cochlear implants at several months post-implantation (T1), shortly after implantation (T0) and in normal hearing controls. In between-group analysis, patients at T1 had greater blood flow in the left middle temporal cortex as compared with T0 and normal hearing controls. In within-group analysis, patients at T0 had a task-related ICA component in the visual cortex, and patients at T1 had one task-related ICA component in the left middle temporal cortex and the other in the visual cortex. The time courses of temporal and visual activities during the positron emission tomography examination at T1 were highly correlated, meaning that synchronized integrative activity occurred. The greater involvement of the visual cortex and its close coupling with the temporal cortex at T1 confirm the importance of audio-visual integration in more experienced cochlear implant subjects at the cortical level.
Resumo:
Spatial data on species distributions are available in two main forms, point locations and distribution maps (polygon ranges and grids). The first are often temporally and spatially biased, and too discontinuous, to be useful (untransformed) in spatial analyses. A variety of modelling approaches are used to transform point locations into maps. We discuss the attributes that point location data and distribution maps must satisfy in order to be useful in conservation planning. We recommend that before point location data are used to produce and/or evaluate distribution models, the dataset should be assessed under a set of criteria, including sample size, age of data, environmental/geographical coverage, independence, accuracy, time relevance and (often forgotten) representation of areas of permanent and natural presence of the species. Distribution maps must satisfy additional attributes if used for conservation analyses and strategies, including minimizing commission and omission errors, credibility of the source/assessors and availability for public screening. We review currently available databases for mammals globally and show that they are highly variable in complying with these attributes. The heterogeneity and weakness of spatial data seriously constrain their utility to global and also sub-global scale conservation analyses.
Resumo:
Human T-cell lymphotropic virus type 1 (HTLV-1) is mainly associated with two diseases: tropical spastic paraparesis/HTLV-1-associated myelopathy (TSP/HAM) and adult T-cell leukaemia/lymphoma. This retrovirus infects five-10 million individuals throughout the world. Previously, we developed a database that annotates sequence data from GenBank and the present study aimed to describe the clinical, molecular and epidemiological scenarios of HTLV-1 infection through the stored sequences in this database. A total of 2,545 registered complete and partial sequences of HTLV-1 were collected and 1,967 (77.3%) of those sequences represented unique isolates. Among these isolates, 93% contained geographic origin information and only 39% were related to any clinical status. A total of 1,091 sequences contained information about the geographic origin and viral subtype and 93% of these sequences were identified as subtype “a”. Ethnicity data are very scarce. Regarding clinical status data, 29% of the sequences were generated from TSP/HAM and 67.8% from healthy carrier individuals. Although the data mining enabled some inferences about specific aspects of HTLV-1 infection to be made, due to the relative scarcity of data of available sequences, it was not possible to delineate a global scenario of HTLV-1 infection.
Resumo:
Position du problème: La mise en place de la tarification à l'activité pour les hôpitaux de court séjour pourrait entraîner une diminution des durées de séjour pour raisons financières. L'impact potentiel de ce phénomène sur la qualité des soins n'est pas connu. Les réadmissions identifiées à l'aide des données administratives hospitalières sont, pour certaines situations cliniques, des indicateurs de qualité des soins valides. Méthode: Étude rétrospective du lien entre la durée de séjour et la survenue de réadmissions imprévues liées au séjour initial, pour les cholécystectomies simples et les accouchements par voie basse sans complication, à partir des données du programme de médicalisation des systèmes d'information de l'Assistance publique-Hôpitaux de Paris des années 2002 à 2005. Résultats: Pour les deux procédures, la probabilité de réadmission suit une courbe en " J ". Après ajustement sur l'âge, le sexe, les comorbidités associées, l'hôpital et l'année d'admission, la probabilité de réadmission est plus élevée pour les durées de séjour les plus courtes : pour les cholécystectomies, odds ratio : 6,03 [IC95 % : 2,67-13,59] pour les hospitalisations d'un jour versus trois jours ; pour les accouchements, odds ratio : 1,74 [IC95 % : 1,05-2,91] pour les hospitalisations de deux jours versus trois jours. Conclusion: Pour deux pathologies communes, les durées de séjour les plus courtes sont associées à des probabilités de réadmission plus élevées. L'utilisation routinière des données du programme de médicalisation des systèmes d'information peut permettre d'assurer le suivi de la relation entre la réduction de la durée de séjour et les réadmissions. The prospective payment system for the French short-stay hospitals creates a financial incentive to reduce length of stay. The potential impact of the resulting decrease in length of stay on the quality of healthcare is unknown. Readmission rates are valid outcome indicators for some clinical procedures. Methods: Retrospective study of the association between length of stay and unplanned readmissions related to the initial stay, for two procedures: cholecystectomy and vaginal delivery. Data: Administrative diagnosis-related groups database of "Assistance publique-Hopitaux de Paris", a large teaching hospital, for years 2002 to 2005. Results: The risk of readmission according to length of stay, taking age, sex, comorbidity, hospital and year of admission into account, followed a J-shaped curve for both procedures. The probability of readmission was higher for very short stays, with odds ratios and 95% confidence intervals of 6.03 [2.67-13.59] for cholecystectomies (1- versus 3-night stays), and of 1.74 [1.05-2.91] for vaginal deliveries (2- versus 3-night stays). Conclusion: For both procedures, the shortest lengths of stay are associated with a higher readmission probability. Suitable indicators derived from administrative databases would enable monitoring of the association between length of stay and readmissions.
Resumo:
Background: Understanding the true prevalence of lymphangioleiomyomatosis (LAM) is important in estimating disease burden and targeting specific interventions. As with all rare diseases, obtaining reliable epidemiological data is difficult and requires innovative approaches.Aim: To determine the prevalence and incidence of LAM using data from patient organizations in seven countries, and to use the extent to which the prevalence of LAM varies regionally and nationally to determine whether prevalence estimates are related to health-care provision.Methods: Numbers of women with LAM were obtained from patient groups and national databases from seven countries (n = 1001). Prevalence was calculated for regions within countries using female population figures from census data. Incidence estimates were calculated for the USA, UK and Switzerland. Regional variation in prevalence and changes in incidence over time were analysed using Poisson regression and linear regression.Results: Prevalence of LAM in the seven countries ranged from 3.4 to 7.8/million women with significant variation, both between countries and between states in the USA. This variation did not relate to the number of pulmonary specialists in the region nor the percentage of population with health insurance, but suggests a large number of patients remain undiagnosed. The incidence of LAM from 2004 to 2008 ranged from 0.23 to 0.31/million women/per year in the USA, UK and Switzerland.Conclusions: Using this method, we have found that the prevalence of LAM is higher than that previously recorded and that many patients with LAM are undiagnosed.
Resumo:
The CIAOW study (Complicated intra-abdominal infections worldwide observational study) is a multicenter observational study underwent in 68 medical institutions worldwide during a six-month study period (October 2012-March 2013). The study included patients older than 18 years undergoing surgery or interventional drainage to address complicated intra-abdominal infections (IAIs). 1898 patients with a mean age of 51.6 years (range 18-99) were enrolled in the study. 777 patients (41%) were women and 1,121 (59%) were men. Among these patients, 1,645 (86.7%) were affected by community-acquired IAIs while the remaining 253 (13.3%) suffered from healthcare-associated infections. Intraperitoneal specimens were collected from 1,190 (62.7%) of the enrolled patients. 827 patients (43.6%) were affected by generalized peritonitis while 1071 (56.4%) suffered from localized peritonitis or abscesses. The overall mortality rate was 10.5% (199/1898). According to stepwise multivariate analysis (PR = 0.005 and PE = 0.001), several criteria were found to be independent variables predictive of mortality, including patient age (OR = 1.1; 95%CI = 1.0-1.1; p < 0.0001), the presence of small bowel perforation (OR = 2.8; 95%CI = 1.5-5.3; p < 0.0001), a delayed initial intervention (a delay exceeding 24 hours) (OR = 1.8; 95%CI = 1.5-3.7; p < 0.0001), ICU admission (OR = 5.9; 95%CI = 3.6-9.5; p < 0.0001) and patient immunosuppression (OR = 3.8; 95%CI = 2.1-6.7; p < 0.0001).
Resumo:
As stated in Aitchison (1986), a proper study of relative variation in a compositional data set should be based on logratios, and dealing with logratios excludes dealing with zeros. Nevertheless, it is clear that zero observations might be present in real data sets, either because the corresponding part is completelyabsent –essential zeros– or because it is below detection limit –rounded zeros. Because the second kind of zeros is usually understood as “a trace too small to measure”, it seems reasonable to replace them by a suitable small value, and this has been the traditional approach. As stated, e.g. by Tauber (1999) and byMartín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000), the principal problem in compositional data analysis is related to rounded zeros. One should be careful to use a replacement strategy that does not seriously distort the general structure of the data. In particular, the covariance structure of the involvedparts –and thus the metric properties– should be preserved, as otherwise further analysis on subpopulations could be misleading. Following this point of view, a non-parametric imputation method isintroduced in Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000). This method is analyzed in depth by Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2003) where it is shown that thetheoretical drawbacks of the additive zero replacement method proposed in Aitchison (1986) can be overcome using a new multiplicative approach on the non-zero parts of a composition. The new approachhas reasonable properties from a compositional point of view. In particular, it is “natural” in the sense thatit recovers the “true” composition if replacement values are identical to the missing values, and it is coherent with the basic operations on the simplex. This coherence implies that the covariance structure of subcompositions with no zeros is preserved. As a generalization of the multiplicative replacement, in thesame paper a substitution method for missing values on compositional data sets is introduced
Resumo:
The main objective of this paper aims at developing a methodology that takes into account the human factor extracted from the data base used by the recommender systems, and which allow to resolve the specific problems of prediction and recommendation. In this work, we propose to extract the user's human values scale from the data base of the users, to improve their suitability in open environments, such as the recommender systems. For this purpose, the methodology is applied with the data of the user after interacting with the system. The methodology is exemplified with a case study
Resumo:
In the eighties, John Aitchison (1986) developed a new methodological approach for the statistical analysis of compositional data. This new methodology was implemented in Basic routines grouped under the name CODA and later NEWCODA inMatlab (Aitchison, 1997). After that, several other authors have published extensions to this methodology: Marín-Fernández and others (2000), Barceló-Vidal and others (2001), Pawlowsky-Glahn and Egozcue (2001, 2002) and Egozcue and others (2003). (...)
Resumo:
The log-ratio methodology makes available powerful tools for analyzing compositionaldata. Nevertheless, the use of this methodology is only possible for those data setswithout null values. Consequently, in those data sets where the zeros are present, aprevious treatment becomes necessary. Last advances in the treatment of compositionalzeros have been centered especially in the zeros of structural nature and in the roundedzeros. These tools do not contemplate the particular case of count compositional datasets with null values. In this work we deal with \count zeros" and we introduce atreatment based on a mixed Bayesian-multiplicative estimation. We use the Dirichletprobability distribution as a prior and we estimate the posterior probabilities. Then weapply a multiplicative modi¯cation for the non-zero values. We present a case studywhere this new methodology is applied.Key words: count data, multiplicative replacement, composition, log-ratio analysis
Resumo:
In a seminal paper, Aitchison and Lauder (1985) introduced classical kernel densityestimation techniques in the context of compositional data analysis. Indeed, they gavetwo options for the choice of the kernel to be used in the kernel estimator. One ofthese kernels is based on the use the alr transformation on the simplex SD jointly withthe normal distribution on RD-1. However, these authors themselves recognized thatthis method has some deficiencies. A method for overcoming these dificulties based onrecent developments for compositional data analysis and multivariate kernel estimationtheory, combining the ilr transformation with the use of the normal density with a fullbandwidth matrix, was recently proposed in Martín-Fernández, Chacón and Mateu-Figueras (2006). Here we present an extensive simulation study that compares bothmethods in practice, thus exploring the finite-sample behaviour of both estimators
Resumo:
The quantitative estimation of Sea Surface Temperatures from fossils assemblages is afundamental issue in palaeoclimatic and paleooceanographic investigations. TheModern Analogue Technique, a widely adopted method based on direct comparison offossil assemblages with modern coretop samples, was revised with the aim ofconforming it to compositional data analysis. The new CODAMAT method wasdeveloped by adopting the Aitchison metric as distance measure. Modern coretopdatasets are characterised by a large amount of zeros. The zero replacement was carriedout by adopting a Bayesian approach to the zero replacement, based on a posteriorestimation of the parameter of the multinomial distribution. The number of modernanalogues from which reconstructing the SST was determined by means of a multipleapproach by considering the Proxies correlation matrix, Standardized Residual Sum ofSquares and Mean Squared Distance. This new CODAMAT method was applied to theplanktonic foraminiferal assemblages of a core recovered in the Tyrrhenian Sea.Kew words: Modern analogues, Aitchison distance, Proxies correlation matrix,Standardized Residual Sum of Squares
Resumo:
Summary : 1. Measuring health literacy in Switzerland: a review of six surveys: 1.1 Comparison of questionnaires - 1.2 Measures of health literacy in Switzerland - 1.3 Discussion of Swiss data on HL - 1.4 Description of the six surveys: 1.4.1 Current health trends and health literacy in the Swiss population (gfs-UNIVOX), 1.4.2 Nutrition, physical exercise and body weight : opinions and perceptions of the Swiss population (USI), 1.4.3 Health Literacy in Switzerland (ISPMZ), 1.4.4 Swiss Health Survey (SHS), 1.4.5 Survey of Health, Ageing and Retirement in Europe (SHARE), 1.4.6 Adult literacy and life skills survey (ALL). - 2 . Economic costs of low health literacy in Switzerland: a rough calculation. Appendix: Screenshots cost model
Resumo:
The 2008 Data Fusion Contest organized by the IEEE Geoscience and Remote Sensing Data Fusion Technical Committee deals with the classification of high-resolution hyperspectral data from an urban area. Unlike in the previous issues of the contest, the goal was not only to identify the best algorithm but also to provide a collaborative effort: The decision fusion of the best individual algorithms was aiming at further improving the classification performances, and the best algorithms were ranked according to their relative contribution to the decision fusion. This paper presents the five awarded algorithms and the conclusions of the contest, stressing the importance of decision fusion, dimension reduction, and supervised classification methods, such as neural networks and support vector machines.