856 resultados para Population set-based methods
Resumo:
INTRODUCTION: We present the protocol of a large population-based case-control study of 5 common tumors in Spain (MCC-Spain) that evaluates environmental exposures and genetic factors. METHODS: Between 2008-2013, 10,183 persons aged 20-85 years were enrolled in 23 hospitals and primary care centres in 12 Spanish provinces including 1,115 cases of a new diagnosis of prostate cancer, 1,750 of breast cancer, 2,171 of colorectal cancer, 492 of gastro-oesophageal cancer, 554 cases of chronic lymphocytic leukaemia (CLL) and 4,101 population-based controls matched by frequency to cases by age, sex and region of residence. Participation rates ranged from 57% (stomach cancer) to 87% (CLL cases) and from 30% to 77% in controls. Participants completed a face-to-face computerized interview on sociodemographic factors, environmental exposures, occupation, medication, lifestyle, and personal and family medical history. In addition, participants completed a self-administered food-frequency questionnaire and telephone interviews. Blood samples were collected from 76% of participants while saliva samples were collected in CLL cases and participants refusing blood extractions. Clinical information was recorded for cases and paraffin blocks and/or fresh tumor samples are available in most collaborating hospitals. Genotyping was done through an exome array enriched with genetic markers in specific pathways. Multiple analyses are planned to assess the association of environmental, personal and genetic risk factors for each tumor and to identify pleiotropic effects. DISCUSSION: This study, conducted within the Spanish Consortium for Biomedical Research in Epidemiology & Public Health (CIBERESP), is a unique initiative to evaluate etiological factors for common cancers and will promote cancer research and prevention in Spain.
Resumo:
Machine learning provides tools for automated construction of predictive models in data intensive areas of engineering and science. The family of regularized kernel methods have in the recent years become one of the mainstream approaches to machine learning, due to a number of advantages the methods share. The approach provides theoretically well-founded solutions to the problems of under- and overfitting, allows learning from structured data, and has been empirically demonstrated to yield high predictive performance on a wide range of application domains. Historically, the problems of classification and regression have gained the majority of attention in the field. In this thesis we focus on another type of learning problem, that of learning to rank. In learning to rank, the aim is from a set of past observations to learn a ranking function that can order new objects according to how well they match some underlying criterion of goodness. As an important special case of the setting, we can recover the bipartite ranking problem, corresponding to maximizing the area under the ROC curve (AUC) in binary classification. Ranking applications appear in a large variety of settings, examples encountered in this thesis include document retrieval in web search, recommender systems, information extraction and automated parsing of natural language. We consider the pairwise approach to learning to rank, where ranking models are learned by minimizing the expected probability of ranking any two randomly drawn test examples incorrectly. The development of computationally efficient kernel methods, based on this approach, has in the past proven to be challenging. Moreover, it is not clear what techniques for estimating the predictive performance of learned models are the most reliable in the ranking setting, and how the techniques can be implemented efficiently. The contributions of this thesis are as follows. First, we develop RankRLS, a computationally efficient kernel method for learning to rank, that is based on minimizing a regularized pairwise least-squares loss. In addition to training methods, we introduce a variety of algorithms for tasks such as model selection, multi-output learning, and cross-validation, based on computational shortcuts from matrix algebra. Second, we improve the fastest known training method for the linear version of the RankSVM algorithm, which is one of the most well established methods for learning to rank. Third, we study the combination of the empirical kernel map and reduced set approximation, which allows the large-scale training of kernel machines using linear solvers, and propose computationally efficient solutions to cross-validation when using the approach. Next, we explore the problem of reliable cross-validation when using AUC as a performance criterion, through an extensive simulation study. We demonstrate that the proposed leave-pair-out cross-validation approach leads to more reliable performance estimation than commonly used alternative approaches. Finally, we present a case study on applying machine learning to information extraction from biomedical literature, which combines several of the approaches considered in the thesis. The thesis is divided into two parts. Part I provides the background for the research work and summarizes the most central results, Part II consists of the five original research articles that are the main contribution of this thesis.
Resumo:
PURPOSE: To analyze the prevalence of and factors associated with fragility fractures in Brazilian women aged 50 years and older. METHODS: This cross-sectional population survey, conducted between May 10 and October 31, 2011, included 622 women aged >50 years living in a city in southeastern Brazil. A questionnaire was administered to each woman by a trained interviewer. The associations between the occurrence of a fragility fracture after age 50 years and sociodemographic data, health-related habits and problems, self-perception of health and evaluation of functional capacity were determined by the χ2 test and Poisson regression using the backward selection criteria. RESULTS: The mean age of the 622 women was 64.1 years. The prevalence of fragility fractures was 10.8%, with 1.8% reporting hip fracture. In the final statistical model, a longer time since menopause (PR 1.03; 95%CI 1.01-1.05; p<0.01) and osteoporosis (PR 1.97; 95%CI 1.27-3.08; p<0.01) were associated with a higher prevalence of fractures. CONCLUSIONS: These findings may provide a better understanding of the risk factors associated with fragility fractures in Brazilian women and emphasize the importance of performing bone densitometry.
Resumo:
Feature selection plays an important role in knowledge discovery and data mining nowadays. In traditional rough set theory, feature selection using reduct - the minimal discerning set of attributes - is an important area. Nevertheless, the original definition of a reduct is restrictive, so in one of the previous research it was proposed to take into account not only the horizontal reduction of information by feature selection, but also a vertical reduction considering suitable subsets of the original set of objects. Following the work mentioned above, a new approach to generate bireducts using a multi--objective genetic algorithm was proposed. Although the genetic algorithms were used to calculate reduct in some previous works, we did not find any work where genetic algorithms were adopted to calculate bireducts. Compared to the works done before in this area, the proposed method has less randomness in generating bireducts. The genetic algorithm system estimated a quality of each bireduct by values of two objective functions as evolution progresses, so consequently a set of bireducts with optimized values of these objectives was obtained. Different fitness evaluation methods and genetic operators, such as crossover and mutation, were applied and the prediction accuracies were compared. Five datasets were used to test the proposed method and two datasets were used to perform a comparison study. Statistical analysis using the one-way ANOVA test was performed to determine the significant difference between the results. The experiment showed that the proposed method was able to reduce the number of bireducts necessary in order to receive a good prediction accuracy. Also, the influence of different genetic operators and fitness evaluation strategies on the prediction accuracy was analyzed. It was shown that the prediction accuracies of the proposed method are comparable with the best results in machine learning literature, and some of them outperformed it.
Resumo:
Naïvement perçu, le processus d’évolution est une succession d’événements de duplication et de mutations graduelles dans le génome qui mènent à des changements dans les fonctions et les interactions du protéome. La famille des hydrolases de guanosine triphosphate (GTPases) similaire à Ras constitue un bon modèle de travail afin de comprendre ce phénomène fondamental, car cette famille de protéines contient un nombre limité d’éléments qui diffèrent en fonctionnalité et en interactions. Globalement, nous désirons comprendre comment les mutations singulières au niveau des GTPases affectent la morphologie des cellules ainsi que leur degré d’impact sur les populations asynchrones. Mon travail de maîtrise vise à classifier de manière significative différents phénotypes de la levure Saccaromyces cerevisiae via l’analyse de plusieurs critères morphologiques de souches exprimant des GTPases mutées et natives. Notre approche à base de microscopie et d’analyses bioinformatique des images DIC (microscopie d’interférence différentielle de contraste) permet de distinguer les phénotypes propres aux cellules natives et aux mutants. L’emploi de cette méthode a permis une détection automatisée et une caractérisation des phénotypes mutants associés à la sur-expression de GTPases constitutivement actives. Les mutants de GTPases constitutivement actifs Cdc42 Q61L, Rho5 Q91H, Ras1 Q68L et Rsr1 G12V ont été analysés avec succès. En effet, l’implémentation de différents algorithmes de partitionnement, permet d’analyser des données qui combinent les mesures morphologiques de population native et mutantes. Nos résultats démontrent que l’algorithme Fuzzy C-Means performe un partitionnement efficace des cellules natives ou mutantes, où les différents types de cellules sont classifiés en fonction de plusieurs facteurs de formes cellulaires obtenus à partir des images DIC. Cette analyse démontre que les mutations Cdc42 Q61L, Rho5 Q91H, Ras1 Q68L et Rsr1 G12V induisent respectivement des phénotypes amorphe, allongé, rond et large qui sont représentés par des vecteurs de facteurs de forme distincts. Ces distinctions sont observées avec différentes proportions (morphologie mutante / morphologie native) dans les populations de mutants. Le développement de nouvelles méthodes automatisées d’analyse morphologique des cellules natives et mutantes s’avère extrêmement utile pour l’étude de la famille des GTPases ainsi que des résidus spécifiques qui dictent leurs fonctions et réseau d’interaction. Nous pouvons maintenant envisager de produire des mutants de GTPases qui inversent leur fonction en ciblant des résidus divergents. La substitution fonctionnelle est ensuite détectée au niveau morphologique grâce à notre nouvelle stratégie quantitative. Ce type d’analyse peut également être transposé à d’autres familles de protéines et contribuer de manière significative au domaine de la biologie évolutive.
Resumo:
Contexte De nombreuses études, utilisant des indicateurs de qualité variés, ont démontré que la qualité des soins pour la dépression n’est pas optimale en première ligne. Peu de ces études ont examiné les facteurs associés à la réception d’un traitement adéquat, en particulier en tenant compte simultanément des caractéristiques individuelles et organisationnelles. L'association entre un traitement adéquat pour un épisode dépressif majeur (EDM) et une amélioration des symptômes dépressifs n'est pas bien établie dans des conditions non-expérimentales. Les objectifs de cette étude étaient de : 1) réaliser une revue systématique des indicateurs mesurant la qualité du traitement de la dépression en première ligne ; 2) estimer la proportion de patients souffrant d’EDM qui reçoivent un traitement adéquat (selon les guides de pratique clinique) en première ligne ; 3) examiner les caractéristiques individuelles et organisationnelles associées à l’adéquation du traitement pour la dépression ; 4) examiner l'association entre un traitement minimalement adéquat au cours des 12 mois précédents et l'évolution des symptômes dépressifs à 6 et 12 mois. Méthodes La littérature sur la qualité du traitement de la dépression a été examinée en utilisant un ensemble de mots-clés (« depression », « depressive disorder », « quality », « treatment », « indicator », « adequacy », « adherence », « concordance », « clinical guideline » et « guideline ») et « 360search », un moteur de recherche fédérée. Les données proviennent d'une étude de cohorte incluant 915 adultes consultant un médecin généraliste, quel que soit le motif de consultation, répondant aux critères du DSM-IV pour l’EDM dans la dernière année, nichés dans 65 cliniques de première ligne au Québec, Canada. Des analyses multiniveaux ont été réalisées. Résultats Bien que majoritairement développés à partir de guides de pratique clinique, une grande variété d'indicateurs a été observée dans la revue systématique de littérature. La plupart des études retenues ont utilisé des indicateurs de qualité rudimentaires, surtout pour la psychothérapie. Les méthodes utilisées étaient très variées, limitant la comparabilité des résultats. Toutefois, quelque soit la méthode choisie, la plupart des études ont révélé qu’une grande proportion des personnes souffrant de dépression n’ont pas reçu de traitement minimalement adéquat en première ligne. Dans notre échantillon, l’adéquation était élevée (> 75 %) pour un tiers des indicateurs de qualité mesurés, mais était faible (< 60 %) pour près de la moitié des mesures. Un peu plus de la moitié de l'échantillon (52,2 %) a reçu au moins un traitement minimalement adéquat pour la dépression. Au niveau individuel, les jeunes adultes (18-24 ans) et les personnes de plus de 65 ans avaient une probabilité moins élevée de recevoir un traitement minimalement adéquat. Cette probabilité était plus élevée pour ceux qui ont un médecin de famille, une assurance complémentaire, un trouble anxieux comorbide et une dépression plus sévère. Au niveau des cliniques, la disponibilité de la psychothérapie sur place, l'utilisation d'algorithmes de traitement, et le mode de rémunération perçu comme adéquat étaient associés à plus de traitement adéquat. Les résultats ont également montré que 1) la réception d'au moins un traitement minimalement adéquat pour la dépression était associée à une plus grande amélioration des symptômes dépressifs à 6 et à 12 mois; 2) la pharmacothérapie adéquate et la psychothérapie adéquate étaient toutes deux associées à de plus grandes améliorations dans les symptômes dépressifs, et 3) l'association entre un traitement adéquat et l'amélioration des symptômes dépressifs varie en fonction de la sévérité des symptômes au moment de l'inclusion dans la cohorte, un niveau de symptômes plus élevé étant associé à une amélioration plus importante à 6 et à 12 mois. Conclusions Nos résultats suggèrent que des interventions sont nécessaires pour améliorer la qualité du traitement de la dépression en première ligne. Ces interventions devraient cibler des populations spécifiques (les jeunes adultes et les personnes âgées), améliorer l'accessibilité à la psychothérapie et à un médecin de famille, et soutenir les médecins de première ligne dans leur pratique clinique avec des patients souffrant de dépression de différentes façons, telles que le développement des connaissances pour traiter la dépression et l'adaptation du mode de rémunération. Cette étude montre également que le traitement adéquat de la dépression en première ligne est associé à une amélioration des symptômes dépressifs dans des conditions non-expérimentales.
Resumo:
Le problème de tournées de véhicules (VRP), introduit par Dantzig and Ramser en 1959, est devenu l'un des problèmes les plus étudiés en recherche opérationnelle, et ce, en raison de son intérêt méthodologique et de ses retombées pratiques dans de nombreux domaines tels que le transport, la logistique, les télécommunications et la production. L'objectif général du VRP est d'optimiser l'utilisation des ressources de transport afin de répondre aux besoins des clients tout en respectant les contraintes découlant des exigences du contexte d’application. Les applications réelles du VRP doivent tenir compte d’une grande variété de contraintes et plus ces contraintes sont nombreuse, plus le problème est difficile à résoudre. Les VRPs qui tiennent compte de l’ensemble de ces contraintes rencontrées en pratique et qui se rapprochent des applications réelles forment la classe des problèmes ‘riches’ de tournées de véhicules. Résoudre ces problèmes de manière efficiente pose des défis considérables pour la communauté de chercheurs qui se penchent sur les VRPs. Cette thèse, composée de deux parties, explore certaines extensions du VRP vers ces problèmes. La première partie de cette thèse porte sur le VRP périodique avec des contraintes de fenêtres de temps (PVRPTW). Celui-ci est une extension du VRP classique avec fenêtres de temps (VRPTW) puisqu’il considère un horizon de planification de plusieurs jours pendant lesquels les clients n'ont généralement pas besoin d’être desservi à tous les jours, mais plutôt peuvent être visités selon un certain nombre de combinaisons possibles de jours de livraison. Cette généralisation étend l'éventail d'applications de ce problème à diverses activités de distributions commerciales, telle la collecte des déchets, le balayage des rues, la distribution de produits alimentaires, la livraison du courrier, etc. La principale contribution scientifique de la première partie de cette thèse est le développement d'une méta-heuristique hybride dans la quelle un ensemble de procédures de recherche locales et de méta-heuristiques basées sur les principes de voisinages coopèrent avec un algorithme génétique afin d’améliorer la qualité des solutions et de promouvoir la diversité de la population. Les résultats obtenus montrent que la méthode proposée est très performante et donne de nouvelles meilleures solutions pour certains grands exemplaires du problème. La deuxième partie de cette étude a pour but de présenter, modéliser et résoudre deux problèmes riches de tournées de véhicules, qui sont des extensions du VRPTW en ce sens qu'ils incluent des demandes dépendantes du temps de ramassage et de livraison avec des restrictions au niveau de la synchronization temporelle. Ces problèmes sont connus respectivement sous le nom de Time-dependent Multi-zone Multi-Trip Vehicle Routing Problem with Time Windows (TMZT-VRPTW) et de Multi-zone Mult-Trip Pickup and Delivery Problem with Time Windows and Synchronization (MZT-PDTWS). Ces deux problèmes proviennent de la planification des opérations de systèmes logistiques urbains à deux niveaux. La difficulté de ces problèmes réside dans la manipulation de deux ensembles entrelacés de décisions: la composante des tournées de véhicules qui vise à déterminer les séquences de clients visités par chaque véhicule, et la composante de planification qui vise à faciliter l'arrivée des véhicules selon des restrictions au niveau de la synchronisation temporelle. Auparavant, ces questions ont été abordées séparément. La combinaison de ces types de décisions dans une seule formulation mathématique et dans une même méthode de résolution devrait donc donner de meilleurs résultats que de considérer ces décisions séparément. Dans cette étude, nous proposons des solutions heuristiques qui tiennent compte de ces deux types de décisions simultanément, et ce, d'une manière complète et efficace. Les résultats de tests expérimentaux confirment la performance de la méthode proposée lorsqu’on la compare aux autres méthodes présentées dans la littérature. En effet, la méthode développée propose des solutions nécessitant moins de véhicules et engendrant de moindres frais de déplacement pour effectuer efficacement la même quantité de travail. Dans le contexte des systèmes logistiques urbains, nos résultats impliquent une réduction de la présence de véhicules dans les rues de la ville et, par conséquent, de leur impact négatif sur la congestion et sur l’environnement.
Resumo:
Background: The most common application of imputation is to infer genotypes of a high-density panel of markers on animals that are genotyped for a low-density panel. However, the increase in accuracy of genomic predictions resulting from an increase in the number of markers tends to reach a plateau beyond a certain density. Another application of imputation is to increase the size of the training set with un-genotyped animals. This strategy can be particularly successful when a set of closely related individuals are genotyped. ----- Methods: Imputation on completely un-genotyped dams was performed using known genotypes from the sire of each dam, one offspring and the offspring’s sire. Two methods were applied based on either allele or haplotype frequencies to infer genotypes at ambiguous loci. Results of these methods and of two available software packages were compared. Quality of imputation under different population structures was assessed. The impact of using imputed dams to enlarge training sets on the accuracy of genomic predictions was evaluated for different populations, heritabilities and sizes of training sets. ----- Results: Imputation accuracy ranged from 0.52 to 0.93 depending on the population structure and the method used. The method that used allele frequencies performed better than the method based on haplotype frequencies. Accuracy of imputation was higher for populations with higher levels of linkage disequilibrium and with larger proportions of markers with more extreme allele frequencies. Inclusion of imputed dams in the training set increased the accuracy of genomic predictions. Gains in accuracy ranged from close to zero to 37.14%, depending on the simulated scenario. Generally, the larger the accuracy already obtained with the genotyped training set, the lower the increase in accuracy achieved by adding imputed dams. ----- Conclusions: Whenever a reference population resembling the family configuration considered here is available, imputation can be used to achieve an extra increase in accuracy of genomic predictions by enlarging the training set with completely un-genotyped dams. This strategy was shown to be particularly useful for populations with lower levels of linkage disequilibrium, for genomic selection on traits with low heritability, and for species or breeds for which the size of the reference population is limited.
Resumo:
1. Species-based indices are frequently employed as surrogates for wider biodiversity health and measures of environmental condition. Species selection is crucial in determining an indicators metric value and hence the validity of the interpretation of ecosystem condition and function it provides, yet an objective process to identify appropriate indicator species is frequently lacking. 2. An effective indicator needs to (i) be representative, reflecting the status of wider biodiversity; (ii) be reactive, acting as early-warning systems for detrimental changes in environmental conditions; (iii) respond to change in a predictable way. We present an objective, niche-based approach for species' selection, founded on a coarse categorisation of species' niche space and key resource requirements, which ensures the resultant indicator has these key attributes. 3. We use UK farmland birds as a case study to demonstrate this approach, identifying an optimal indicator set containing 12 species. In contrast to the 19 species included in the farmland bird index (FBI), a key UK biodiversity indicator that contributes to one of the UK Government's headline indicators of sustainability, the niche space occupied by these species fully encompasses that occupied by the wider community of 62 species. 4. We demonstrate that the response of these 12 species to land-use change is a strong correlate to that of the wider farmland bird community. Furthermore, the temporal dynamics of the index based on their population trends closely matches the population dynamics of the wider community. However, in both analyses, the magnitude of the change in our indicator was significantly greater, allowing this indicator to act as an early-warning system. 5. Ecological indicators are embedded in environmental management, sustainable development and biodiversity conservation policy and practice where they act as metrics against which progress towards national, regional and global targets can be measured. Adopting this niche-based approach for objective selection of indicator species will facilitate the development of sensitive and representative indices for a range of taxonomic groups, habitats and spatial scales.
Resumo:
Background: Exposure to solar ultraviolet-B (UV-B) radiation is a major source of vitamin D3. Chemistry climate models project decreases in ground-level solar erythemal UV over the current century. It is unclear what impact this will have on vitamin D status at the population level. The purpose of this study was to measure the association between ground-level solar UV-B and serum concentrations of 25-hydroxyvitamin D (25(OH)D) using a secondary analysis of the 2007 to 2009 Canadian Health Measures Survey (CHMS). Methods: Blood samples collected from individuals aged 12 to 79 years sampled across Canada were analyzed for 25(OH)D (n=4,398). Solar UV-B irradiance was calculated for the 15 CHMS collection sites using the Tropospheric Ultraviolet and Visible Radiation Model. Multivariable linear regression was used to evaluate the association between 25(OH)D and solar UV-B adjusted for other predictors and to explore effect modification. Results: Cumulative solar UV-B irradiance averaged over 91 days (91-day UV-B) prior to blood draw correlated significantly with 25(OH)D. Independent of other predictors, a 1 kJ/m 2 increase in 91-day UV-B was associated with a significant 0.5 nmol/L (95% CI 0.3-0.8) increase in mean 25(OH)D (P =0.0001). The relationship was stronger among younger individuals and those spending more time outdoors. Based on current projections of decreases in ground-level solar UV-B, we predict less than a 1 nmol/L decrease in mean 25(OH)D for the population. Conclusions: In Canada, cumulative exposure to ambient solar UV-B has a small but significant association with 25(OH)D concentrations. Public health messages to improve vitamin D status should target safe sun exposure with sunscreen use, and also enhanced dietary and supplemental intake and maintenance of a healthy body weight.
Resumo:
Crystallization must occur in honey in order to produce set or creamed honey; however, the process must occur in a controlled manner in order to obtain an acceptable product. As a consequence, reliable methods are needed to measure the crystal content of honey (φ expressed as kg crystal per kg honey), which can also be implemented with relative ease in industrial production facilities. Unfortunately, suitable methods do not currently exist. This article reports on the development of 2 independent offline methods to measure the crystal content in honey based on differential scanning calorimetry and high-performance liquid chromatography. The 2 methods gave highly consistent results on the basis of paired t-test involving 143 experimental points (P > 0.05, r**2 = 0.99). The crystal content also correlated with the relative viscosity, defined as the ratio of the viscosity of crystal containing honey to that of the same honey when all crystals are dissolved, giving the following correlation: μr = 1 + 1398.8∅**2.318. This correlation can be used to estimate the crystal content of honey in industrial production facilities. The crystal growth rate at a temperature of 14 ◦C—the normal crystallization temperature used in practice—was linear, and the growth rate also increased with the total glucose content in the honey.
Resumo:
Earthworms are important organisms in soil communities and so are used as model organisms in environmental risk assessments of chemicals. However current risk assessments of soil invertebrates are based on short-term laboratory studies, of limited ecological relevance, supplemented if necessary by site-specific field trials, which sometimes are challenging to apply across the whole agricultural landscape. Here, we investigate whether population responses to environmental stressors and pesticide exposure can be accurately predicted by combining energy budget and agent-based models (ABMs), based on knowledge of how individuals respond to their local circumstances. A simple energy budget model was implemented within each earthworm Eisenia fetida in the ABM, based on a priori parameter estimates. From broadly accepted physiological principles, simple algorithms specify how energy acquisition and expenditure drive life cycle processes. Each individual allocates energy between maintenance, growth and/or reproduction under varying conditions of food density, soil temperature and soil moisture. When simulating published experiments, good model fits were obtained to experimental data on individual growth, reproduction and starvation. Using the energy budget model as a platform we developed methods to identify which of the physiological parameters in the energy budget model (rates of ingestion, maintenance, growth or reproduction) are primarily affected by pesticide applications, producing four hypotheses about how toxicity acts. We tested these hypotheses by comparing model outputs with published toxicity data on the effects of copper oxychloride and chlorpyrifos on E. fetida. Both growth and reproduction were directly affected in experiments in which sufficient food was provided, whilst maintenance was targeted under food limitation. Although we only incorporate toxic effects at the individual level we show how ABMs can readily extrapolate to larger scales by providing good model fits to field population data. The ability of the presented model to fit the available field and laboratory data for E. fetida demonstrates the promise of the agent-based approach in ecology, by showing how biological knowledge can be used to make ecological inferences. Further work is required to extend the approach to populations of more ecologically relevant species studied at the field scale. Such a model could help extrapolate from laboratory to field conditions and from one set of field conditions to another or from species to species.
Resumo:
BACKGROUND AND OBJECTIVE: Given the role of uncoupling protein 2 (UCP2) in the accumulation of fat in the hepatocytes and in the enhancement of protective mechanisms in acute ethanol intake, we hypothesised that UCP2 polymorphisms are likely to cause liver disease through their interactions with obesity and alcohol intake. To test this hypothesis, we investigated the interaction between tagging polymorphisms in the UCP2 gene (rs2306819, rs599277 and rs659366), alcohol intake and obesity traits such as BMI and waist circumference (WC) on alanine aminotransferase (ALT) and gamma glutamyl transferase (GGT) in a large meta-analysis of data sets from three populations (n=20 242). DESIGN AND METHODS: The study populations included the Northern Finland Birth Cohort 1966 (n=4996), Netherlands Study of Depression and Anxiety (n=1883) and LifeLines Cohort Study (n=13 363). Interactions between the polymorphisms and obesity and alcohol intake on dichotomised ALT and GGT levels were assessed using logistic regression and the likelihood ratio test. RESULTS: In the meta-analysis of the three cohorts, none of the three UCP2 polymorphisms were associated with GGT or ALT levels. There was no evidence for interaction between the polymorphisms and alcohol intake on GGT and ALT levels. In contrast, the association of WC and BMI with GGT levels varied by rs659366 genotype (Pinteraction=0.03 and 0.007, respectively; adjusted for age, gender, high alcohol intake, diabetes, hypertension and serum lipid concentrations). CONCLUSION: In conclusion, our findings in 20 242 individuals suggest that UCP2 gene polymorphisms may cause liver dysfunction through the interaction with body fat rather than alcohol intake.
Resumo:
Background: Health care literature supports the development of accessible interventions that integrate behavioral economics, wearable devices, principles of evidence-based behavior change, and community support. However, there are limited real-world examples of large scale, population-based, member-driven reward platforms. Subsequently, a paucity of outcome data exists and health economic effects remain largely theoretical. To complicate matters, an emerging area of research is defining the role of Superusers, the small percentage of unusually engaged digital health participants who may influence other members. Objective: The objective of this preliminary study is to analyze descriptive data from GOODcoins, a self-guided, free-to-consumer engagement and rewards platform incentivizing walking, running and cycling. Registered members accessed the GOODcoins platform through PCs, tablets or mobile devices, and had the opportunity to sync wearables to track activity. Following registration, members were encouraged to join gamified group challenges and compare their progress with that of others. As members met challenge targets, they were rewarded with GOODcoins, which could be redeemed for planet- or people-friendly products. Methods: Outcome data were obtained from the GOODcoins custom SQL database. The reporting period was December 1, 2014 to May 1, 2015. Descriptive self-report data were analyzed using MySQL and MS Excel. Results: The study period includes data from 1298 users who were connected to an exercise tracking device. Females consisted of 52.6% (n=683) of the study population, 33.7% (n=438) were between the ages of 20-29, and 24.8% (n=322) were between the ages of 30-39. 77.5% (n=1006) of connected and active members met daily-recommended physical activity guidelines of 30 minutes, with a total daily average activity of 107 minutes (95% CI 90, 124). Of all connected and active users, 96.1% (n=1248) listed walking as their primary activity. For members who exchanged GOODcoins, the mean balance was 4,000 (95% CI 3850, 4150) at time of redemption, and 50.4% (n=61) of exchanges were for fitness or outdoor products, while 4.1% (n=5) were for food-related items. Participants were most likely to complete challenges when rewards were between 201-300 GOODcoins. Conclusions: The purpose of this study is to form a baseline for future research. Overall, results indicate that challenges and incentives may be effective for connected and active members, and may play a role in achieving daily-recommended activity guidelines. Registrants were typically younger, walking was the primary activity, and rewards were mainly exchanged for fitness or outdoor products. Remaining to be determined is whether members were already physically active at time of registration and are representative of healthy adherers, or were previously inactive and were incentivized to change their behavior. As challenges are gamified, there is an opportunity to investigate the role of superusers and healthy adherers, impacts on behavioral norms, and how cooperative games and incentives can be leveraged across stratified populations. Study limitations and future research agendas are discussed.
Resumo:
Objectives. To investigate health self-assessment and to estimate the prevalence of chronic diseases and recent illnesses in people with and without physical disabilities (PD) in the state of Sao Paulo, southeastern Brazil. Study design. A Cross-sectional study comprising two population-based health surveys conducted in 2002 and 2003. Methods. A total of 8317 persons (165 with PD) were interviewed in the two studies. Variables concerning to health self-assessment; chronic disease and recent illness were compared in the people with and without PD. Negative binomial regression was used in the analysis. Results. Subjects with PD more often assessed their health as poor/very poor compared to non-disabled ones. They reported more illnesses in the 15 days prior to interview as well as more chronic diseases (skin conditions, anaemia, chronic kidney disease, stroke, depression/anxiety, migraine/headache, pulmonary diseases, hypertension, diabetes, arthritis/arthrosis/rheumatic conditions and heart disease). This higher disease prevalence can be either attributed to disability itself or be associated to gender, age and schooling. Conclusions. Subjects with PD had more recent illnesses and chronic diseases and poorer health self-assessment than non-disabled ones. Age, gender, schooling and disability have individual roles in disease development among disabled people.