985 resultados para participatory methods
Resumo:
Flow cytometry (FCM) is emerging as an important tool in environmental microbiology. Although flow cytometry applications have to date largely been restricted to certain specialized fields of microbiology, such as the bacterial cell cycle and marine phytoplankton communities, technical advances in instrumentation and methodology are leading to its increased popularity and extending its range of applications. Here we will focus on a number of recent flow cytometry developments important for addressing questions in environmental microbiology. These include (i) the study of microbial physiology under environmentally relevant conditions, (ii) new methods to identify active microbial populations and to isolate previously uncultured microorganisms, and (iii) the development of high-throughput autofluorescence bioreporter assays
Resumo:
Elucidating the molecular and neural basis of complex social behaviors such as communal living, division of labor and warfare requires model organisms that exhibit these multi-faceted behavioral phenotypes. Social insects, such as ants, bees, wasps and termites, are attractive models to address this problem, with rich ecological and ethological foundations. However, their atypical systems of reproduction have hindered application of classical genetic approaches. In this review, we discuss how recent advances in social insect genomics, transcriptomics, and functional manipulations have enhanced our ability to observe and perturb gene expression, physiology and behavior in these species. Such developments begin to provide an integrated view of the molecular and cellular underpinnings of complex social behavior.
Resumo:
Recently, kernel-based Machine Learning methods have gained great popularity in many data analysis and data mining fields: pattern recognition, biocomputing, speech and vision, engineering, remote sensing etc. The paper describes the use of kernel methods to approach the processing of large datasets from environmental monitoring networks. Several typical problems of the environmental sciences and their solutions provided by kernel-based methods are considered: classification of categorical data (soil type classification), mapping of environmental and pollution continuous information (pollution of soil by radionuclides), mapping with auxiliary information (climatic data from Aral Sea region). The promising developments, such as automatic emergency hot spot detection and monitoring network optimization are discussed as well.
Resumo:
This paper presents 3-D brain tissue classificationschemes using three recent promising energy minimizationmethods for Markov random fields: graph cuts, loopybelief propagation and tree-reweighted message passing.The classification is performed using the well knownfinite Gaussian mixture Markov Random Field model.Results from the above methods are compared with widelyused iterative conditional modes algorithm. Theevaluation is performed on a dataset containing simulatedT1-weighted MR brain volumes with varying noise andintensity non-uniformities. The comparisons are performedin terms of energies as well as based on ground truthsegmentations, using various quantitative metrics.
Resumo:
The paper contrasts empirically the results of alternative methods for estimating thevalue and the depreciation of mineral resources. The historical data of Mexico andVenezuela, covering the period 1920s-1980s, is used to contrast the results of severalmethods. These are the present value, the net price method, the user cost method andthe imputed income method. The paper establishes that the net price and the user costare not competing methods as such, but alternative adjustments to different scenariosof closed and open economies. The results prove that the biases of the methods, ascommonly described in the theoretical literature, only hold under the most restrictedscenario of constant rents over time. It is argued that the difference between what isexpected to happen and what actually did happen is for the most part due to a missingvariable, namely technological change. This is an important caveat to therecommendations made based on these models.
Resumo:
Consider the problem of testing k hypotheses simultaneously. In this paper,we discuss finite and large sample theory of stepdown methods that providecontrol of the familywise error rate (FWE). In order to improve upon theBonferroni method or Holm's (1979) stepdown method, Westfall and Young(1993) make eective use of resampling to construct stepdown methods thatimplicitly estimate the dependence structure of the test statistics. However,their methods depend on an assumption called subset pivotality. The goalof this paper is to construct general stepdown methods that do not requiresuch an assumption. In order to accomplish this, we take a close look atwhat makes stepdown procedures work, and a key component is a monotonicityrequirement of critical values. By imposing such monotonicity on estimatedcritical values (which is not an assumption on the model but an assumptionon the method), it is demonstrated that the problem of constructing a validmultiple test procedure which controls the FWE can be reduced to the problemof contructing a single test which controls the usual probability of a Type 1error. This reduction allows us to draw upon an enormous resamplingliterature as a general means of test contruction.
Resumo:
Bacteria are generally difficult specimens to prepare for conventional resin section electron microscopy and mycobacteria, with their thick and complex cell envelope layers being especially prone to artefacts. Here we made a systematic comparison of different methods for preparing Mycobacterium smegmatis for thin section electron microscopy analysis. These methods were: (1) conventional preparation by fixatives and epoxy resins at ambient temperature. (2) Tokuyasu cryo-section of chemically fixed bacteria. (3) rapid freezing followed by freeze substitution and embedding in epoxy resin at room temperature or (4) combined with Lowicryl HM20 embedding and ultraviolet (UV) polymerization at low temperature and (5) CEMOVIS, or cryo electron microscopy of vitreous sections. The best preservation of bacteria was obtained with the cryo electron microscopy of vitreous sections method, as expected, especially with respect to the preservation of the cell envelope and lipid bodies. By comparison with cryo electron microscopy of vitreous sections both the conventional and Tokuyasu methods produced different, undesirable artefacts. The two different types of freeze-substitution protocols showed variable preservation of the cell envelope but gave acceptable preservation of the cytoplasm, but not lipid bodies, and bacterial DNA. In conclusion although cryo electron microscopy of vitreous sections must be considered the 'gold standard' among sectioning methods for electron microscopy, because it avoids solvents and stains, the use of optimally prepared freeze substitution also offers some advantages for ultrastructural analysis of bacteria.
Resumo:
The question of where retroviral DNA becomes integrated in chromosomes is important for understanding (i) the mechanisms of viral growth, (ii) devising new anti-retroviral therapy, (iii) understanding how genomes evolve, and (iv) developing safer methods for gene therapy. With the completion of genome sequences for many organisms, it has become possible to study integration targeting by cloning and sequencing large numbers of host-virus DNA junctions, then mapping the host DNA segments back onto the genomic sequence. This allows statistical analysis of the distribution of integration sites relative to the myriad types of genomic features that are also being mapped onto the sequence scaffold. Here we present methods for recovering and analyzing integration site sequences.
Resumo:
AbstractText BACKGROUND: Profiling sperm DNA present on vaginal swabs taken from rape victims often contributes to identifying and incarcerating rapists. Large amounts of the victim's epithelial cells contaminate the sperm present on swabs, however, and complicate this process. The standard method for obtaining relatively pure sperm DNA from a vaginal swab is to digest the epithelial cells with Proteinase K in order to solubilize the victim's DNA, and to then physically separate the soluble DNA from the intact sperm by pelleting the sperm, removing the victim's fraction, and repeatedly washing the sperm pellet. An alternative approach that does not require washing steps is to digest with Proteinase K, pellet the sperm, remove the victim's fraction, and then digest the residual victim's DNA with a nuclease. METHODS: The nuclease approach has been commercialized in a product, the Erase Sperm Isolation Kit (PTC Labs, Columbia, MO, USA), and five crime laboratories have tested it on semen-spiked female buccal swabs in a direct comparison with their standard methods. Comparisons have also been performed on timed post-coital vaginal swabs and evidence collected from sexual assault cases. RESULTS: For the semen-spiked buccal swabs, Erase outperformed the standard methods in all five laboratories and in most cases was able to provide a clean male profile from buccal swabs spiked with only 1,500 sperm. The vaginal swabs taken after consensual sex and the evidence collected from rape victims showed a similar pattern of Erase providing superior profiles. CONCLUSIONS: In all samples tested, STR profiles of the male DNA fractions obtained with Erase were as good as or better than those obtained using the standard methods.
Resumo:
Many multivariate methods that are apparently distinct can be linked by introducing oneor more parameters in their definition. Methods that can be linked in this way arecorrespondence analysis, unweighted or weighted logratio analysis (the latter alsoknown as "spectral mapping"), nonsymmetric correspondence analysis, principalcomponent analysis (with and without logarithmic transformation of the data) andmultidimensional scaling. In this presentation I will show how several of thesemethods, which are frequently used in compositional data analysis, may be linkedthrough parametrizations such as power transformations, linear transformations andconvex linear combinations. Since the methods of interest here all lead to visual mapsof data, a "movie" can be made where where the linking parameter is allowed to vary insmall steps: the results are recalculated "frame by frame" and one can see the smoothchange from one method to another. Several of these "movies" will be shown, giving adeeper insight into the similarities and differences between these methods.
Resumo:
Executive SummaryIn Nepal, landslides are one of the major natural hazards after epidemics, killing over 100 persons per year. However, this figure is an underreported reflection of the actual impact that landslides have on livelihoods and food security in rural Nepal. With predictions of more intense rainfall patterns, landslide occurrence in the Himalayas is likely to increase and continue to be one of the major impediments to development. Due to the remoteness of many localities and lack of resources, responsibilities for disaster preparedness and response in mountain areas usually lie with the communities themselves. Everyday life is full of risk in mountains of Nepal. This is why mountain populations, as well as other populations living in harsh conditions have developed a number of coping strategies for dealing with adverse situations. Perhaps due to the dispersed and remote nature of landslides in Nepal, there have been few studies on vulnerability, coping- and mitigation strategies of landslide affected populations. There are also few recommendations available to guide authorities and populations how to reduce losses due to landslides in Nepal, and even less so, how to operationalize resilience and vulnerability.Many policy makers, international donors, NGOs and national authorities are currently asking what investments are needed to increase the so-called 'resilience' of mountain populations to deal with climate risks. However, mountain populations are already quite resilient to seasonal fluctuations, temperature variations, rainfall patterns and market prices. In spite of their resilience, they continue to live in places at risk due to high vulnerability caused by structural inequalities: access to land, resources, markets, education. This interdisciplinary thesis examines the concept of resilience by questioning its usefulness and validity as the current goal of international development and disaster risk reduction policies, its conceptual limitations and its possible scope of action. The goal of this study is two-fold: to better define and distinguish factors and relationships between resilience, vulnerability, capacities and risk; and to test and improve a participatory methodology for evaluating landslide risk that can serve as a guidance tool for improving community-based disaster risk reduction. The objective is to develop a simple methodology that can be used by NGOs, local authorities and communities to reduce losses from landslides.Through its six case studies in Central-Eastern Nepal, this study explores the relation between resilience, vulnerability and landslide risk based on interdisciplinary methods, including geological assessments of landslides, semi-structured interviews, focus groups and participatory risk mapping. For comparison, the study sites were chosen in Tehrathum, Sunsari and Dolakha Districts of Central/Eastern Nepal, to reflect a variety of landslide types, from chronic to acute, and a variety of communities, from very marginalized to very high status. The study uses the Sustainable Livelihoods Approach as its conceptual basis, which is based on the notion that access and rights to resources (natural, human/institutional, economic, environmental, physical) are the basis for coping with adversity, such as landslides. The study is also intended as a contribution to the growing literature and practices on Community Based Disaster Risk Reduction specifically adapted to landslide- prone areas.In addition to the six case studies, results include an indicator based methodology for assessing and measuring vulnerability and resilience, a composite risk assessment methodology, a typology of coping strategies and risk perceptions and a thorough analysis of the relation between risk, vulnerability and resilience. The methodology forassessing vulnerability, resilience and risk is relatively cost-effective and replicable in a low-data environment. Perhaps the major finding is that resilience is a process that defines a community's (or system's) capacity to rebound following adversity but it does not necessarily reduce vulnerability or risk, which requires addressing more structural issues related to poverty. Therefore, conclusions include a critical view of resilience as a main goal of international development and disaster risk reduction policies. It is a useful concept in the context of recovery after a disaster but it needs to be addressed in parallel with vulnerability and risk.This research was funded by an interdisciplinary grant (#26083591) from the Swiss National Science Foundation for the period 2009-2011 and a seed grant from the Faculty of Geosciences and Environment at the University of Lausanne in 2008.Résumé en françaisAu Népal, les glissements de terrain sont un des aléas les plus dévastateurs après les épidémies, causant 100 morts par an. Pourtant, ce chiffre est une sous-estimation de l'impact réel de l'effet des glissements sur les moyens de subsistance et la sécurité alimentaire au Népal. Avec des prévisions de pluies plus intenses, l'occurrence des glissements dans les Himalayas augmente et présente un obstacle au développement. Du fait de l'éloignement et du manque de ressources dans les montagnes au Népal, la responsabilité de la préparation et la réponse aux catastrophes se trouve chez les communautés elles-mêmes. Le risque fait partie de la vie quotidienne dans les montagnes du Népal. C'est pourquoi les populations montagnardes, comme d'autres populations vivant dans des milieux contraignants, ont développé des stratégies pour faire face aux situations défavorables. Peu d'études existent sur la vulnérabilité, ceci étant probablement dû à l'éloignement et pourtant, les stratégies d'adaptation et de mitigation des populations touchées par des glissements au Népal existent.Beaucoup de décideurs politiques, bailleurs de fonds, ONG et autorités nationales se demandent quels investissements sont nécessaires afin d'augmenter la 'resilience' des populations de montagne pour faire face aux changements climatiques. Pourtant, ces populations sont déjà résilientes aux fluctuations des saisons, des variations de température, des pluies et des prix des marchés. En dépit de leur résilience, ils continuent de vivre dans des endroits à fort risque à cause des vulnérabilités créées par les inégalités structurelles : l'accès à la terre, aux ressources, aux marchés et à l'éducation. Cette thèse interdisciplinaire examine le concept de la résilience en mettant en cause son utilité et sa validité en tant que but actuel des politiques internationales de développement et de réduction des risques, ainsi que ses limitations conceptuelles et ses possibles champs d'action. Le but de cette étude est double : mieux définir et distinguer les facteurs et relations entre la résilience, la vulnérabilité, les capacités et le risque ; Et tester et améliorer une méthode participative pour évaluer le risque des glissements qui peut servir en tant qu'outil indicatif pour améliorer la réduction des risques des communautés. Le but est de développer une méthodologie simple qui peut être utilisée par des ONG, autorités locales et communautés pour réduire les pertes dues aux glissements.A travers les études de cas au centre-est du Népal, cette étude explore le rapport entre la résilience, la vulnérabilité et les glissements basée sur des méthodes interdisciplinaires ; Y sont inclus des évaluations géologiques des glissements, des entretiens semi-dirigés, des discussions de groupes et des cartes de risques participatives. Pour la comparaison, les zones d'études ont été sélectionnées dans les districts de Tehrathum, Sunsari et Dolakha dans le centre-est du Népal, afin de refléter différents types de glissements, de chroniques à urgents, ainsi que différentes communautés, variant de très marginalisées à très haut statut. Pour son cadre conceptuel, cette étude s'appuie sur l'approche de moyens de subsistance durable, qui est basée sur les notions d'accès et de droit aux ressources (naturelles, humaines/institutionnelles, économiques, environnementales, physiques) et qui sont le minimum pour faire face à des situations difficiles, comme des glissements. Cette étude se veut aussi une contribution à la littérature et aux pratiques en croissantes sur la réduction des risques communautaires, spécifiquement adaptées aux zones affectées par des glissements.En plus des six études de cas, les résultats incluent une méthodologie basée sur des indicateurs pour évaluer et mesurer la vulnérabilité et la résilience, une méthodologie sur le risque composé, une typologie de stratégies d'adaptation et perceptions des risques ainsi qu'une analyse fondamentale de la relation entre risque, vulnérabilité et résilience. Les méthodologies pour l'évaluation de la vulnérabilité, de la résilience et du risque sont relativement peu coûteuses et reproductibles dans des endroits avec peu de données disponibles. Le résultat probablement le plus pertinent est que la résilience est un processus qui définit la capacité d'une communauté (ou d'un système) à rebondir suite à une situation défavorable, mais qui ne réduit pas forcement la vulnérabilité ou le risque, et qui requiert une approche plus fondamentale s'adressant aux questions de pauvreté. Les conclusions incluent une vue critique de la résilience comme but principal des politiques internationales de développement et de réduction des risques. C'est un concept utile dans le contexte de la récupération après une catastrophe mais il doit être pris en compte au même titre que la vulnérabilité et le risque.Cette recherche a été financée par un fonds interdisciplinaire (#26083591) du Fonds National Suisse pour la période 2009-2011 et un fonds de préparation de recherches par la Faculté des Géosciences et Environnement à l'Université de Lausanne en 2008.
Resumo:
Nowadays, genome-wide association studies (GWAS) and genomic selection (GS) methods which use genome-wide marker data for phenotype prediction are of much potential interest in plant breeding. However, to our knowledge, no studies have been performed yet on the predictive ability of these methods for structured traits when using training populations with high levels of genetic diversity. Such an example of a highly heterozygous, perennial species is grapevine. The present study compares the accuracy of models based on GWAS or GS alone, or in combination, for predicting simple or complex traits, linked or not with population structure. In order to explore the relevance of these methods in this context, we performed simulations using approx 90,000 SNPs on a population of 3,000 individuals structured into three groups and corresponding to published diversity grapevine data. To estimate the parameters of the prediction models, we defined four training populations of 1,000 individuals, corresponding to these three groups and a core collection. Finally, to estimate the accuracy of the models, we also simulated four breeding populations of 200 individuals. Although prediction accuracy was low when breeding populations were too distant from the training populations, high accuracy levels were obtained using the sole core-collection as training population. The highest prediction accuracy was obtained (up to 0.9) using the combined GWAS-GS model. We thus recommend using the combined prediction model and a core-collection as training population for grapevine breeding or for other important economic crops with the same characteristics.