128 resultados para Rejection-sampling Algorithm
em Université de Lausanne, Switzerland
Resumo:
The geometry and connectivity of fractures exert a strong influence on the flow and transport properties of fracture networks. We present a novel approach to stochastically generate three-dimensional discrete networks of connected fractures that are conditioned to hydrological and geophysical data. A hierarchical rejection sampling algorithm is used to draw realizations from the posterior probability density function at different conditioning levels. The method is applied to a well-studied granitic formation using data acquired within two boreholes located 6 m apart. The prior models include 27 fractures with their geometry (position and orientation) bounded by information derived from single-hole ground-penetrating radar (GPR) data acquired during saline tracer tests and optical televiewer logs. Eleven cross-hole hydraulic connections between fractures in neighboring boreholes and the order in which the tracer arrives at different fractures are used for conditioning. Furthermore, the networks are conditioned to the observed relative hydraulic importance of the different hydraulic connections by numerically simulating the flow response. Among the conditioning data considered, constraints on the relative flow contributions were the most effective in determining the variability among the network realizations. Nevertheless, we find that the posterior model space is strongly determined by the imposed prior bounds. Strong prior bounds were derived from GPR measurements and helped to make the approach computationally feasible. We analyze a set of 230 posterior realizations that reproduce all data given their uncertainties assuming the same uniform transmissivity in all fractures. The posterior models provide valuable statistics on length scales and density of connected fractures, as well as their connectivity. In an additional analysis, effective transmissivity estimates of the posterior realizations indicate a strong influence of the DFN structure, in that it induces large variations of equivalent transmissivities between realizations. The transmissivity estimates agree well with previous estimates at the site based on pumping, flowmeter and temperature data.
Resumo:
SUMMARY : Eukaryotic DNA interacts with the nuclear proteins using non-covalent ionic interactions. Proteins can recognize specific nucleotide sequences based on the sterical interactions with the DNA and these specific protein-DNA interactions are the basis for many nuclear processes, e.g. gene transcription, chromosomal replication, and recombination. New technology termed ChIP-Seq has been recently developed for the analysis of protein-DNA interactions on a whole genome scale and it is based on immunoprecipitation of chromatin and high-throughput DNA sequencing procedure. ChIP-Seq is a novel technique with a great potential to replace older techniques for mapping of protein-DNA interactions. In this thesis, we bring some new insights into the ChIP-Seq data analysis. First, we point out to some common and so far unknown artifacts of the method. Sequence tag distribution in the genome does not follow uniform distribution and we have found extreme hot-spots of tag accumulation over specific loci in the human and mouse genomes. These artifactual sequence tags accumulations will create false peaks in every ChIP-Seq dataset and we propose different filtering methods to reduce the number of false positives. Next, we propose random sampling as a powerful analytical tool in the ChIP-Seq data analysis that could be used to infer biological knowledge from the massive ChIP-Seq datasets. We created unbiased random sampling algorithm and we used this methodology to reveal some of the important biological properties of Nuclear Factor I DNA binding proteins. Finally, by analyzing the ChIP-Seq data in detail, we revealed that Nuclear Factor I transcription factors mainly act as activators of transcription, and that they are associated with specific chromatin modifications that are markers of open chromatin. We speculate that NFI factors only interact with the DNA wrapped around the nucleosome. We also found multiple loci that indicate possible chromatin barrier activity of NFI proteins, which could suggest the use of NFI binding sequences as chromatin insulators in biotechnology applications. RESUME : L'ADN des eucaryotes interagit avec les protéines nucléaires par des interactions noncovalentes ioniques. Les protéines peuvent reconnaître les séquences nucléotidiques spécifiques basées sur l'interaction stérique avec l'ADN, et des interactions spécifiques contrôlent de nombreux processus nucléaire, p.ex. transcription du gène, la réplication chromosomique, et la recombinaison. Une nouvelle technologie appelée ChIP-Seq a été récemment développée pour l'analyse des interactions protéine-ADN à l'échelle du génome entier et cette approche est basée sur l'immuno-précipitation de la chromatine et sur la procédure de séquençage de l'ADN à haut débit. La nouvelle approche ChIP-Seq a donc un fort potentiel pour remplacer les anciennes techniques de cartographie des interactions protéine-ADN. Dans cette thèse, nous apportons de nouvelles perspectives dans l'analyse des données ChIP-Seq. Tout d'abord, nous avons identifié des artefacts très communs associés à cette méthode qui étaient jusqu'à présent insoupçonnés. La distribution des séquences dans le génome ne suit pas une distribution uniforme et nous avons constaté des positions extrêmes d'accumulation de séquence à des régions spécifiques, des génomes humains et de la souris. Ces accumulations des séquences artéfactuelles créera de faux pics dans toutes les données ChIP-Seq, et nous proposons différentes méthodes de filtrage pour réduire le nombre de faux positifs. Ensuite, nous proposons un nouvel échantillonnage aléatoire comme un outil puissant d'analyse des données ChIP-Seq, ce qui pourraient augmenter l'acquisition de connaissances biologiques à partir des données ChIP-Seq. Nous avons créé un algorithme d'échantillonnage aléatoire et nous avons utilisé cette méthode pour révéler certaines des propriétés biologiques importantes de protéines liant à l'ADN nommés Facteur Nucléaire I (NFI). Enfin, en analysant en détail les données de ChIP-Seq pour la famille de facteurs de transcription nommés Facteur Nucléaire I, nous avons révélé que ces protéines agissent principalement comme des activateurs de transcription, et qu'elles sont associées à des modifications de la chromatine spécifiques qui sont des marqueurs de la chromatine ouverte. Nous pensons que lés facteurs NFI interagir uniquement avec l'ADN enroulé autour du nucléosome. Nous avons également constaté plusieurs régions génomiques qui indiquent une éventuelle activité de barrière chromatinienne des protéines NFI, ce qui pourrait suggérer l'utilisation de séquences de liaison NFI comme séquences isolatrices dans des applications de la biotechnologie.
Resumo:
BACKGROUND: The estimation of demographic parameters from genetic data often requires the computation of likelihoods. However, the likelihood function is computationally intractable for many realistic evolutionary models, and the use of Bayesian inference has therefore been limited to very simple models. The situation changed recently with the advent of Approximate Bayesian Computation (ABC) algorithms allowing one to obtain parameter posterior distributions based on simulations not requiring likelihood computations. RESULTS: Here we present ABCtoolbox, a series of open source programs to perform Approximate Bayesian Computations (ABC). It implements various ABC algorithms including rejection sampling, MCMC without likelihood, a Particle-based sampler and ABC-GLM. ABCtoolbox is bundled with, but not limited to, a program that allows parameter inference in a population genetics context and the simultaneous use of different types of markers with different ploidy levels. In addition, ABCtoolbox can also interact with most simulation and summary statistics computation programs. The usability of the ABCtoolbox is demonstrated by inferring the evolutionary history of two evolutionary lineages of Microtus arvalis. Using nuclear microsatellites and mitochondrial sequence data in the same estimation procedure enabled us to infer sex-specific population sizes and migration rates and to find that males show smaller population sizes but much higher levels of migration than females. CONCLUSION: ABCtoolbox allows a user to perform all the necessary steps of a full ABC analysis, from parameter sampling from prior distributions, data simulations, computation of summary statistics, estimation of posterior distributions, model choice, validation of the estimation procedure, and visualization of the results.
Resumo:
Geophysical tomography captures the spatial distribution of the underlying geophysical property at a relatively high resolution, but the tomographic images tend to be blurred representations of reality and generally fail to reproduce sharp interfaces. Such models may cause significant bias when taken as a basis for predictive flow and transport modeling and are unsuitable for uncertainty assessment. We present a methodology in which tomograms are used to condition multiple-point statistics (MPS) simulations. A large set of geologically reasonable facies realizations and their corresponding synthetically calculated cross-hole radar tomograms are used as a training image. The training image is scanned with a direct sampling algorithm for patterns in the conditioning tomogram, while accounting for the spatially varying resolution of the tomograms. In a post-processing step, only those conditional simulations that predicted the radar traveltimes within the expected data error levels are accepted. The methodology is demonstrated on a two-facies example featuring channels and an aquifer analog of alluvial sedimentary structures with five facies. For both cases, MPS simulations exhibit the sharp interfaces and the geological patterns found in the training image. Compared to unconditioned MPS simulations, the uncertainty in transport predictions is markedly decreased for simulations conditioned to tomograms. As an improvement to other approaches relying on classical smoothness-constrained geophysical tomography, the proposed method allows for: (1) reproduction of sharp interfaces, (2) incorporation of realistic geological constraints and (3) generation of multiple realizations that enables uncertainty assessment.
Resumo:
An active learning method is proposed for the semi-automatic selection of training sets in remote sensing image classification. The method adds iteratively to the current training set the unlabeled pixels for which the prediction of an ensemble of classifiers based on bagged training sets show maximum entropy. This way, the algorithm selects the pixels that are the most uncertain and that will improve the model if added in the training set. The user is asked to label such pixels at each iteration. Experiments using support vector machines (SVM) on an 8 classes QuickBird image show the excellent performances of the methods, that equals accuracies of both a model trained with ten times more pixels and a model whose training set has been built using a state-of-the-art SVM specific active learning method
Resumo:
3 Summary 3. 1 English The pharmaceutical industry has been facing several challenges during the last years, and the optimization of their drug discovery pipeline is believed to be the only viable solution. High-throughput techniques do participate actively to this optimization, especially when complemented by computational approaches aiming at rationalizing the enormous amount of information that they can produce. In siiico techniques, such as virtual screening or rational drug design, are now routinely used to guide drug discovery. Both heavily rely on the prediction of the molecular interaction (docking) occurring between drug-like molecules and a therapeutically relevant target. Several softwares are available to this end, but despite the very promising picture drawn in most benchmarks, they still hold several hidden weaknesses. As pointed out in several recent reviews, the docking problem is far from being solved, and there is now a need for methods able to identify binding modes with a high accuracy, which is essential to reliably compute the binding free energy of the ligand. This quantity is directly linked to its affinity and can be related to its biological activity. Accurate docking algorithms are thus critical for both the discovery and the rational optimization of new drugs. In this thesis, a new docking software aiming at this goal is presented, EADock. It uses a hybrid evolutionary algorithm with two fitness functions, in combination with a sophisticated management of the diversity. EADock is interfaced with .the CHARMM package for energy calculations and coordinate handling. A validation was carried out on 37 crystallized protein-ligand complexes featuring 11 different proteins. The search space was defined as a sphere of 15 R around the center of mass of the ligand position in the crystal structure, and conversely to other benchmarks, our algorithms was fed with optimized ligand positions up to 10 A root mean square deviation 2MSD) from the crystal structure. This validation illustrates the efficiency of our sampling heuristic, as correct binding modes, defined by a RMSD to the crystal structure lower than 2 A, were identified and ranked first for 68% of the complexes. The success rate increases to 78% when considering the five best-ranked clusters, and 92% when all clusters present in the last generation are taken into account. Most failures in this benchmark could be explained by the presence of crystal contacts in the experimental structure. EADock has been used to understand molecular interactions involved in the regulation of the Na,K ATPase, and in the activation of the nuclear hormone peroxisome proliferatoractivated receptors a (PPARa). It also helped to understand the action of common pollutants (phthalates) on PPARy, and the impact of biotransformations of the anticancer drug Imatinib (Gleevec®) on its binding mode to the Bcr-Abl tyrosine kinase. Finally, a fragment-based rational drug design approach using EADock was developed, and led to the successful design of new peptidic ligands for the a5ß1 integrin, and for the human PPARa. In both cases, the designed peptides presented activities comparable to that of well-established ligands such as the anticancer drug Cilengitide and Wy14,643, respectively. 3.2 French Les récentes difficultés de l'industrie pharmaceutique ne semblent pouvoir se résoudre que par l'optimisation de leur processus de développement de médicaments. Cette dernière implique de plus en plus. de techniques dites "haut-débit", particulièrement efficaces lorsqu'elles sont couplées aux outils informatiques permettant de gérer la masse de données produite. Désormais, les approches in silico telles que le criblage virtuel ou la conception rationnelle de nouvelles molécules sont utilisées couramment. Toutes deux reposent sur la capacité à prédire les détails de l'interaction moléculaire entre une molécule ressemblant à un principe actif (PA) et une protéine cible ayant un intérêt thérapeutique. Les comparatifs de logiciels s'attaquant à cette prédiction sont flatteurs, mais plusieurs problèmes subsistent. La littérature récente tend à remettre en cause leur fiabilité, affirmant l'émergence .d'un besoin pour des approches plus précises du mode d'interaction. Cette précision est essentielle au calcul de l'énergie libre de liaison, qui est directement liée à l'affinité du PA potentiel pour la protéine cible, et indirectement liée à son activité biologique. Une prédiction précise est d'une importance toute particulière pour la découverte et l'optimisation de nouvelles molécules actives. Cette thèse présente un nouveau logiciel, EADock, mettant en avant une telle précision. Cet algorithme évolutionnaire hybride utilise deux pressions de sélections, combinées à une gestion de la diversité sophistiquée. EADock repose sur CHARMM pour les calculs d'énergie et la gestion des coordonnées atomiques. Sa validation a été effectuée sur 37 complexes protéine-ligand cristallisés, incluant 11 protéines différentes. L'espace de recherche a été étendu à une sphère de 151 de rayon autour du centre de masse du ligand cristallisé, et contrairement aux comparatifs habituels, l'algorithme est parti de solutions optimisées présentant un RMSD jusqu'à 10 R par rapport à la structure cristalline. Cette validation a permis de mettre en évidence l'efficacité de notre heuristique de recherche car des modes d'interactions présentant un RMSD inférieur à 2 R par rapport à la structure cristalline ont été classés premier pour 68% des complexes. Lorsque les cinq meilleures solutions sont prises en compte, le taux de succès grimpe à 78%, et 92% lorsque la totalité de la dernière génération est prise en compte. La plupart des erreurs de prédiction sont imputables à la présence de contacts cristallins. Depuis, EADock a été utilisé pour comprendre les mécanismes moléculaires impliqués dans la régulation de la Na,K ATPase et dans l'activation du peroxisome proliferatoractivated receptor a (PPARa). Il a également permis de décrire l'interaction de polluants couramment rencontrés sur PPARy, ainsi que l'influence de la métabolisation de l'Imatinib (PA anticancéreux) sur la fixation à la kinase Bcr-Abl. Une approche basée sur la prédiction des interactions de fragments moléculaires avec protéine cible est également proposée. Elle a permis la découverte de nouveaux ligands peptidiques de PPARa et de l'intégrine a5ß1. Dans les deux cas, l'activité de ces nouveaux peptides est comparable à celles de ligands bien établis, comme le Wy14,643 pour le premier, et le Cilengitide (PA anticancéreux) pour la seconde.
Resumo:
rejection can lead to loss of function. Histological reading of endomyocardial biopsy remains the "gold standard" for guiding immunosuppression, despite its methodological limitations (sampling error and interobserver variability). The measurement of the T2 relaxation time has been suggested for detection of allograft rejection, on the pathophysiological basis that the T2 relaxation time prolongs with local edema resulting from acute allograft rejection. Using breath-held cardiac magnetic resonance T2 mapping at 1.5 T, Usman et al. (CircCardiovascImaging2012) detected moderate allograft rejection (grade 2R, ISHLT 2004). With modern immunosuppression grade 2R rejection has become a rare event, but the need remains for a technique that permits the discrimination of absent (grade 0R) and mild rejection (grade 1R). We therefore investigated whether an increase of magnetic field strength to 3T and the use of real-time navigator-gated respiration compensation allow for an increase in the sensitivity of T2 relaxation time detection that is necessary to achieve this discrimination. Methods: Eighteen patients received EMB (Tan et al., ArchPatholLabMed2007) and cardiac T2 mapping on the same day. Reading of T2 maps was blinded to the histological results. For final analysis, 3 cases with known 2R rejection at the time of T2 mapping were added, yielding 21 T2 mapping sessions. A respiration-navigator-gated radial gradient-recalled-echo pulse sequence (resolution 1.17 mm2, matrix 2562, trigger time 3 heartbeats, T2 preparation duration TET2 Prep = 60/30/0 ms) was applied to obtain 3 short-axis T2 maps (van Heeswijk et al., JACCCardiovascImaging2012), which were segmented according to AHA guidelines (Cerqueira et al, Circulation2001). The highest segmental T2 values were grouped according to histological rejection grade and differences were analyzed by Student's t-test, except for the non-blinded cases with 2R rejection. The degree of discrimination was determined using the Spearman's ranked correlation test. Results: The high-quality T2 maps allowed for visual differentiation of the rejection degrees (Figure 1), and the correlation of T2 mapping with the histological grade of acute cellular rejection was significant (Spearman's r = 0.56, p = 0.007). The 0R (n = 15) and 1R (n = 3) degrees demonstrated significantly different T2 values (46.9 ± 5.0 and 54.3 ± 3.0 ms, p = 0.02, Figure 2). Cases with 2R rejection showed clear T2 elevation (T2 = 60.3 ± 16.2 ms). Conclusions: This pilot study demonstrates that non-invasive free-breathing cardiac T2 mapping at 3T discriminates between no and mild cardiac allograft rejection. Confirmation of these encouraging results in a larger cohort should consider a study able to show equivalency or superiority of T2 mapping.
Resumo:
Purpose: In extreme situations, such as hyperacute rejection of heart transplant or major bleeding per-operating complications, an urgent heart explantation might be the only means of survival. The aim of this experimental study was to improve the surgical technique and the hemodynamics of an Extracorporeal Membrane Oxygenation (ECMO) support through a peripheral vascular access in an acardia model. Methods: An ECMO support was established in 7 bovine experiments (59±6.1 kg) by the transjugular insertion to the caval axis of a self-expanded cannula, with return through a carotid artery. After baseline measurements of pump flow and arterial and central venous pressure, ventricular fibrillation was induced (B), the great arteries were clamped, the heart was excised and right and left atria remnants, containing the pulmonary veins, were sutured together leaving an atrial septal defect (ASD) over the cannula in the caval axis. Measurements were taken with the pulmonary artery (PA) clamped (C) and anastomosed with the caval axis (D). Regular arterial and central venous blood gases tests were performed. The ANOVA test for repeated measures was used to test the null hypothesis and a Bonferroni t method for assessing the significance in the between groups pairwise comparison of mean pump flow. Results: Initial pump flow (A) was 4.3±0.6 L/min dropping to 2.8±0.7 L/min (P B-A= 0.003) 10 minutes after induction of ventricular fibrillation (B). After cardiectomy, with the pulmonary artery clamped (C) it augmented not significantly to 3.5±0.8 L/min (P C-B= 0.33, P C-A= 0.029). Finally, PA anastomosis to the caval axis was followed by an almost to baseline pump flow augmentation (4.1±0.7 L/min, P D-B= 0.009, P D-C= 0.006, P D-A= 0.597), permitting a full ECMO support in acardia by a peripheral vascular access. Conclusions: ECMO support in acardia is feasible, providing new opportunities in situations where heart must urgently be explanted, as in hyperacute rejection of heart transplant. Adequate drainage of pulmonary circulation is pivotal in order to avoid pulmonary congestion and loss of volume from the normal right to left shunt of bronchial vessels. Furthermore, the PA anastomosis to the caval axis not only improves pump flow but it also permits an ECMO support by a peripheral vascular access and the closure of the chest.
Resumo:
An assessment of sewage workers' exposure to airborne cultivable bacteria, fungi and inhaled endotoxins was performed at 11 sewage treatment plants. We sampled the enclosed and unenclosed treatment areas in each plant and evaluated the influence of seasons (summer and winter) on bioaerosol levels. We also measured personal exposure to endotoxins of workers during special operation where a higher risk of bioaerosol inhalation was assumed. Results show that only fungi are present in significantly higher concentrations in summer than in winter (2331 +/- 858 versus 329 +/- 95 CFU m(-3)). We also found that there are significantly more bacteria in the enclosed area, near the particle grids for incoming water, than in the unenclosed area near the aeration basins (9455 +/- 2661 versus 2435 +/- 985 CFU m(-3) in summer and 11 081 +/- 2299 versus 2002 +/- 839 CFU m(-3) in winter). All bioaerosols were frequently above the recommended values of occupational exposure. Workers carrying out special tasks such as cleaning tanks were exposed to very high levels of endotoxins (up to 500 EU m(-3)) compared to routine work. The species composition and concentration of airborne Gram-negative bacteria were also studied. A broad spectrum of different species within the Pseudomonadaceae and the Enterobacteriaceae families were predominant in nearly all plants investigated. [Authors]
Resumo:
Introduction/objectives: Multipatient use of a single-patient CBSD occurred inan outpatient clinic during 4 to 16 months before itsnotification. We looked for transmission of blood-bornepathogens among exposed patients.Methods: Exposed patients underwent serology testing for HBV,HCV and HIV. Patients with isolated anti-HBc receivedone dose of hepatitis B vaccine to look for a memoryimmune response. Possible transmissions were investigatedby mapping visits and sequencing of the viral genomeif needed.Results: Of 280 exposed patients, 9 had died without suspicionof blood-borne infection, 3 could not be tested, and 5declined investigations. Among the 263 (93%) testedpatients, 218 (83%) had negative results. We confirmeda known history of HCV infection in 6 patients (1 coinfectedby HIV), and also identified resolved HBVinfection in 37 patients, of whom 18 were alreadyknown. 2 patients were found to have a previouslyunknown HCV infection. According to the time elapsedfrom the closest previous visit of a HCV-infected potentialsource patient, we could rule out nosocomial transmissionin one case (14 weeks) but not in the other (1day). In the latter, however, transmission was deemedvery unlikely by 2 reference centers based on thesequences of the E1 and HVR1 regions of the virus.Conclusion: We did not identify any transmission of blood-bornepathogens in 263 patients exposed to a single-patientCBSD, despite the presence of potential source cases.Change of needle and disinfection of the device betweenpatients may have contributed to this outcome.Although we cannot exclude transmission of HBV, previousacquisition in endemic countries is a more likelyexplanation in this multi-national population.
Resumo:
The role of land cover change as a significant component of global change has become increasingly recognized in recent decades. Large databases measuring land cover change, and the data which can potentially be used to explain the observed changes, are also becoming more commonly available. When developing statistical models to investigate observed changes, it is important to be aware that the chosen sampling strategy and modelling techniques can influence results. We present a comparison of three sampling strategies and two forms of grouped logistic regression models (multinomial and ordinal) in the investigation of patterns of successional change after agricultural land abandonment in Switzerland. Results indicated that both ordinal and nominal transitional change occurs in the landscape and that the use of different sampling regimes and modelling techniques as investigative tools yield different results. Synthesis and applications. Our multimodel inference identified successfully a set of consistently selected indicators of land cover change, which can be used to predict further change, including annual average temperature, the number of already overgrown neighbouring areas of land and distance to historically destructive avalanche sites. This allows for more reliable decision making and planning with respect to landscape management. Although both model approaches gave similar results, ordinal regression yielded more parsimonious models that identified the important predictors of land cover change more efficiently. Thus, this approach is favourable where land cover change pattern can be interpreted as an ordinal process. Otherwise, multinomial logistic regression is a viable alternative.
Resumo:
A study on lead pollution was carried out on a sample of ca. 300 city children. This paper presents the errors producing bias in the sample. It is emphasized that, in Switzerland, the difference between the Swiss and the migrant population (the latter being mainly Italian and Spanish) must be taken into account.
Resumo:
The implicit projection algorithm of isotropic plasticity is extended to an objective anisotropic elastic perfectly plastic model. The recursion formula developed to project the trial stress on the yield surface, is applicable to any non linear elastic law and any plastic yield function.A curvilinear transverse isotropic model based on a quadratic elastic potential and on Hill's quadratic yield criterion is then developed and implemented in a computer program for bone mechanics perspectives. The paper concludes with a numerical study of a schematic bone-prosthesis system to illustrate the potential of the model.
Resumo:
Defining an efficient training set is one of the most delicate phases for the success of remote sensing image classification routines. The complexity of the problem, the limited temporal and financial resources, as well as the high intraclass variance can make an algorithm fail if it is trained with a suboptimal dataset. Active learning aims at building efficient training sets by iteratively improving the model performance through sampling. A user-defined heuristic ranks the unlabeled pixels according to a function of the uncertainty of their class membership and then the user is asked to provide labels for the most uncertain pixels. This paper reviews and tests the main families of active learning algorithms: committee, large margin, and posterior probability-based. For each of them, the most recent advances in the remote sensing community are discussed and some heuristics are detailed and tested. Several challenging remote sensing scenarios are considered, including very high spatial resolution and hyperspectral image classification. Finally, guidelines for choosing the good architecture are provided for new and/or unexperienced user.