76 resultados para Error correction codes
em Université de Lausanne, Switzerland
Resumo:
Among the largest resources for biological sequence data is the large amount of expressed sequence tags (ESTs) available in public and proprietary databases. ESTs provide information on transcripts but for technical reasons they often contain sequencing errors. Therefore, when analyzing EST sequences computationally, such errors must be taken into account. Earlier attempts to model error prone coding regions have shown good performance in detecting and predicting these while correcting sequencing errors using codon usage frequencies. In the research presented here, we improve the detection of translation start and stop sites by integrating a more complex mRNA model with codon usage bias based error correction into one hidden Markov model (HMM), thus generalizing this error correction approach to more complex HMMs. We show that our method maintains the performance in detecting coding sequences.
Resumo:
This corrects the article on p. e73445 in vol. 8.]. This corrects the article "Topographical Body Fat Distribution Links to Amino Acid and Lipid Metabolism in Healthy Non-Obese Women" , e73445. There was an error in the title of the article. The correct version of the title in the article is: Topographical Body Fat Distribution Links to Amino Acid and Lipid Metabolism in Healthy Obese Women The correct citation is: Martin F-PJ, Montoliu I, Collino S, Scherer M, Guy P, et al. (2013) Topographical Body Fat Distribution Links to Amino Acid and Lipid Metabolism in Healthy Obese Women. PLoS ONE 8(9): e73445. doi:10.1371/journal.pone.0073445
Resumo:
Notre consommation en eau souterraine, en particulier comme eau potable ou pour l'irrigation, a considérablement augmenté au cours des années. De nombreux problèmes font alors leur apparition, allant de la prospection de nouvelles ressources à la remédiation des aquifères pollués. Indépendamment du problème hydrogéologique considéré, le principal défi reste la caractérisation des propriétés du sous-sol. Une approche stochastique est alors nécessaire afin de représenter cette incertitude en considérant de multiples scénarios géologiques et en générant un grand nombre de réalisations géostatistiques. Nous rencontrons alors la principale limitation de ces approches qui est le coût de calcul dû à la simulation des processus d'écoulements complexes pour chacune de ces réalisations. Dans la première partie de la thèse, ce problème est investigué dans le contexte de propagation de l'incertitude, oú un ensemble de réalisations est identifié comme représentant les propriétés du sous-sol. Afin de propager cette incertitude à la quantité d'intérêt tout en limitant le coût de calcul, les méthodes actuelles font appel à des modèles d'écoulement approximés. Cela permet l'identification d'un sous-ensemble de réalisations représentant la variabilité de l'ensemble initial. Le modèle complexe d'écoulement est alors évalué uniquement pour ce sousensemble, et, sur la base de ces réponses complexes, l'inférence est faite. Notre objectif est d'améliorer la performance de cette approche en utilisant toute l'information à disposition. Pour cela, le sous-ensemble de réponses approximées et exactes est utilisé afin de construire un modèle d'erreur, qui sert ensuite à corriger le reste des réponses approximées et prédire la réponse du modèle complexe. Cette méthode permet de maximiser l'utilisation de l'information à disposition sans augmentation perceptible du temps de calcul. La propagation de l'incertitude est alors plus précise et plus robuste. La stratégie explorée dans le premier chapitre consiste à apprendre d'un sous-ensemble de réalisations la relation entre les modèles d'écoulement approximé et complexe. Dans la seconde partie de la thèse, cette méthodologie est formalisée mathématiquement en introduisant un modèle de régression entre les réponses fonctionnelles. Comme ce problème est mal posé, il est nécessaire d'en réduire la dimensionnalité. Dans cette optique, l'innovation du travail présenté provient de l'utilisation de l'analyse en composantes principales fonctionnelles (ACPF), qui non seulement effectue la réduction de dimensionnalités tout en maximisant l'information retenue, mais permet aussi de diagnostiquer la qualité du modèle d'erreur dans cet espace fonctionnel. La méthodologie proposée est appliquée à un problème de pollution par une phase liquide nonaqueuse et les résultats obtenus montrent que le modèle d'erreur permet une forte réduction du temps de calcul tout en estimant correctement l'incertitude. De plus, pour chaque réponse approximée, une prédiction de la réponse complexe est fournie par le modèle d'erreur. Le concept de modèle d'erreur fonctionnel est donc pertinent pour la propagation de l'incertitude, mais aussi pour les problèmes d'inférence bayésienne. Les méthodes de Monte Carlo par chaîne de Markov (MCMC) sont les algorithmes les plus communément utilisés afin de générer des réalisations géostatistiques en accord avec les observations. Cependant, ces méthodes souffrent d'un taux d'acceptation très bas pour les problèmes de grande dimensionnalité, résultant en un grand nombre de simulations d'écoulement gaspillées. Une approche en deux temps, le "MCMC en deux étapes", a été introduite afin d'éviter les simulations du modèle complexe inutiles par une évaluation préliminaire de la réalisation. Dans la troisième partie de la thèse, le modèle d'écoulement approximé couplé à un modèle d'erreur sert d'évaluation préliminaire pour le "MCMC en deux étapes". Nous démontrons une augmentation du taux d'acceptation par un facteur de 1.5 à 3 en comparaison avec une implémentation classique de MCMC. Une question reste sans réponse : comment choisir la taille de l'ensemble d'entrainement et comment identifier les réalisations permettant d'optimiser la construction du modèle d'erreur. Cela requiert une stratégie itérative afin que, à chaque nouvelle simulation d'écoulement, le modèle d'erreur soit amélioré en incorporant les nouvelles informations. Ceci est développé dans la quatrième partie de la thèse, oú cette méthodologie est appliquée à un problème d'intrusion saline dans un aquifère côtier. -- Our consumption of groundwater, in particular as drinking water and for irrigation, has considerably increased over the years and groundwater is becoming an increasingly scarce and endangered resource. Nofadays, we are facing many problems ranging from water prospection to sustainable management and remediation of polluted aquifers. Independently of the hydrogeological problem, the main challenge remains dealing with the incomplete knofledge of the underground properties. Stochastic approaches have been developed to represent this uncertainty by considering multiple geological scenarios and generating a large number of realizations. The main limitation of this approach is the computational cost associated with performing complex of simulations in each realization. In the first part of the thesis, we explore this issue in the context of uncertainty propagation, where an ensemble of geostatistical realizations is identified as representative of the subsurface uncertainty. To propagate this lack of knofledge to the quantity of interest (e.g., the concentration of pollutant in extracted water), it is necessary to evaluate the of response of each realization. Due to computational constraints, state-of-the-art methods make use of approximate of simulation, to identify a subset of realizations that represents the variability of the ensemble. The complex and computationally heavy of model is then run for this subset based on which inference is made. Our objective is to increase the performance of this approach by using all of the available information and not solely the subset of exact responses. Two error models are proposed to correct the approximate responses follofing a machine learning approach. For the subset identified by a classical approach (here the distance kernel method) both the approximate and the exact responses are knofn. This information is used to construct an error model and correct the ensemble of approximate responses to predict the "expected" responses of the exact model. The proposed methodology makes use of all the available information without perceptible additional computational costs and leads to an increase in accuracy and robustness of the uncertainty propagation. The strategy explored in the first chapter consists in learning from a subset of realizations the relationship between proxy and exact curves. In the second part of this thesis, the strategy is formalized in a rigorous mathematical framework by defining a regression model between functions. As this problem is ill-posed, it is necessary to reduce its dimensionality. The novelty of the work comes from the use of functional principal component analysis (FPCA), which not only performs the dimensionality reduction while maximizing the retained information, but also allofs a diagnostic of the quality of the error model in the functional space. The proposed methodology is applied to a pollution problem by a non-aqueous phase-liquid. The error model allofs a strong reduction of the computational cost while providing a good estimate of the uncertainty. The individual correction of the proxy response by the error model leads to an excellent prediction of the exact response, opening the door to many applications. The concept of functional error model is useful not only in the context of uncertainty propagation, but also, and maybe even more so, to perform Bayesian inference. Monte Carlo Markov Chain (MCMC) algorithms are the most common choice to ensure that the generated realizations are sampled in accordance with the observations. Hofever, this approach suffers from lof acceptance rate in high dimensional problems, resulting in a large number of wasted of simulations. This led to the introduction of two-stage MCMC, where the computational cost is decreased by avoiding unnecessary simulation of the exact of thanks to a preliminary evaluation of the proposal. In the third part of the thesis, a proxy is coupled to an error model to provide an approximate response for the two-stage MCMC set-up. We demonstrate an increase in acceptance rate by a factor three with respect to one-stage MCMC results. An open question remains: hof do we choose the size of the learning set and identify the realizations to optimize the construction of the error model. This requires devising an iterative strategy to construct the error model, such that, as new of simulations are performed, the error model is iteratively improved by incorporating the new information. This is discussed in the fourth part of the thesis, in which we apply this methodology to a problem of saline intrusion in a coastal aquifer.
Resumo:
Restriction site-associated DNA sequencing (RADseq) provides researchers with the ability to record genetic polymorphism across thousands of loci for nonmodel organisms, potentially revolutionizing the field of molecular ecology. However, as with other genotyping methods, RADseq is prone to a number of sources of error that may have consequential effects for population genetic inferences, and these have received only limited attention in terms of the estimation and reporting of genotyping error rates. Here we use individual sample replicates, under the expectation of identical genotypes, to quantify genotyping error in the absence of a reference genome. We then use sample replicates to (i) optimize de novo assembly parameters within the program Stacks, by minimizing error and maximizing the retrieval of informative loci; and (ii) quantify error rates for loci, alleles and single-nucleotide polymorphisms. As an empirical example, we use a double-digest RAD data set of a nonmodel plant species, Berberis alpina, collected from high-altitude mountains in Mexico.
Resumo:
BACKGROUND: Cone-beam computed tomography (CBCT) image-guided radiotherapy (IGRT) systems are widely used tools to verify and correct the target position before each fraction, allowing to maximize treatment accuracy and precision. In this study, we evaluate automatic three-dimensional intensity-based rigid registration (RR) methods for prostate setup correction using CBCT scans and study the impact of rectal distension on registration quality. METHODS: We retrospectively analyzed 115 CBCT scans of 10 prostate patients. CT-to-CBCT registration was performed using (a) global RR, (b) bony RR, or (c) bony RR refined by a local prostate RR using the CT clinical target volume (CTV) expanded with 1-to-20-mm varying margins. After propagation of the manual CT contours, automatic CBCT contours were generated. For evaluation, a radiation oncologist manually delineated the CTV on the CBCT scans. The propagated and manual CBCT contours were compared using the Dice similarity and a measure based on the bidirectional local distance (BLD). We also conducted a blind visual assessment of the quality of the propagated segmentations. Moreover, we automatically quantified rectal distension between the CT and CBCT scans without using the manual CBCT contours and we investigated its correlation with the registration failures. To improve the registration quality, the air in the rectum was replaced with soft tissue using a filter. The results with and without filtering were compared. RESULTS: The statistical analysis of the Dice coefficients and the BLD values resulted in highly significant differences (p<10(-6)) for the 5-mm and 8-mm local RRs vs the global, bony and 1-mm local RRs. The 8-mm local RR provided the best compromise between accuracy and robustness (Dice median of 0.814 and 97% of success with filtering the air in the rectum). We observed that all failures were due to high rectal distension. Moreover, the visual assessment confirmed the superiority of the 8-mm local RR over the bony RR. CONCLUSION: The most successful CT-to-CBCT RR method proved to be the 8-mm local RR. We have shown the correlation between its registration failures and rectal distension. Furthermore, we have provided a simple (easily applicable in routine) and automatic method to quantify rectal distension and to predict registration failure using only the manual CT contours.
Resumo:
BACKGROUND: Surgical correction of complete atrio-ventricular septal defect (AVSD) achieves satisfactory results with low morbidity and mortality, but may require reoperation. Our recent operative results at mid-term were followed-up. METHODS: From June 2000 to December 2007, 81 patients (Down syndrome; n=60), median age 4.0 months (range 0.7-118.6) and weight 4.7kg (range 2.2-33), underwent complete AVSD correction. Patch closure for the ventricular septal defect (VSD; n=69) and atrial septal defect (ASD; n=42) was performed with left atrio-ventricular valve (LAVV) cleft closure (n=76) and right atrio-ventricular valve (RAVV) repair (n=57). Mortality, morbidity, and indications for reoperation were retrospectively studied; the end point 'time to reoperation' was analyzed using Kaplan-Meier curves. Follow-up was complete except in two patients and spanned a median of 28 months (range 0.4-6.1 years). RESULTS: In-hospital mortality was 3.7% (n=3) and one late death occurred. Reoperation was required in 7/79 patients (8.9%) for LAVV insufficiency (n=4), for a residual ASD (n=1), for right atrio-ventricular valve insufficiency (n=1), and for subaortic stenosis (n=1). At last follow-up, no or only mild LAVV and RAVV insufficiency was present in 81.3% and 92.1% of patients, respectively, and 2/3 of patients were medication-free. Risk factors for reoperation were younger age (<3 months; p=0.001) and lower weight (<4kg; p=0.003), and a trend towards less and later reoperations in Down syndrome (p<0.2). CONCLUSIONS: Surgical correction of AVSD can be achieved with low mortality and need for reoperation, regardless of Down syndrome or not. Immediate postoperative moderate or more residual atrio-ventricular valve insufficiency will eventually require a reoperation, and could be anticipated in patients younger than 3 months and weighing <4kg.
Resumo:
Patients with defective ectodysplasin A (EDA) are affected by X-linked hypohidrotic ectodermal dysplasia (XLHED), a condition characterized by sparse hair, inability to sweat, decreased lacrimation, frequent pulmonary infections, and missing and malformed teeth. The canine model of XLHED was used to study the developmental impact of EDA on secondary dentition, since dogs have an entirely brachyodont, diphyodont dentition similar to that in humans, as opposed to mice, which have only permanent teeth (monophyodont dentition), some of which are very different (aradicular hypsodont) than brachyodont human teeth. Also, clinical signs in humans and dogs with XLHED are virtually identical, whereas several are missing in the murine equivalent. In our model, the genetically missing EDA was compensated for by postnatal intravenous administration of soluble recombinant EDA. Untreated XLHED dogs have an incomplete set of conically shaped teeth similar to those seen in human patients with XLHED. After treatment with EDA, significant normalization of adult teeth was achieved in four of five XLHED dogs. Moreover, treatment restored normal lacrimation and resistance to eye and airway infections and improved sweating ability. These results not only provide proof of concept for a potential treatment of this orphan disease but also demonstrate an essential role of EDA in the development of secondary dentition.
Resumo:
Restriction site-associated DNA sequencing (RADseq) provides researchers with the ability to record genetic polymorphism across thousands of loci for nonmodel organisms, potentially revolutionizing the field of molecular ecology. However, as with other genotyping methods, RADseq is prone to a number of sources of error that may have consequential effects for population genetic inferences, and these have received only limited attention in terms of the estimation and reporting of genotyping error rates. Here we use individual sample replicates, under the expectation of identical genotypes, to quantify genotyping error in the absence of a reference genome. We then use sample replicates to (i) optimize de novo assembly parameters within the program Stacks, by minimizing error and maximizing the retrieval of informative loci; and (ii) quantify error rates for loci, alleles and single-nucleotide polymorphisms. As an empirical example, we use a double-digest RAD data set of a nonmodel plant species, Berberis alpina, collected from high-altitude mountains in Mexico.
Resumo:
OBJECTIVES: Residual mitral regurgitation after valve repair worsens patients' clinical outcome. Postimplant adjustable mitral rings potentially address this issue, allowing the reshaping of the annulus on the beating heart under echocardiography control. We developed an original mitral ring allowing valve geometry remodelling after the implantation and designed an animal study to assess device effectiveness in correcting residual mitral regurgitation. METHODS: The device consists of two concentric rings: one internal and flexible, sutured to the mitral annulus and a second external and rigid. A third conic element slides between the two rings, modifying the shape of the flexible ring. This sliding element is remotely activated with a rotating tool. Animal model: in adult swine, under cardio pulmonary bypass and cardiac arrest, we shortened the primary chordae of P2 segment to reproduce Type III regurgitation and implanted the active ring. We used intracardiac ultrasound to assess mitral regurgitation and the efficacy of the active ring to correct it. RESULTS: Severe mitral regurgitation (3+ and 4+) was induced in eight animals, 54 ± 6 kg in weight. Vena contracta width decreased from 0.8 ± 0.2 to 0.1 cm; proximal isovelocity surface area radius decreased from 0.8 ± 0.2 to 0.1 cm and effective regurgitant orifice area decreased from 0.50 ± 0.1 to 0.1 ± 0.1 cm(2). Six animals had a reversal of systolic pulmonary flow that normalized following the activation of the device. All corrections were reversible. CONCLUSIONS: Postimplant adjustable mitral ring corrects severe mitral regurgitation through the reversible modification of the annulus geometry on the beating heart. It addresses the frequent and morbid issue of recurrent mitral valve regurgitation.
Resumo:
Ce travail décrit les opérations théoriques et pratiques concernant le transcodage des opérations d'une part, des diagnostics d'autre part; il décrit également le programme informatique de transcodage.
Resumo:
Zero correlation between measurement error and model error has been assumed in existing panel data models dealing specifically with measurement error. We extend this literature and propose a simple model where one regressor is mismeasured, allowing the measurement error to correlate with model error. Zero correlation between measurement error and model error is a special case in our model where correlated measurement error equals zero. We ask two research questions. First, we wonder if the correlated measurement error can be identified in the context of panel data. Second, we wonder if classical instrumental variables in panel data need to be adjusted when correlation between measurement error and model error cannot be ignored. Under some regularity conditions the answer is yes to both questions. We then propose a two-step estimation corresponding to the two questions. The first step estimates correlated measurement error from a reverse regression; and the second step estimates usual coefficients of interest using adjusted instruments.
Resumo:
The involvement of the cerebellum in migraine pathophysiology is not well understood. We used a biparametric approach at high-field MRI (3 T) to assess the structural integrity of the cerebellum in 15 migraineurs with aura (MWA), 23 migraineurs without aura (MWoA), and 20 healthy controls (HC). High-resolution T1 relaxation maps were acquired together with magnetization transfer images in order to probe microstructural and myelin integrity. Clusterwise analysis was performed on T1 and magnetization transfer ratio (MTR) maps of the cerebellum of MWA, MWoA, and HC using an ANOVA and a non-parametric clusterwise permutation F test, with age and gender as covariates and correction for familywise error rate. In addition, mean MTR and T1 in frontal regions known to be highly connected to the cerebellum were computed. Clusterwise comparison among groups showed a cluster of lower MTR in the right Crus I of MWoA patients vs. HC and MWA subjects (p = 0.04). Univariate and bivariate analysis on T1 and MTR contrasts showed that MWoA patients had longer T1 and lower MTR in the right and left pars orbitalis compared to MWA (p < 0.01 and 0.05, respectively), but no differences were found with HC. Lower MTR and longer T1 point at a loss of macromolecules and/or micro-edema in Crus I and pars orbitalis in MWoA patients vs. HC and vs. MWA. The pathophysiological implications of these findings are discussed in light of recent literature.
Resumo:
Whole-body (WB) planar imaging has long been one of the staple methods of dosimetry, and its quantification has been formalized by the MIRD Committee in pamphlet no 16. One of the issues not specifically addressed in the formalism occurs when the count rates reaching the detector are sufficiently high to result in camera count saturation. Camera dead-time effects have been extensively studied, but all of the developed correction methods assume static acquisitions. However, during WB planar (sweep) imaging, a variable amount of imaged activity exists in the detector's field of view as a function of time and therefore the camera saturation is time dependent. A new time-dependent algorithm was developed to correct for dead-time effects during WB planar acquisitions that accounts for relative motion between detector heads and imaged object. Static camera dead-time parameters were acquired by imaging decaying activity in a phantom and obtaining a saturation curve. Using these parameters, an iterative algorithm akin to Newton's method was developed, which takes into account the variable count rate seen by the detector as a function of time. The algorithm was tested on simulated data as well as on a whole-body scan of high activity Samarium-153 in an ellipsoid phantom. A complete set of parameters from unsaturated phantom data necessary for count rate to activity conversion was also obtained, including build-up and attenuation coefficients, in order to convert corrected count rate values to activity. The algorithm proved successful in accounting for motion- and time-dependent saturation effects in both the simulated and measured data and converged to any desired degree of precision. The clearance half-life calculated from the ellipsoid phantom data was calculated to be 45.1 h after dead-time correction and 51.4 h with no correction; the physical decay half-life of Samarium-153 is 46.3 h. Accurate WB planar dosimetry of high activities relies on successfully compensating for camera saturation which takes into account the variable activity in the field of view, i.e. time-dependent dead-time effects. The algorithm presented here accomplishes this task.
Resumo:
n this paper the iterative MSFV method is extended to include the sequential implicit simulation of time dependent problems involving the solution of a system of pressure-saturation equations. To control numerical errors in simulation results, an error estimate, based on the residual of the MSFV approximate pressure field, is introduced. In the initial time steps in simulation iterations are employed until a specified accuracy in pressure is achieved. This initial solution is then used to improve the localization assumption at later time steps. Additional iterations in pressure solution are employed only when the pressure residual becomes larger than a specified threshold value. Efficiency of the strategy and the error control criteria are numerically investigated. This paper also shows that it is possible to derive an a-priori estimate and control based on the allowed pressure-equation residual to guarantee the desired accuracy in saturation calculation.