140 resultados para Error-resilient Applications

em Université de Lausanne, Switzerland


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Zero correlation between measurement error and model error has been assumed in existing panel data models dealing specifically with measurement error. We extend this literature and propose a simple model where one regressor is mismeasured, allowing the measurement error to correlate with model error. Zero correlation between measurement error and model error is a special case in our model where correlated measurement error equals zero. We ask two research questions. First, we wonder if the correlated measurement error can be identified in the context of panel data. Second, we wonder if classical instrumental variables in panel data need to be adjusted when correlation between measurement error and model error cannot be ignored. Under some regularity conditions the answer is yes to both questions. We then propose a two-step estimation corresponding to the two questions. The first step estimates correlated measurement error from a reverse regression; and the second step estimates usual coefficients of interest using adjusted instruments.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Notre consommation en eau souterraine, en particulier comme eau potable ou pour l'irrigation, a considérablement augmenté au cours des années. De nombreux problèmes font alors leur apparition, allant de la prospection de nouvelles ressources à la remédiation des aquifères pollués. Indépendamment du problème hydrogéologique considéré, le principal défi reste la caractérisation des propriétés du sous-sol. Une approche stochastique est alors nécessaire afin de représenter cette incertitude en considérant de multiples scénarios géologiques et en générant un grand nombre de réalisations géostatistiques. Nous rencontrons alors la principale limitation de ces approches qui est le coût de calcul dû à la simulation des processus d'écoulements complexes pour chacune de ces réalisations. Dans la première partie de la thèse, ce problème est investigué dans le contexte de propagation de l'incertitude, oú un ensemble de réalisations est identifié comme représentant les propriétés du sous-sol. Afin de propager cette incertitude à la quantité d'intérêt tout en limitant le coût de calcul, les méthodes actuelles font appel à des modèles d'écoulement approximés. Cela permet l'identification d'un sous-ensemble de réalisations représentant la variabilité de l'ensemble initial. Le modèle complexe d'écoulement est alors évalué uniquement pour ce sousensemble, et, sur la base de ces réponses complexes, l'inférence est faite. Notre objectif est d'améliorer la performance de cette approche en utilisant toute l'information à disposition. Pour cela, le sous-ensemble de réponses approximées et exactes est utilisé afin de construire un modèle d'erreur, qui sert ensuite à corriger le reste des réponses approximées et prédire la réponse du modèle complexe. Cette méthode permet de maximiser l'utilisation de l'information à disposition sans augmentation perceptible du temps de calcul. La propagation de l'incertitude est alors plus précise et plus robuste. La stratégie explorée dans le premier chapitre consiste à apprendre d'un sous-ensemble de réalisations la relation entre les modèles d'écoulement approximé et complexe. Dans la seconde partie de la thèse, cette méthodologie est formalisée mathématiquement en introduisant un modèle de régression entre les réponses fonctionnelles. Comme ce problème est mal posé, il est nécessaire d'en réduire la dimensionnalité. Dans cette optique, l'innovation du travail présenté provient de l'utilisation de l'analyse en composantes principales fonctionnelles (ACPF), qui non seulement effectue la réduction de dimensionnalités tout en maximisant l'information retenue, mais permet aussi de diagnostiquer la qualité du modèle d'erreur dans cet espace fonctionnel. La méthodologie proposée est appliquée à un problème de pollution par une phase liquide nonaqueuse et les résultats obtenus montrent que le modèle d'erreur permet une forte réduction du temps de calcul tout en estimant correctement l'incertitude. De plus, pour chaque réponse approximée, une prédiction de la réponse complexe est fournie par le modèle d'erreur. Le concept de modèle d'erreur fonctionnel est donc pertinent pour la propagation de l'incertitude, mais aussi pour les problèmes d'inférence bayésienne. Les méthodes de Monte Carlo par chaîne de Markov (MCMC) sont les algorithmes les plus communément utilisés afin de générer des réalisations géostatistiques en accord avec les observations. Cependant, ces méthodes souffrent d'un taux d'acceptation très bas pour les problèmes de grande dimensionnalité, résultant en un grand nombre de simulations d'écoulement gaspillées. Une approche en deux temps, le "MCMC en deux étapes", a été introduite afin d'éviter les simulations du modèle complexe inutiles par une évaluation préliminaire de la réalisation. Dans la troisième partie de la thèse, le modèle d'écoulement approximé couplé à un modèle d'erreur sert d'évaluation préliminaire pour le "MCMC en deux étapes". Nous démontrons une augmentation du taux d'acceptation par un facteur de 1.5 à 3 en comparaison avec une implémentation classique de MCMC. Une question reste sans réponse : comment choisir la taille de l'ensemble d'entrainement et comment identifier les réalisations permettant d'optimiser la construction du modèle d'erreur. Cela requiert une stratégie itérative afin que, à chaque nouvelle simulation d'écoulement, le modèle d'erreur soit amélioré en incorporant les nouvelles informations. Ceci est développé dans la quatrième partie de la thèse, oú cette méthodologie est appliquée à un problème d'intrusion saline dans un aquifère côtier. -- Our consumption of groundwater, in particular as drinking water and for irrigation, has considerably increased over the years and groundwater is becoming an increasingly scarce and endangered resource. Nofadays, we are facing many problems ranging from water prospection to sustainable management and remediation of polluted aquifers. Independently of the hydrogeological problem, the main challenge remains dealing with the incomplete knofledge of the underground properties. Stochastic approaches have been developed to represent this uncertainty by considering multiple geological scenarios and generating a large number of realizations. The main limitation of this approach is the computational cost associated with performing complex of simulations in each realization. In the first part of the thesis, we explore this issue in the context of uncertainty propagation, where an ensemble of geostatistical realizations is identified as representative of the subsurface uncertainty. To propagate this lack of knofledge to the quantity of interest (e.g., the concentration of pollutant in extracted water), it is necessary to evaluate the of response of each realization. Due to computational constraints, state-of-the-art methods make use of approximate of simulation, to identify a subset of realizations that represents the variability of the ensemble. The complex and computationally heavy of model is then run for this subset based on which inference is made. Our objective is to increase the performance of this approach by using all of the available information and not solely the subset of exact responses. Two error models are proposed to correct the approximate responses follofing a machine learning approach. For the subset identified by a classical approach (here the distance kernel method) both the approximate and the exact responses are knofn. This information is used to construct an error model and correct the ensemble of approximate responses to predict the "expected" responses of the exact model. The proposed methodology makes use of all the available information without perceptible additional computational costs and leads to an increase in accuracy and robustness of the uncertainty propagation. The strategy explored in the first chapter consists in learning from a subset of realizations the relationship between proxy and exact curves. In the second part of this thesis, the strategy is formalized in a rigorous mathematical framework by defining a regression model between functions. As this problem is ill-posed, it is necessary to reduce its dimensionality. The novelty of the work comes from the use of functional principal component analysis (FPCA), which not only performs the dimensionality reduction while maximizing the retained information, but also allofs a diagnostic of the quality of the error model in the functional space. The proposed methodology is applied to a pollution problem by a non-aqueous phase-liquid. The error model allofs a strong reduction of the computational cost while providing a good estimate of the uncertainty. The individual correction of the proxy response by the error model leads to an excellent prediction of the exact response, opening the door to many applications. The concept of functional error model is useful not only in the context of uncertainty propagation, but also, and maybe even more so, to perform Bayesian inference. Monte Carlo Markov Chain (MCMC) algorithms are the most common choice to ensure that the generated realizations are sampled in accordance with the observations. Hofever, this approach suffers from lof acceptance rate in high dimensional problems, resulting in a large number of wasted of simulations. This led to the introduction of two-stage MCMC, where the computational cost is decreased by avoiding unnecessary simulation of the exact of thanks to a preliminary evaluation of the proposal. In the third part of the thesis, a proxy is coupled to an error model to provide an approximate response for the two-stage MCMC set-up. We demonstrate an increase in acceptance rate by a factor three with respect to one-stage MCMC results. An open question remains: hof do we choose the size of the learning set and identify the realizations to optimize the construction of the error model. This requires devising an iterative strategy to construct the error model, such that, as new of simulations are performed, the error model is iteratively improved by incorporating the new information. This is discussed in the fourth part of the thesis, in which we apply this methodology to a problem of saline intrusion in a coastal aquifer.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This study aimed to design and validate the measurement of ankle kinetics (force, moment, and power) during consecutive gait cycles and in the field using an ambulatory system. An ambulatory system consisting of plantar pressure insole and inertial sensors (3D gyroscopes and 3D accelerometers) on foot and shank was used. To test this system, 12 patients and 10 healthy elderly subjects wore shoes embedding this system and walked many times across a gait lab including a force-plate surrounded by seven cameras considered as the reference system. Then, the participants walked two 50-meter trials where only the ambulatory system was used. Ankle force components and sagittal moment of ankle measured by ambulatory system showed correlation coefficient (R) and normalized RMS error (NRMSE) of more than 0.94 and less than 13% in comparison with the references system for both patients and healthy subjects. Transverse moment of ankle and ankle power showed R>0.85 and NRMSE<23%. These parameters also showed high repeatability (CMC>0.7). In contrast, the ankle coronal moment of ankle demonstrated high error and lower repeatability. Except for ankle coronal moment, the kinetic features obtained by the ambulatory system could distinguish the patients with ankle osteoarthritis from healthy subjects when measured in 50-meter trials. The proposed ambulatory system can be easily accessible in most clinics and could assess main ankle kinetics quantities with acceptable error and repeatability for clinical evaluations. This system is therefore suggested for field measurement in clinical applications.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Analysis of variance is commonly used in morphometry in order to ascertain differences in parameters between several populations. Failure to detect significant differences between populations (type II error) may be due to suboptimal sampling and lead to erroneous conclusions; the concept of statistical power allows one to avoid such failures by means of an adequate sampling. Several examples are given in the morphometry of the nervous system, showing the use of the power of a hierarchical analysis of variance test for the choice of appropriate sample and subsample sizes. In the first case chosen, neuronal densities in the human visual cortex, we find the number of observations to be of little effect. For dendritic spine densities in the visual cortex of mice and humans, the effect is somewhat larger. A substantial effect is shown in our last example, dendritic segmental lengths in monkey lateral geniculate nucleus. It is in the nature of the hierarchical model that sample size is always more important than subsample size. The relative weight to be attributed to subsample size thus depends on the relative magnitude of the between observations variance compared to the between individuals variance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In groundwater applications, Monte Carlo methods are employed to model the uncertainty on geological parameters. However, their brute-force application becomes computationally prohibitive for highly detailed geological descriptions, complex physical processes, and a large number of realizations. The Distance Kernel Method (DKM) overcomes this issue by clustering the realizations in a multidimensional space based on the flow responses obtained by means of an approximate (computationally cheaper) model; then, the uncertainty is estimated from the exact responses that are computed only for one representative realization per cluster (the medoid). Usually, DKM is employed to decrease the size of the sample of realizations that are considered to estimate the uncertainty. We propose to use the information from the approximate responses for uncertainty quantification. The subset of exact solutions provided by DKM is then employed to construct an error model and correct the potential bias of the approximate model. Two error models are devised that both employ the difference between approximate and exact medoid solutions, but differ in the way medoid errors are interpolated to correct the whole set of realizations. The Local Error Model rests upon the clustering defined by DKM and can be seen as a natural way to account for intra-cluster variability; the Global Error Model employs a linear interpolation of all medoid errors regardless of the cluster to which the single realization belongs. These error models are evaluated for an idealized pollution problem in which the uncertainty of the breakthrough curve needs to be estimated. For this numerical test case, we demonstrate that the error models improve the uncertainty quantification provided by the DKM algorithm and are effective in correcting the bias of the estimate computed solely from the MsFV results. The framework presented here is not specific to the methods considered and can be applied to other combinations of approximate models and techniques to select a subset of realizations

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Restriction site-associated DNA sequencing (RADseq) provides researchers with the ability to record genetic polymorphism across thousands of loci for nonmodel organisms, potentially revolutionizing the field of molecular ecology. However, as with other genotyping methods, RADseq is prone to a number of sources of error that may have consequential effects for population genetic inferences, and these have received only limited attention in terms of the estimation and reporting of genotyping error rates. Here we use individual sample replicates, under the expectation of identical genotypes, to quantify genotyping error in the absence of a reference genome. We then use sample replicates to (i) optimize de novo assembly parameters within the program Stacks, by minimizing error and maximizing the retrieval of informative loci; and (ii) quantify error rates for loci, alleles and single-nucleotide polymorphisms. As an empirical example, we use a double-digest RAD data set of a nonmodel plant species, Berberis alpina, collected from high-altitude mountains in Mexico.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Biological materials are increasingly used in abdominal surgery for ventral, pelvic and perineal reconstructions, especially in contaminated fields. Future applications are multi-fold and include prevention and one-step closure of infected areas. This includes prevention of abdominal, parastomal and pelvic hernia, but could also include prevention of separation of multiple anastomoses, suture- or staple-lines. Further indications could be a containment of infected and/or inflammatory areas and protection of vital implants such as vascular grafts. Reinforcement patches of high-risk anastomoses or unresectable perforation sites are possibilities at least. Current applications are based mostly on case series and better data is urgently needed. Clinical benefits need to be assessed in prospective studies to provide reliable proof of efficacy with a sufficient follow-up. Only superior results compared with standard treatment will justify the higher costs of these materials. To date, the use of biological materials is not standard and applications should be limited to case-by-case decision.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This letter describes a data telemetry biomedical experiment. An implant, consisting of a biometric data sensor, electronics, an antenna, and a biocompatible capsule, is described. All the elements were co-designed in order to maximize the transmission distance. The device was implanted in a pig for an in vivo experiment of temperature monitoring.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Pharmacogenomics is a field with origins in the study of monogenic variations in drug metabolism in the 1950s. Perhaps because of these historical underpinnings, there has been an intensive investigation of 'hepatic pharmacogenes' such as CYP450s and liver drug metabolism using pharmacogenomics approaches over the past five decades. Surprisingly, kidney pathophysiology, attendant diseases and treatment outcomes have been vastly under-studied and under-theorized despite their central importance in maintenance of health, susceptibility to disease and rational personalized therapeutics. Indeed, chronic kidney disease (CKD) represents an increasing public health burden worldwide, both in developed and developing countries. Patients with CKD suffer from high cardiovascular morbidity and mortality, which is mainly attributable to cardiovascular events before reaching end-stage renal disease. In this paper, we focus our analyses on renal function before end-stage renal disease, as seen through the lens of pharmacogenomics and human genomic variation. We herein synthesize the recent evidence linking selected Very Important Pharmacogenes (VIP) to renal function, blood pressure and salt-sensitivity in humans, and ways in which these insights might inform rational personalized therapeutics. Notably, we highlight and present the rationale for three applications that we consider as important and actionable therapeutic and preventive focus areas in renal pharmacogenomics: 1) ACE inhibitors, as a confirmed application, 2) VDR agonists, as a promising application, and 3) moderate dietary salt intake, as a suggested novel application. Additionally, we emphasize the putative contributions of gene-environment interactions, discuss the implications of these findings to treat and prevent hypertension and CKD. Finally, we conclude with a strategic agenda and vision required to accelerate advances in this under-studied field of renal pharmacogenomics with vast significance for global public health.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Age is the main clinical determinant of large artery stiffness. Central arteries stiffen progressively with age, whereas peripheral muscular arteries change little with age. A number of clinical studies have analyzed the effects of age on aortic stiffness. Increase of central artery stiffness with age is responsible for earlier wave reflections and changes in pressure wave contours. The stiffening of aorta and other central arteries is a potential risk factor for increased cardiovascular morbidity and mortality. Arterial stiffening with aging is accompanied by an elevation in systolic blood pressure (BP) and pulse pressure (PP). Although arterial stiffening with age is a common situation, it has now been confirmed that older subjects with increased arterial stiffness and elevated PP have higher cardiovascular morbidity and mortality. Increase in aortic stiffness with age occurs gradually and continuously, similarly for men and women. Cross-sectional studies have shown that aortic and carotid stiffness (evaluated by the pulse wave velocity) increase with age by approximately 10% to 15% during a period of 10 years. Women always have 5% to 10% lower stiffness than men of the same age. Although large artery stiffness increases with age independently of the presence of cardiovascular risk factors or other associated conditions, the extent of this increase may depend on several environmental or genetic factors. Hypertension may increase arterial stiffness, especially in older subjects. Among other cardiovascular risk factors, diabetes type 1 and 2 accelerates arterial stiffness, whereas the role of dyslipidemia and tobacco smoking is unclear. Arterial stiffness is also present in several cardiovascular and renal diseases. Patients with heart failure, end stage renal disease, and those with atherosclerotic lesions often develop central artery stiffness. Decreased carotid distensibility, increased arterial thickness, and presence of calcifications and plaques often coexist in the same subject. However, relationships between these three alterations of the arterial wall remain to be explored.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Microarray transcript profiling and RNA interference are two new technologies crucial for large-scale gene function studies in multicellular eukaryotes. Both rely on sequence-specific hybridization between complementary nucleic acid strands, inciting us to create a collection of gene-specific sequence tags (GSTs) representing at least 21,500 Arabidopsis genes and which are compatible with both approaches. The GSTs were carefully selected to ensure that each of them shared no significant similarity with any other region in the Arabidopsis genome. They were synthesized by PCR amplification from genomic DNA. Spotted microarrays fabricated from the GSTs show good dynamic range, specificity, and sensitivity in transcript profiling experiments. The GSTs have also been transferred to bacterial plasmid vectors via recombinational cloning protocols. These cloned GSTs constitute the ideal starting point for a variety of functional approaches, including reverse genetics. We have subcloned GSTs on a large scale into vectors designed for gene silencing in plant cells. We show that in planta expression of GST hairpin RNA results in the expected phenotypes in silenced Arabidopsis lines. These versatile GST resources provide novel and powerful tools for functional genomics.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Grâce à des développements technologiques récents, tels que le système de positionnement par satellites GPS (Global Positioning System) en mode différentie on est maintenant capable de mesurer avec une grande précision non seulement le profil de vitesse de déplacement d'un sujet sur la terre mais aussi sa trajectoire en faisant totalement abstraction de la chronométrie classique. De plus, des capteurs accélérométriques miniaturisés permettent d'obtenir un complément d'information biomécanique utile (fréquence et longueur du pas, signature accélérométrique individuelle). Dans cet artide, un exemple d'application de ces deux techniques à des sports tels que le ski alpin (descente ou Super G) et le sprint est présenté. La combinaison de plusieurs mesures physiologiques, cinétiques et biomécaniques permettra de mieux comprendre les facteurs endogènes et exogènes qui jouent un rôle dans l'amélioration des performances de l'homme.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

So-called online Voting Advice Applications (VAAs) have become very popular all over Europe. Millions of voters are using them as an assistance to make up their minds for which party they should vote. Despite this popularity there are only very few studies about the impact of these tools on individual electoral choice. On the basis of the Swiss VAA smartvote we present some first findings about the question whether VAAs do have a direct impact on the actual vote of their users. In deed, we find strong evidence that Swiss voters were affected by smartvote. However, our findings are somewhat contrary to the results of previous studies from other countries. Furthermore, the quality of available data for such studies needs to be improved. Future studies should pay attention to both: the improvement of the available data, as well as the explanation of the large variance of findings between the specific European countries.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

1. Species distribution modelling is used increasingly in both applied and theoretical research to predict how species are distributed and to understand attributes of species' environmental requirements. In species distribution modelling, various statistical methods are used that combine species occurrence data with environmental spatial data layers to predict the suitability of any site for that species. While the number of data sharing initiatives involving species' occurrences in the scientific community has increased dramatically over the past few years, various data quality and methodological concerns related to using these data for species distribution modelling have not been addressed adequately. 2. We evaluated how uncertainty in georeferences and associated locational error in occurrences influence species distribution modelling using two treatments: (1) a control treatment where models were calibrated with original, accurate data and (2) an error treatment where data were first degraded spatially to simulate locational error. To incorporate error into the coordinates, we moved each coordinate with a random number drawn from the normal distribution with a mean of zero and a standard deviation of 5 km. We evaluated the influence of error on the performance of 10 commonly used distributional modelling techniques applied to 40 species in four distinct geographical regions. 3. Locational error in occurrences reduced model performance in three of these regions; relatively accurate predictions of species distributions were possible for most species, even with degraded occurrences. Two species distribution modelling techniques, boosted regression trees and maximum entropy, were the best performing models in the face of locational errors. The results obtained with boosted regression trees were only slightly degraded by errors in location, and the results obtained with the maximum entropy approach were not affected by such errors. 4. Synthesis and applications. To use the vast array of occurrence data that exists currently for research and management relating to the geographical ranges of species, modellers need to know the influence of locational error on model quality and whether some modelling techniques are particularly robust to error. We show that certain modelling techniques are particularly robust to a moderate level of locational error and that useful predictions of species distributions can be made even when occurrence data include some error.