1000 resultados para Simulation methodologies


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The identification of genetically homogeneous groups of individuals is a long standing issue in population genetics. A recent Bayesian algorithm implemented in the software STRUCTURE allows the identification of such groups. However, the ability of this algorithm to detect the true number of clusters (K) in a sample of individuals when patterns of dispersal among populations are not homogeneous has not been tested. The goal of this study is to carry out such tests, using various dispersal scenarios from data generated with an individual-based model. We found that in most cases the estimated 'log probability of data' does not provide a correct estimation of the number of clusters, K. However, using an ad hoc statistic DeltaK based on the rate of change in the log probability of data between successive K values, we found that STRUCTURE accurately detects the uppermost hierarchical level of structure for the scenarios we tested. As might be expected, the results are sensitive to the type of genetic marker used (AFLP vs. microsatellite), the number of loci scored, the number of populations sampled, and the number of individuals typed in each sample.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The pharmacokinetic determinants of successful antibiotic prophylaxis of endocarditis are not precisely known. Differences in half-lives of antibiotics between animals and humans preclude extrapolation of animal results to human situations. To overcome this limitation, we have mimicked in rats the amoxicillin kinetics in humans following a 3-g oral dose (as often used for prophylaxis of endocarditis) by delivering the drug through a computerized pump. Rats with catheter-induced vegetations were challenged with either of two strains of antibiotic-tolerant viridans group streptococci. Antibiotics were given either through the pump (to simulate the whole kinetic profile during prophylaxis in humans) or as an intravenous bolus which imitated only the peak level of amoxicillin (18 mg/liter) in human serum. Prophylaxis by intravenous bolus was inoculum dependent and afforded a limited protection only in rats challenged with the minimum inoculum size infecting > or = 90% of untreated controls. In contrast, simulation of kinetics in humans significantly protected animals challenged with 10 to 100 times the inoculum of either of the test organisms infecting > or = 90% of untreated controls. Thus, simulation of the profiles of amoxicillin prophylaxis in human serum was more efficacious than mere imitation of the transient peak level in rats. This confirms previous studies suggesting that the duration for which the serum amoxicillin level remained detectable (not only the magnitude of the peak) was an important parameter in successful prophylaxis of endocarditis. The results also suggest that single-dose prophylaxis with 3 g of amoxicillin in humans might be more effective than predicted by conventional animal models in which only peak levels of antibiotic in human serum were stimulated.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Significant progress has been made with regard to the quantitative integration of geophysical and hydrological data at the local scale for the purpose of improving predictions of groundwater flow and solute transport. However, extending corresponding approaches to the regional scale still represents one of the major challenges in the domain of hydrogeophysics. To address this problem, we have developed a regional-scale data integration methodology based on a two-step Bayesian sequential simulation approach. Our objective is to generate high-resolution stochastic realizations of the regional-scale hydraulic conductivity field in the common case where there exist spatially exhaustive but poorly resolved measurements of a related geophysical parameter, as well as highly resolved but spatially sparse collocated measurements of this geophysical parameter and the hydraulic conductivity. To integrate this multi-scale, multi-parameter database, we first link the low- and high-resolution geophysical data via a stochastic downscaling procedure. This is followed by relating the downscaled geophysical data to the high-resolution hydraulic conductivity distribution. After outlining the general methodology of the approach, we demonstrate its application to a realistic synthetic example where we consider as data high-resolution measurements of the hydraulic and electrical conductivities at a small number of borehole locations, as well as spatially exhaustive, low-resolution estimates of the electrical conductivity obtained from surface-based electrical resistivity tomography. The different stochastic realizations of the hydraulic conductivity field obtained using our procedure are validated by comparing their solute transport behaviour with that of the underlying ?true? hydraulic conductivity field. We find that, even in the presence of strong subsurface heterogeneity, our proposed procedure allows for the generation of faithful representations of the regional-scale hydraulic conductivity structure and reliable predictions of solute transport over long, regional-scale distances.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Barnardos supports children whose well-being is under threat, by working with them, their families and communities and by campaigning for the rights of children. Barnardos was established in Ireland in 1962 and is Ireland’s leading independent Children's charity.  

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Le modèle développé à l'Institut universitaire de médecine sociale et préventive de Lausanne utilise un programme informatique pour simuler les mouvements d'entrées et de sorties des hôpitaux de soins généraux. Cette simulation se fonde sur les données récoltées de routine dans les hôpitaux; elle tient notamment compte de certaines variations journalières et saisonnières, du nombre d'entrées, ainsi que du "Case-Mix" de l'hôpital, c'est-à-dire de la répartition des cas selon les groupes cliniques et l'âge des patients.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

OBJECTIVES: Human papillomavirus (HPV) is a sexually transmitted infection of particular interest because of its high prevalence rate and strong causal association with cervical cancer. Two prophylactic vaccines have been developed and different countries have made or will soon make recommendations for the vaccination of girls. Even if there is a consensus to recommend a vaccination before the beginning of sexual activity, there are, however, large discrepancies between countries concerning the perceived usefulness of a catch-up procedure and of boosters. The main objective of this article is to simulate the impact on different vaccination policies upon the mid- and long-term HPV 16/18 age-specific infection rates. METHODS: We developed an epidemiological model based on the susceptible-infective-recovered approach using Swiss data. The mid- and long-term impact of different vaccination scenarios was then compared. RESULTS: The generalization of a catch-up procedure is always beneficial, whatever its extent. Moreover, pending on the length of the protection offered by the vaccine, boosters will also be very useful. CONCLUSIONS: To be really effective, a vaccination campaign against HPV infection should at least include a catch-up to early reach a drop in HPV 16/18 prevalence, and maybe boosters. Otherwise, the protection insured for women in their 20s could be lower than expected, resulting in higher risks to later develop cervical cancer.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The dynamic properties of helix 12 in the ligand binding domain of nuclear receptors are a major determinant of AF-2 domain activity. We investigated the molecular and structural basis of helix 12 mobility, as well as the involvement of individual residues with regard to peroxisome proliferator-activated receptor alpha (PPARalpha) constitutive and ligand-dependent transcriptional activity. Functional assays of the activity of PPARalpha helix 12 mutants were combined with free energy molecular dynamics simulations. The agreement between the results from these approaches allows us to make robust claims concerning the mechanisms that govern helix 12 functions. Our data support a model in which PPARalpha helix 12 transiently adopts a relatively stable active conformation even in the absence of a ligand. This conformation provides the interface for the recruitment of a coactivator and results in constitutive activity. The receptor agonists stabilize this conformation and increase PPARalpha transcription activation potential. Finally, we disclose important functions of residues in PPARalpha AF-2, which determine the positioning of helix 12 in the active conformation in the absence of a ligand. Substitution of these residues suppresses PPARalpha constitutive activity, without changing PPARalpha ligand-dependent activation potential.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Our essay aims at studying suitable statistical methods for the clustering ofcompositional data in situations where observations are constituted by trajectories ofcompositional data, that is, by sequences of composition measurements along a domain.Observed trajectories are known as “functional data” and several methods have beenproposed for their analysis.In particular, methods for clustering functional data, known as Functional ClusterAnalysis (FCA), have been applied by practitioners and scientists in many fields. To ourknowledge, FCA techniques have not been extended to cope with the problem ofclustering compositional data trajectories. In order to extend FCA techniques to theanalysis of compositional data, FCA clustering techniques have to be adapted by using asuitable compositional algebra.The present work centres on the following question: given a sample of compositionaldata trajectories, how can we formulate a segmentation procedure giving homogeneousclasses? To address this problem we follow the steps described below.First of all we adapt the well-known spline smoothing techniques in order to cope withthe smoothing of compositional data trajectories. In fact, an observed curve can bethought of as the sum of a smooth part plus some noise due to measurement errors.Spline smoothing techniques are used to isolate the smooth part of the trajectory:clustering algorithms are then applied to these smooth curves.The second step consists in building suitable metrics for measuring the dissimilaritybetween trajectories: we propose a metric that accounts for difference in both shape andlevel, and a metric accounting for differences in shape only.A simulation study is performed in order to evaluate the proposed methodologies, usingboth hierarchical and partitional clustering algorithm. The quality of the obtained resultsis assessed by means of several indices

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes an heuristic for the scheduling of capacity requests and the periodic assignment of radio resources in geostationary (GEO) satellite networks with star topology, using the Demand Assigned Multiple Access (DAMA) protocol in the link layer, and Multi-Frequency Time Division Multiple Access (MF-TDMA) and Adaptive Coding and Modulation (ACM) in the physical layer.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We developed a procedure that combines three complementary computational methodologies to improve the theoretical description of the electronic structure of nickel oxide. The starting point is a Car-Parrinello molecular dynamics simulation to incorporate vibrorotational degrees of freedom into the material model. By means ofcomplete active space self-consistent field second-order perturbation theory (CASPT2) calculations on embedded clusters extracted from the resulting trajectory, we describe localized spectroscopic phenomena on NiO with an efficient treatment of electron correlation. The inclusion of thermal motion into the theoretical description allowsus to study electronic transitions that, otherwise, would be dipole forbidden in the ideal structure and results in a natural reproduction of the band broadening. Moreover, we improved the embedded cluster model by incorporating self-consistently at the complete active space self-consistent field (CASSCF) level a discrete (or direct) reaction field (DRF) in the cluster surroundings. The DRF approach offers an efficient treatment ofelectric response effects of the crystalline embedding to the electronic transitions localized in the cluster. We offer accurate theoretical estimates of the absorption spectrum and the density of states around the Fermi level of NiO, and a comprehensive explanation of the source of the broadening and the relaxation of the charge transferstates due to the adaptation of the environment

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Purpose: To investigate the effect of incremental increases in intraocular straylight on threshold measurements made by three modern forms of perimetry: Standard Automated Perimetry (SAP) using Octopus (Dynamic, G-Pattern), Pulsar Perimetry (PP) (TOP, 66 points) and the Moorfields Motion Displacement Test (MDT) (WEBS, 32 points).Methods: Four healthy young observers were recruited (mean age 26yrs [25yrs, 28yrs]), refractive correction [+2 D, -4.25D]). Five white opacity filters (WOF), each scattering light by different amounts were used to create incremental increases in intraocular straylight (IS). Resultant IS values were measured with each WOF and at baseline (no WOF) for each subject using a C-Quant Straylight Meter (Oculus, Wetzlar, Germany). A 25 yr old has an IS value of ~0.85 log(s). An increase of 40% in IS to 1.2log(s) corresponds to the physiological value of a 70yr old. Each WOFs created an increase in IS between 10-150% from baseline, ranging from effects similar to normal aging to those found with considerable cataract. Each subject underwent 6 test sessions over a 2-week period; each session consisted of the 3 perimetric tests using one of the five WOFs and baseline (both instrument and filter were randomised).Results: The reduction in sensitivity from baseline was calculated. A two-way ANOVA on mean change in threshold (where subjects were treated as rows in the block and each increment in fog filters was treated as column) was used to examine the effect of incremental increases in straylight. Both SAP (p<0.001) and Pulsar (p<0.001) were significantly affected by increases in straylight. The MDT (p=0.35) remained comparatively robust to increases in straylight.Conclusions: The Moorfields MDT measurement of threshold is robust to effects of additional straylight as compared to SAP and PP.