190 resultados para Agent Oriented Modeling
em Université de Lausanne, Switzerland
Resumo:
Computational modeling has become a widely used tool for unraveling the mechanisms of higher level cooperative cell behavior during vascular morphogenesis. However, experimenting with published simulation models or adding new assumptions to those models can be daunting for novice and even for experienced computational scientists. Here, we present a step-by-step, practical tutorial for building cell-based simulations of vascular morphogenesis using the Tissue Simulation Toolkit (TST). The TST is a freely available, open-source C++ library for developing simulations with the two-dimensional cellular Potts model, a stochastic, agent-based framework to simulate collective cell behavior. We will show the basic use of the TST to simulate and experiment with published simulations of vascular network formation. Then, we will present step-by-step instructions and explanations for building a recent simulation model of tumor angiogenesis. Demonstrated mechanisms include cell-cell adhesion, chemotaxis, cell elongation, haptotaxis, and haptokinesis.
Resumo:
Methods like Event History Analysis can show the existence of diffusion and part of its nature, but do not study the process itself. Nowadays, thanks to the increasing performance of computers, processes can be studied using computational modeling. This thesis presents an agent-based model of policy diffusion mainly inspired from the model developed by Braun and Gilardi (2006). I first start by developing a theoretical framework of policy diffusion that presents the main internal drivers of policy diffusion - such as the preference for the policy, the effectiveness of the policy, the institutional constraints, and the ideology - and its main mechanisms, namely learning, competition, emulation, and coercion. Therefore diffusion, expressed by these interdependencies, is a complex process that needs to be studied with computational agent-based modeling. In a second step, computational agent-based modeling is defined along with its most significant concepts: complexity and emergence. Using computational agent-based modeling implies the development of an algorithm and its programming. When this latter has been developed, we let the different agents interact. Consequently, a phenomenon of diffusion, derived from learning, emerges, meaning that the choice made by an agent is conditional to that made by its neighbors. As a result, learning follows an inverted S-curve, which leads to partial convergence - global divergence and local convergence - that triggers the emergence of political clusters; i.e. the creation of regions with the same policy. Furthermore, the average effectiveness in this computational world tends to follow a J-shaped curve, meaning that not only time is needed for a policy to deploy its effects, but that it also takes time for a country to find the best-suited policy. To conclude, diffusion is an emergent phenomenon from complex interactions and its outcomes as ensued from my model are in line with the theoretical expectations and the empirical evidence.Les méthodes d'analyse de biographie (event history analysis) permettent de mettre en évidence l'existence de phénomènes de diffusion et de les décrire, mais ne permettent pas d'en étudier le processus. Les simulations informatiques, grâce aux performances croissantes des ordinateurs, rendent possible l'étude des processus en tant que tels. Cette thèse, basée sur le modèle théorique développé par Braun et Gilardi (2006), présente une simulation centrée sur les agents des phénomènes de diffusion des politiques. Le point de départ de ce travail met en lumière, au niveau théorique, les principaux facteurs de changement internes à un pays : la préférence pour une politique donnée, l'efficacité de cette dernière, les contraintes institutionnelles, l'idéologie, et les principaux mécanismes de diffusion que sont l'apprentissage, la compétition, l'émulation et la coercition. La diffusion, définie par l'interdépendance des différents acteurs, est un système complexe dont l'étude est rendue possible par les simulations centrées sur les agents. Au niveau méthodologique, nous présenterons également les principaux concepts sous-jacents aux simulations, notamment la complexité et l'émergence. De plus, l'utilisation de simulations informatiques implique le développement d'un algorithme et sa programmation. Cette dernière réalisée, les agents peuvent interagir, avec comme résultat l'émergence d'un phénomène de diffusion, dérivé de l'apprentissage, où le choix d'un agent dépend en grande partie de ceux faits par ses voisins. De plus, ce phénomène suit une courbe en S caractéristique, poussant à la création de régions politiquement identiques, mais divergentes au niveau globale. Enfin, l'efficacité moyenne, dans ce monde simulé, suit une courbe en J, ce qui signifie qu'il faut du temps, non seulement pour que la politique montre ses effets, mais également pour qu'un pays introduise la politique la plus efficace. En conclusion, la diffusion est un phénomène émergent résultant d'interactions complexes dont les résultats du processus tel que développé dans ce modèle correspondent tant aux attentes théoriques qu'aux résultats pratiques.
Resumo:
Because of the increase in workplace automation and the diversification of industrial processes, workplaces have become more and more complex. The classical approaches used to address workplace hazard concerns, such as checklists or sequence models, are, therefore, of limited use in such complex systems. Moreover, because of the multifaceted nature of workplaces, the use of single-oriented methods, such as AEA (man oriented), FMEA (system oriented), or HAZOP (process oriented), is not satisfactory. The use of a dynamic modeling approach in order to allow multiple-oriented analyses may constitute an alternative to overcome this limitation. The qualitative modeling aspects of the MORM (man-machine occupational risk modeling) model are discussed in this article. The model, realized on an object-oriented Petri net tool (CO-OPN), has been developed to simulate and analyze industrial processes in an OH&S perspective. The industrial process is modeled as a set of interconnected subnets (state spaces), which describe its constitutive machines. Process-related factors are introduced, in an explicit way, through machine interconnections and flow properties. While man-machine interactions are modeled as triggering events for the state spaces of the machines, the CREAM cognitive behavior model is used in order to establish the relevant triggering events. In the CO-OPN formalism, the model is expressed as a set of interconnected CO-OPN objects defined over data types expressing the measure attached to the flow of entities transiting through the machines. Constraints on the measures assigned to these entities are used to determine the state changes in each machine. Interconnecting machines implies the composition of such flow and consequently the interconnection of the measure constraints. This is reflected by the construction of constraint enrichment hierarchies, which can be used for simulation and analysis optimization in a clear mathematical framework. The use of Petri nets to perform multiple-oriented analysis opens perspectives in the field of industrial risk management. It may significantly reduce the duration of the assessment process. But, most of all, it opens perspectives in the field of risk comparisons and integrated risk management. Moreover, because of the generic nature of the model and tool used, the same concepts and patterns may be used to model a wide range of systems and application fields.
Resumo:
Abstract in English : Ubiquitous Computing is the emerging trend in computing systems. Based on this observation this thesis proposes an analysis of the hardware and environmental constraints that rule pervasive platforms. These constraints have a strong impact on the programming of such platforms. Therefore solutions are proposed to facilitate this programming both at the platform and node levels. The first contribution presented in this document proposes a combination of agentoriented programming with the principles of bio-inspiration (Phylogenesys, Ontogenesys and Epigenesys) to program pervasive platforms such as the PERvasive computing framework for modeling comPLEX virtually Unbounded Systems platform. The second contribution proposes a method to program efficiently parallelizable applications on each computing node of this platform. Résumé en Français : Basée sur le constat que les calculs ubiquitaires vont devenir le paradigme de programmation dans les années à venir, cette thèse propose une analyse des contraintes matérielles et environnementale auxquelles sont soumises les plateformes pervasives. Ces contraintes ayant un impact fort sur la programmation des plateformes. Des solutions sont donc proposées pour faciliter cette programmation tant au niveau de l'ensemble des noeuds qu'au niveau de chacun des noeuds de la plateforme. La première contribution présentée dans ce document propose d'utiliser une alliance de programmation orientée agent avec les grands principes de la bio-inspiration (Phylogénèse, Ontogénèse et Épigénèse). Ceci pour répondres aux contraintes de programmation de plateformes pervasives comme la plateforme PERvasive computing framework for modeling comPLEX virtually Unbounded Systems . La seconde contribution propose quant à elle une méthode permettant de programmer efficacement des applications parallélisable sur chaque noeud de calcul de la plateforme
Resumo:
BACKGROUND: An ADME (absorption, distribution, metabolism and excretion)-pharmacogenetics association study may identify functional variants relevant to the pharmacokinetics of lopinavir co-formulated with ritonavir (LPV/r), a first-line anti-HIV agent. METHODS: An extensive search of literature and web resources helped select ADME genes and single nucleotide polymorphisms (SNPs, functional and HapMap tagging SNPs) with a proven or potentially relevant role in LPV/r pharmacokinetics. The study followed a two-stage design. Stage 1 (discovery) considered a Caucasian population (n=638) receiving LPV/r, where we selected 117 individuals with low LPV clearance (cases) and 90 individuals with high clearance (controls). Genotyping was performed by a 1536-SNP customized GoldenGate Illumina BeadArray. Stage 2 (confirmation) represented a replication study of candidate SNPs from the stage 1 in 148 individuals receiving LPV/r. The analysis led to formal population pharmacokinetic-pharmacogenetic modeling of demographic, environmental and candidate SNP effects. RESULTS: One thousand three hundred and eighty SNPs were successfully genotyped. Nine SNPs prioritized by the stage 1 analysis were brought to replication. Stage 2 confirmed the contribution of two functional SNPs in SLCO1B1, one functional SNP in ABCC2 and a tag SNP of the CYP3A locus in addition to body weight effect and ritonavir coadministration. According to the population pharmacokinetic-pharmacogenetic model, genetic variants explained 5% of LPV variability. Individuals homozygous rs11045819 (SLCO1B1*4) had a clearance of 12.6 l/h, compared with 5.4 l/h in the reference group, and 3.9 l/h in individuals with two or more variant alleles of rs4149056 (SLCO1B1*5), rs717620 (ABCC2) or rs6945984 (CYP3A). A subanalysis confirmed that although a significant part of the variance in LPV clearance was attributed to fluctuation in ritonavir levels, genetic variants had an additional effect on LPV clearance. CONCLUSION: The two-stage strategy successfully identified genetic variants affecting LPV/r pharmacokinetics. Such a general approach of ADME pharmacogenetics should be generalized to other drugs.
Resumo:
The paper presents some contemporary approaches to spatial environmental data analysis. The main topics are concentrated on the decision-oriented problems of environmental spatial data mining and modeling: valorization and representativity of data with the help of exploratory data analysis, spatial predictions, probabilistic and risk mapping, development and application of conditional stochastic simulation models. The innovative part of the paper presents integrated/hybrid model-machine learning (ML) residuals sequential simulations-MLRSS. The models are based on multilayer perceptron and support vector regression ML algorithms used for modeling long-range spatial trends and sequential simulations of the residuals. NIL algorithms deliver non-linear solution for the spatial non-stationary problems, which are difficult for geostatistical approach. Geostatistical tools (variography) are used to characterize performance of ML algorithms, by analyzing quality and quantity of the spatially structured information extracted from data with ML algorithms. Sequential simulations provide efficient assessment of uncertainty and spatial variability. Case study from the Chernobyl fallouts illustrates the performance of the proposed model. It is shown that probability mapping, provided by the combination of ML data driven and geostatistical model based approaches, can be efficiently used in decision-making process. (C) 2003 Elsevier Ltd. All rights reserved.
Resumo:
Approximate models (proxies) can be employed to reduce the computational costs of estimating uncertainty. The price to pay is that the approximations introduced by the proxy model can lead to a biased estimation. To avoid this problem and ensure a reliable uncertainty quantification, we propose to combine functional data analysis and machine learning to build error models that allow us to obtain an accurate prediction of the exact response without solving the exact model for all realizations. We build the relationship between proxy and exact model on a learning set of geostatistical realizations for which both exact and approximate solvers are run. Functional principal components analysis (FPCA) is used to investigate the variability in the two sets of curves and reduce the dimensionality of the problem while maximizing the retained information. Once obtained, the error model can be used to predict the exact response of any realization on the basis of the sole proxy response. This methodology is purpose-oriented as the error model is constructed directly for the quantity of interest, rather than for the state of the system. Also, the dimensionality reduction performed by FPCA allows a diagnostic of the quality of the error model to assess the informativeness of the learning set and the fidelity of the proxy to the exact model. The possibility of obtaining a prediction of the exact response for any newly generated realization suggests that the methodology can be effectively used beyond the context of uncertainty quantification, in particular for Bayesian inference and optimization.
Resumo:
Na,K-ATPase is the main active transport system that maintains the large gradients of Na(+) and K(+) across the plasma membrane of animal cells. The crystal structure of a K(+)-occluding conformation of this protein has been recently published, but the movements of its different domains allowing for the cation pumping mechanism are not yet known. The structure of many more conformations is known for the related calcium ATPase SERCA, but the reliability of homology modeling is poor for several domains with low sequence identity, in particular the extracellular loops. To better define the structure of the large fourth extracellular loop between the seventh and eighth transmembrane segments of the alpha subunit, we have studied the formation of a disulfide bond between pairs of cysteine residues introduced by site-directed mutagenesis in the second and the fourth extracellular loop. We found a specific pair of cysteine positions (Y308C and D884C) for which extracellular treatment with an oxidizing agent inhibited the Na,K pump function, which could be rapidly restored by a reducing agent. The formation of the disulfide bond occurred preferentially under the E2-P conformation of Na,K-ATPase, in the absence of extracellular cations. Using recently published crystal structure and a distance constraint reproducing the existence of disulfide bond, we performed an extensive conformational space search using simulated annealing and showed that the Tyr(308) and Asp(884) residues can be in close proximity, and simultaneously, the SYGQ motif of the fourth extracellular loop, known to interact with the extracellular domain of the beta subunit, can be exposed to the exterior of the protein and can easily interact with the beta subunit.
Ab initio modeling and molecular dynamics simulation of the alpha 1b-adrenergic receptor activation.
Resumo:
This work describes the ab initio procedure employed to build an activation model for the alpha 1b-adrenergic receptor (alpha 1b-AR). The first version of the model was progressively modified and complicated by means of a many-step iterative procedure characterized by the employment of experimental validations of the model in each upgrading step. A combined simulated (molecular dynamics) and experimental mutagenesis approach was used to determine the structural and dynamic features characterizing the inactive and active states of alpha 1b-AR. The latest version of the model has been successfully challenged with respect to its ability to interpret and predict the functional properties of a large number of mutants. The iterative approach employed to describe alpha 1b-AR activation in terms of molecular structure and dynamics allows further complications of the model to allow prediction and interpretation of an ever-increasing number of experimental data.
Resumo:
Among the largest resources for biological sequence data is the large amount of expressed sequence tags (ESTs) available in public and proprietary databases. ESTs provide information on transcripts but for technical reasons they often contain sequencing errors. Therefore, when analyzing EST sequences computationally, such errors must be taken into account. Earlier attempts to model error prone coding regions have shown good performance in detecting and predicting these while correcting sequencing errors using codon usage frequencies. In the research presented here, we improve the detection of translation start and stop sites by integrating a more complex mRNA model with codon usage bias based error correction into one hidden Markov model (HMM), thus generalizing this error correction approach to more complex HMMs. We show that our method maintains the performance in detecting coding sequences.
Resumo:
Background: Motive-oriented therapeutic relationship (MOTR) was postulated to be a particularly helpful therapeutic ingredient in the early treatment phase of patients with personality disorders, in particular with borderline personality disorder (BPD). The present randomized controlled study using an add-on design is the first study to test this assumption in a 10-session general psychiatric treatment with patients presenting with BPD on symptom reduction and therapeutic alliance. Methods: A total of 85 patients were randomized. They were either allocated to a manual-based short variant of the general psychiatric management (GPM) treatment (in 10 sessions) or to the same treatment where MOTR was deliberately added to the treatment. Treatment attrition and integrity analyses yielded satisfactory results. Results: The results of the intent-to-treat analyses suggested a global efficacy of MOTR, in the sense of an additional reduction of general problems, i.e. symptoms, interpersonal and social problems (F1, 73 = 7.25, p < 0.05). However, they also showed that MOTR did not yield an additional reduction of specific borderline symptoms. It was also shown that a stronger therapeutic alliance, as assessed by the therapist, developed in MOTR treatments compared to GPM (Z55 = 0.99, p < 0.04). Conclusions: These results suggest that adding MOTR to psychiatric and psychotherapeutic treatments of BPD is promising. Moreover, the findings shed additional light on the perspective of shortening treatments for patients presenting with BPD. © 2014 S. Karger AG, Basel.
Resumo:
The Mont Collon mafic complex is one of the best preserved examples of the Early Permian magmatism in the Central Alps, related to the intra-continental collapse of the Variscan belt. It mostly consists (> 95 vol.%) of ol+hy-nonnative plagioclase-wehrlites, olivine- and cpx-gabbros with cumulitic structures, crosscut by acid dikes. Pegmatitic gabbros, troctolites and anorthosites outcrop locally. A well-preserved cumulative, sequence is exposed in the Dents de Bertol area (center of intrusion). PT-calculations indicate that this layered magma chamber emplaced at mid-crustal levels at about 0.5 GPa and 1100 degrees C. The Mont Collon cumulitic rocks record little magmatic differentiation, as illustrated by the restricted range of clinopyroxene mg-number (Mg#(cpx)=83-89). Whole-rock incompatible trace-element contents (e.g. Nb, Zr, Ba) vary largely and without correlation with major-element composition. These features are characteristic of an in-situ crystallization process with variable amounts of interstitial liquid L trapped between the cumulus mineral phases. LA-ICPMS measurements show that trace-element distribution in the latter is homogeneous, pointing to subsolidus re-equilibration between crystals and interstitial melts. A quantitative modeling based on Langmuir's in-situ crystallization equation successfully duplicated the REE concentrations in cumulitic minerals of all rock facies of the intrusion. The calculated amounts of interstitial liquid L vary between 0 and 35% for degrees of differentiation F of 0 to 20%, relative to the least evolved facies of the intrusion. L values are well correlated with the modal proportions of interstitial amphibole and whole-rock incompatible trace-element concentrations (e.g. Zr, Nb) of the tested samples. However, the in-situ crystallization model reaches its limitations with rock containing high modal content of REE-bearing minerals (i.e. zircon), such as pegmatitic gabbros. Dikes of anorthositic composition, locally crosscutting the layered lithologies, evidence that the Mont Collon rocks evolved in open system with mixing of intercumulus liquids of different origins and possibly contrasting compositions. The proposed model is not able to resolve these complex open systems, but migrating liquids could be partly responsible for the observed dispersion of points in some correlation diagrams. Absence of significant differentiation with recurrent lithologies in the cumulitic pile of Dents de Bertol points to an efficiently convective magma chamber, with possible periodic replenishment, (c) 2005 Elsevier B.V. All rights reserved.
Resumo:
The transition from wakefulness to sleep represents the most conspicuous change in behavior and the level of consciousness occurring in the healthy brain. It is accompanied by similarly conspicuous changes in neural dynamics, traditionally exemplified by the change from "desynchronized" electroencephalogram activity in wake to globally synchronized slow wave activity of early sleep. However, unit and local field recordings indicate that the transition is more gradual than it might appear: On one hand, local slow waves already appear during wake; on the other hand, slow sleep waves are only rarely global. Studies with functional magnetic resonance imaging also reveal changes in resting-state functional connectivity (FC) between wake and slow wave sleep. However, it remains unclear how resting-state networks may change during this transition period. Here, we employ large-scale modeling of the human cortico-cortical anatomical connectivity to evaluate changes in resting-state FC when the model "falls asleep" due to the progressive decrease in arousal-promoting neuromodulation. When cholinergic neuromodulation is parametrically decreased, local slow waves appear, while the overall organization of resting-state networks does not change. Furthermore, we show that these local slow waves are structured macroscopically in networks that resemble the resting-state networks. In contrast, when the neuromodulator decrease further to very low levels, slow waves become global and resting-state networks merge into a single undifferentiated, broadly synchronized network.