899 resultados para two-Gaussian mixture model
Resumo:
Running economy (RE), i.e. the oxygen consumption at a given submaximal speed, is an important determinant of endurance running performance. So far, investigators have widely attempted to individuate the factors affecting RE in competitive athletes, focusing mainly on the relationships between RE and running biomechanics. However, the current results are inconsistent and a clear mechanical profile of an economic runner has not been yet established. The present work aimed to better understand how the running technique influences RE in sub-elite middle-distance runners by investigating the biomechanical parameters acting on RE and the underlying mechanisms. Special emphasis was given to accounting for intra-individual variability in RE at different speeds and to assessing track running rather than treadmill running. In Study One, a factor analysis was used to reduce the 30 considered mechanical parameters to few global descriptors of the running mechanics. Then, a biomechanical comparison between economic and non economic runners and a multiple regression analysis (with RE as criterion variable and mechanical indices as independent variables) were performed. It was found that a better RE was associated to higher knee and ankle flexion in the support phase, and that the combination of seven individuated mechanical measures explains ∼72% of the variability in RE. In Study Two, a mathematical model predicting RE a priori from the rate of force production, originally developed and used in the field of comparative biology, was adapted and tested in competitive athletes. The model showed a very good fit (R2=0.86). In conclusion, the results of this dissertation suggest that the very complex interrelationships among the mechanical parameters affecting RE may be successfully dealt with through multivariate statistical analyses and the application of theoretical mathematical models. Thanks to these results, coaches are provided with useful tools to assess the biomechanical profile of their athletes. Thus, individual weaknesses in the running technique may be identified and removed, with the ultimate goal to improve RE.
Resumo:
The Assimilation in the Unstable Subspace (AUS) was introduced by Trevisan and Uboldi in 2004, and developed by Trevisan, Uboldi and Carrassi, to minimize the analysis and forecast errors by exploiting the flow-dependent instabilities of the forecast-analysis cycle system, which may be thought of as a system forced by observations. In the AUS scheme the assimilation is obtained by confining the analysis increment in the unstable subspace of the forecast-analysis cycle system so that it will have the same structure of the dominant instabilities of the system. The unstable subspace is estimated by Breeding on the Data Assimilation System (BDAS). AUS- BDAS has already been tested in realistic models and observational configurations, including a Quasi-Geostrophicmodel and a high dimensional, primitive equation ocean model; the experiments include both fixed and“adaptive”observations. In these contexts, the AUS-BDAS approach greatly reduces the analysis error, with reasonable computational costs for data assimilation with respect, for example, to a prohibitive full Extended Kalman Filter. This is a follow-up study in which we revisit the AUS-BDAS approach in the more basic, highly nonlinear Lorenz 1963 convective model. We run observation system simulation experiments in a perfect model setting, and with two types of model error as well: random and systematic. In the different configurations examined, and in a perfect model setting, AUS once again shows better efficiency than other advanced data assimilation schemes. In the present study, we develop an iterative scheme that leads to a significant improvement of the overall assimilation performance with respect also to standard AUS. In particular, it boosts the efficiency of regime’s changes tracking, with a low computational cost. Other data assimilation schemes need estimates of ad hoc parameters, which have to be tuned for the specific model at hand. In Numerical Weather Prediction models, tuning of parameters — and in particular an estimate of the model error covariance matrix — may turn out to be quite difficult. Our proposed approach, instead, may be easier to implement in operational models.
Resumo:
This thesis is dedicated to the analysis of non-linear pricing in oligopoly. Non-linear pricing is a fairly predominant practice in most real markets, mostly characterized by some amount of competition. The sophistication of pricing practices has increased in the latest decades due to the technological advances that have allowed companies to gather more and more data on consumers preferences. The first essay of the thesis highlights the main characteristics of oligopolistic non-linear pricing. Non-linear pricing is a special case of price discrimination. The theory of price discrimination has to be modified in presence of oligopoly: in particular, a crucial role is played by the competitive externality that implies that product differentiation is closely related to the possibility of discriminating. The essay reviews the theory of competitive non-linear pricing by starting from its foundations, mechanism design under common agency. The different approaches to model non-linear pricing are then reviewed. In particular, the difference between price and quantity competition is highlighted. Finally, the close link between non-linear pricing and the recent developments in the theory of vertical differentiation is explored. The second essay shows how the effects of non-linear pricing are determined by the relationship between the demand and the technological structure of the market. The chapter focuses on a model in which firms supply a homogeneous product in two different sizes. Information about consumers' reservation prices is incomplete and the production technology is characterized by size economies. The model provides insights on the size of the products that one finds in the market. Four equilibrium regions are identified depending on the relative intensity of size economies with respect to consumers' evaluation of the good. Regions for which the product is supplied in a single unit or in several different sizes or in only a very large one. Both the private and social desirability of non-linear pricing varies across different equilibrium regions. The third essay considers the broadband internet market. Non discriminatory issues seem the core of the recent debate on the opportunity or not of regulating the internet. One of the main questions posed is whether the telecom companies, owning the networks constituting the internet, should be allowed to offer quality-contingent contracts to content providers. The aim of this essay is to analyze the issue through a stylized two-sided market model of the web that highlights the effects of such a discrimination over quality, prices and participation to the internet of providers and final users. An overall welfare comparison is proposed, concluding that the final effects of regulation crucially depend on both the technology and preferences of agents.
Resumo:
In a large number of problems the high dimensionality of the search space, the vast number of variables and the economical constrains limit the ability of classical techniques to reach the optimum of a function, known or unknown. In this thesis we investigate the possibility to combine approaches from advanced statistics and optimization algorithms in such a way to better explore the combinatorial search space and to increase the performance of the approaches. To this purpose we propose two methods: (i) Model Based Ant Colony Design and (ii) Naïve Bayes Ant Colony Optimization. We test the performance of the two proposed solutions on a simulation study and we apply the novel techniques on an appplication in the field of Enzyme Engineering and Design.
Resumo:
In dieser Arbeit wurden die Phasenübergänge einer einzelnen Polymerkette mit Hilfe der Monte Carlo Methode untersucht. Das Bondfluktuationsmodell wurde zur Simulation benutzt, wobei ein attraktives Kastenpotential zwischen allen Monomeren der Polymerkette gewirkt hat. Drei Arten von Bewegungen sind eingeführt worden, um die Polymerkette richtig zu relaxieren. Diese sind die Hüpfbewegung, die Reptationsbewegung und die Pivotbewegung. Um die Volumenausschlußwechselwirkung zu prüfen und um die Anzahl der Nachbarn jedes Monomers zu bestimmen ist ein hierarchischer Suchalgorithmus eingeführt worden. Die Zustandsdichte des Modells ist mittels des Wang-Landau Algorithmus bestimmt worden. Damit sind thermodynamische Größen berechnet worden, um die Phasenübergänge der einzelnen Polymerkette zu studieren. Wir haben zuerst eine freie Polymerkette untersucht. Der Knäuel-Kügelchen Übergang zeigt sich als ein kontinuierlicher Übergang, bei dem der Knäuel zum Kügelchen zusammenfällt. Der Kügelchen-Kügelchen Übergang bei niedrigeren Temperaturen ist ein Phasenübergang der ersten Ordnung, mit einer Koexistenz des flüssigen und festen Kügelchens, das eine kristalline Struktur hat. Im thermodynamischen Limes sind die Übergangstemperaturen identisch. Das entspricht einem Verschwinden der flüssigen Phase. In zwei Dimensionen zeigt das Modell einen kontinuierlichen Knäuel-Kügelchen Übergang mit einer lokal geordneten Struktur. Wir haben ferner einen Polymermushroom, das ist eine verankerte Polymerkette, zwischen zwei repulsiven Wänden im Abstand D untersucht. Das Phasenverhalten der Polymerkette zeigt einen dimensionalen crossover. Sowohl die Verankerung als auch die Beschränkung fördern den Knäuel-Kügelchen Übergang, wobei es eine Symmetriebrechung gibt, da die Ausdehnung der Polymerkette parallel zu den Wänden schneller schrumpft als die senkrecht zu den Wänden. Die Beschränkung hindert den Kügelchen-Kügelchen Übergang, wobei die Verankerung keinen Einfluss zu haben scheint. Die Übergangstemperaturen im thermodynamischen Limes sind wiederum identisch im Rahmen des Fehlers. Die spezifische Wärme des gleichen Modells aber mit einem abstoßendem Kastenpotential zeigt eine Schottky Anomalie, typisch für ein Zwei-Niveau System.
Resumo:
Conjugated polymers have attracted tremendous academical and industrial research interest over the past decades due to the appealing advantages that organic / polymeric materials offer for electronic applications and devices such as organic light emitting diodes (OLED), organic field effect transistors (OFET), organic solar cells (OSC), photodiodes and plastic lasers. The optimization of organic materials for applications in optoelectronic devices requires detailed knowledge of their photophysical properties, for instance energy levels of excited singlet and triplet states, excited state decay mechanisms and charge carrier mobilities. In the present work a variety of different conjugated (co)polymers, mainly polyspirobifluorene- and polyfluorene-type materials, was investigated using time-resolved photoluminescence spectroscopy in the picosecond to second time domain to study their elementary photophysical properties and to get a deeper insight into structure-property relationships. The experiments cover fluorescence spectroscopy using Streak Camera techniques as well as time-delayed gated detection techniques for the investigation of delayed fluorescence and phosphorescence. All measurements were performed on the solid state, i.e. thin polymer films and on diluted solutions. Starting from the elementary photophysical properties of conjugated polymers the experiments were extended to studies of singlet and triplet energy transfer processes in polymer blends, polymer-triplet emitter blends and copolymers. The phenomenon of photonenergy upconversion was investigated in blue light-emitting polymer matrices doped with metallated porphyrin derivatives supposing an bimolecular annihilation upconversion mechanism which could be experimentally verified on a series of copolymers. This mechanism allows for more efficient photonenergy upconversion than previously reported for polyfluorene derivatives. In addition to the above described spectroscopical experiments, amplified spontaneous emission (ASE) in thin film polymer waveguides was studied employing a fully-arylated poly(indenofluorene) as the gain medium. It was found that the material exhibits a very low threshold value for amplification of blue light combined with an excellent oxidative stability, which makes it interesting as active material for organic solid state lasers. Apart from spectroscopical experiments, transient photocurrent measurements on conjugated polymers were performed as well to elucidate the charge carrier mobility in the solid state, which is an important material parameter for device applications. A modified time-of-flight (TOF) technique using a charge carrier generation layer allowed to study hole transport in a series of spirobifluorene copolymers to unravel the structure-mobility relationship by comparison with the homopolymer. Not only the charge carrier mobility could be determined for the series of polymers but also field- and temperature-dependent measurements analyzed in the framework of the Gaussian disorder model showed that results coincide very well with the predictions of the model. Thus, the validity of the disorder concept for charge carrier transport in amorphous glassy materials could be verified for the investigated series of copolymers.
Resumo:
In this thesis we consider three different models for strongly correlated electrons, namely a multi-band Hubbard model as well as the spinless Falicov-Kimball model, both with a semi-elliptical density of states in the limit of infinite dimensions d, and the attractive Hubbard model on a square lattice in d=2.
In the first part, we study a two-band Hubbard model with unequal bandwidths and anisotropic Hund's rule coupling (J_z-model) in the limit of infinite dimensions within the dynamical mean-field theory (DMFT). Here, the DMFT impurity problem is solved with the use of quantum Monte Carlo (QMC) simulations. Our main result is that the J_z-model describes the occurrence of an orbital-selective Mott transition (OSMT), in contrast to earlier findings. We investigate the model with a high-precision DMFT algorithm, which was developed as part of this thesis and which supplements QMC with a high-frequency expansion of the self-energy.
The main advantage of this scheme is the extraordinary accuracy of the numerical solutions, which can be obtained already with moderate computational effort, so that studies of multi-orbital systems within the DMFT+QMC are strongly improved. We also found that a suitably defined
Falicov-Kimball (FK) model exhibits an OSMT, revealing the close connection of the Falicov-Kimball physics to the J_z-model in the OSM phase.
In the second part of this thesis we study the attractive Hubbard model in two spatial dimensions within second-order self-consistent perturbation theory.
This model is considered on a square lattice at finite doping and at low temperatures. Our main result is that the predictions of first-order perturbation theory (Hartree-Fock approximation) are renormalized by a factor of the order of unity even at arbitrarily weak interaction (U->0). The renormalization factor q can be evaluated as a function of the filling n for 0
Resumo:
This thesis analyses problems related to the applicability, in business environments, of Process Mining tools and techniques. The first contribution is a presentation of the state of the art of Process Mining and a characterization of companies, in terms of their "process awareness". The work continues identifying circumstance where problems can emerge: data preparation; actual mining; and results interpretation. Other problems are the configuration of parameters by not-expert users and computational complexity. We concentrate on two possible scenarios: "batch" and "on-line" Process Mining. Concerning the batch Process Mining, we first investigated the data preparation problem and we proposed a solution for the identification of the "case-ids" whenever this field is not explicitly indicated. After that, we concentrated on problems at mining time and we propose the generalization of a well-known control-flow discovery algorithm in order to exploit non instantaneous events. The usage of interval-based recording leads to an important improvement of performance. Later on, we report our work on the parameters configuration for not-expert users. We present two approaches to select the "best" parameters configuration: one is completely autonomous; the other requires human interaction to navigate a hierarchy of candidate models. Concerning the data interpretation and results evaluation, we propose two metrics: a model-to-model and a model-to-log. Finally, we present an automatic approach for the extension of a control-flow model with social information, in order to simplify the analysis of these perspectives. The second part of this thesis deals with control-flow discovery algorithms in on-line settings. We propose a formal definition of the problem, and two baseline approaches. The actual mining algorithms proposed are two: the first is the adaptation, to the control-flow discovery problem, of a frequency counting algorithm; the second constitutes a framework of models which can be used for different kinds of streams (stationary versus evolving).
Resumo:
Der Haupt-Lichtsammelkomplex (LHCII) des Photosyntheseapparates höherer Pflanzen gehört zu den häufigsten Membranproteinen der Erde. Seine Kristallstruktur ist bekannt. Das Apoprotein kann rekombinant in Escherichia coli überexprimiert und somit molekularbiologisch vielfältig verändert werden. In Detergenzlösung besitzt das denaturierte Protein die erstaunliche Fähigkeit, sich spontan zu funktionalen Protein-Pigment-Komplexen zu organisieren, welche strukturell nahezu identisch sind mit nativem LHCII. Der Faltungsprozess findet in vitro im Zeitbereich von Sekunden bis Minuten statt und ist abhängig von der Bindung der Cofaktoren Chlorophyll a und b sowie verschiedenen Carotinoiden.rn Diese Eigenschaften machen LHCII besonders geeignet für Strukturuntersuchungen mittels der elektronenparamagnetischen Resonanz (EPR)-Spektrokopie. Diese setzt eine punktspezifische Spinmarkierung des LHCII voraus, die in dieser Arbeit zunächst optimiert wurde. Einschließlich der Beiträge Anderer stand eine breite Auswahl von über 40 spinmarkierten Mutanten des LHCII bereit, einen N-terminalen „Cys walk“ eingeschlossen. Weder der hierfür notwendige Austausch einzelner Aminosäuren noch die Anknüpfung des Spinmarkers beeinträchtigten die Funktion des LHCII. Zudem konnte ein Protokoll zur Präparation heterogen spinmarkierter LHCII-Trimere entwickelt werden, also von Trimeren, die jeweils nur ein Monomer mit einer Spinmarkierung enthalten.rn Spinmarkierte Proben des Detergenz-solubilisierten LHCII wurden unter Verwendung verschiedener EPR-Techniken strukturell analysiert. Als besonders aussagekräftig erwies sich die Messung der Wasserzugänglichkeit einzelner Aminosäurepositionen anhand der Electron Spin Echo Envelope Modulation (ESEEM). In Kombination mit der etablierten Double Electron-Electron Resonance (DEER)-Technik zur Detektion von Abständen zwischen zwei Spinmarkern wurde der membranständige Kernbereich des LHCII in Lösung eingehend untersucht und strukturell der Kristallstruktur für sehr ähnlich befunden. Die Vermessung kristallographisch nicht erfasster Bereiche nahe dem N-Terminus offenbarte die schon früher detektierte Strukturdynamik der Domäne in Abhängigkeit des Oligomerisierungsgrades. Der neue, noch zu vervollständigende Datensatz aus Abstandsverteilungen und ESEEM-Wasserzugänglichkeiten monomerer wie trimerer Proben sollte in naher Zukunft die sehr genaue Modellierung der N-terminalen Domäne des LHCII ermöglichen.rn In einem weiteren Abschnitt der Arbeit wurde die Faltung des LHCII-Apoproteins bei der LHCII-Assemblierung in vitro untersucht. Vorausgegangene fluoreszenzspektroskopi-sche Arbeiten hatten gezeigt, dass die Bindung von Chlorophyll a und b in aufeinanderfolgenden Schritten im Zeitbereich von weniger als einer Minute bzw. mehreren Minuten erfolgten. Sowohl die Wasserzugänglichkeit einzelner Aminosäurepositionen als auch Spin-Spin-Abstände änderten sich in ähnlichen Zeitbereichen. Die Daten deuten darauf hin, dass die Ausbildung der mittleren Transmembran-Helix mit der schnelleren Chlorophyll-a-Bindung einhergeht, während sich die Superhelix aus den beiden anderen Transmembranhelices erst im langsameren Schritt, zusammen mit der Chlorophyll-b-Bindung, ausbildet.rn
Resumo:
Stylolites are rough paired surfaces, indicative of localized stress-induced dissolution under a non-hydrostatic state of stress, separated by a clay parting which is believed to be the residuum of the dissolved rock. These structures are the most frequent deformation pattern in monomineralic rocks and thus provide important information about low temperature deformation and mass transfer. The intriguing roughness of stylolites can be used to assess amount of volume loss and paleo-stress directions, and to infer the destabilizing processes during pressure solution. But there is little agreement on how stylolites form and why these localized pressure solution patterns develop their characteristic roughness.rnNatural bedding parallel and vertical stylolites were studied in this work to obtain a quantitative description of the stylolite roughness and understand the governing processes during their formation. Adapting scaling approaches based on fractal principles it is demonstrated that stylolites show two self affine scaling regimes with roughness exponents of 1.1 and 0.5 for small and large length scales separated by a crossover length at the millimeter scale. Analysis of stylolites from various depths proved that this crossover length is a function of the stress field during formation, as analytically predicted. For bedding parallel stylolites the crossover length is a function of the normal stress on the interface, but vertical stylolites show a clear in-plane anisotropy of the crossover length owing to the fact that the in-plane stresses (σ2 and σ3) are dissimilar. Therefore stylolite roughness contains a signature of the stress field during formation.rnTo address the origin of stylolite roughness a combined microstructural (SEM/EBSD) and numerical approach is employed. Microstructural investigations of natural stylolites in limestones reveal that heterogeneities initially present in the host rock (clay particles, quartz grains) are responsible for the formation of the distinctive stylolite roughness. A two-dimensional numerical model, i.e. a discrete linear elastic lattice spring model, is used to investigate the roughness evolving from an initially flat fluid filled interface induced by heterogeneities in the matrix. This model generates rough interfaces with the same scaling properties as natural stylolites. Furthermore two coinciding crossover phenomena in space and in time exist that separate length and timescales for which the roughening is either balanced by surface or elastic energies. The roughness and growth exponents are independent of the size, amount and the dissolution rate of the heterogeneities. This allows to conclude that the location of asperities is determined by a polimict multi-scale quenched noise, while the roughening process is governed by inherent processes i.e. the transition from a surface to an elastic energy dominated regime.rn
Resumo:
In dieser Arbeit werden vier unterschiedliche, stark korrelierte, fermionische Mehrbandsysteme untersucht. Es handelt sich dabei um ein Mehrstörstellen-Anderson-Modell, zwei Hubbard-Modelle sowie ein Mehrbandsystem, wie es sich aus einer ab initio-Beschreibung für ein korreliertes Halbmetall ergibt.rnrnDie Betrachtung des Mehrstörstellen-Anderson-Modells konzentriert sich auf die Untersuchung des Einflusses der Austauschwechselwirkung und der nicht-lokalen Korrelationen zwischen zwei Störstellen in einem einfach-kubischen Gitter. Das zentrale Resultat ist die Abstandsabhängigkeit der Korrelationen der Störstellenelektronen, welche stark von der Gitterdimension und der relativen Position der Störstellen abhängen. Bemerkenswert ist hier die lange Reichweite der Korrelationen in der Diagonalrichtung des Gitters. Außerdem ergibt sich, dass eine antiferromagnetische Austauschwechselwirkung ein Singulett zwischen den Störstellenelektronen gegenüber den Kondo-Singuletts der einzelnen Störstellen favorisiert und so den Kondo-Effekt der einzelnen Störstellen behindert.rnrnEin Zweiband-Hubbard-Modell, das Jz-Modell, wird im Hinblick auf seine Mott-Phasen in Abhängigkeit von Dotierung und Kristallfeldaufspaltung auf dem Bethe-Gitter untersucht. Die Entartung der Bänder ist durch eine unterschiedliche Bandbreite aufgehoben. Wichtigstes Ergebnis sind die Phasendiagramme in Bezug auf Wechselwirkung, Gesamtfüllung und Kristallfeldparameter. Im Vergleich zu Einbandmodellen kommen im Jz-Modell sogenannte orbital-selektive Mott-Phasen hinzu, die, abhängig von Wechselwirkung, Gesamtfüllung und Kristallfeldparameter, einerseits metallischen und andererseits isolierenden Charakter haben. Ein neuer Aspekt ergibt sich durch den Kristallfeldparameter, der die ionischen Einteilchenniveaus relativ zueinander verschiebt, und für bestimmte Werte eine orbital-selektive Mott-Phase des breiten Bands ermöglicht. Im Vergleich mit analytischen Näherungslösungen und Einbandmodellen lassen sich generische Vielteilchen- und Korrelationseffekte von typischen Mehrband- und Einteilcheneffekten differenzieren.rnrnDas zweite untersuchte Hubbard-Modell beschreibt eine magneto-optische Falle mit einer endlichen Anzahl Gitterplätze, in welcher fermionische Atome platziert sind. Es wird eine z-antiferromagnetische Phase unter Berücksichtigung nicht-lokaler Vielteilchenkorrelationen erhalten, und dabei werden bekannte Ergebnisse einer effektiven Einteilchenbeschreibung verbessert.rnrnDas korrelierte Halbmetall wird im Rahmen einer Mehrbandrechnung im Hinblick auf Korrelationseffekte untersucht. Ausgangspunkt ist eine ab initio-Beschreibung durch die Dichtefunktionaltheorie (DFT), welche dann durch die Hinzunahme lokaler Korrelationen ergänzt wird. Die Vielteilcheneffekte werden an Hand einer einfachen Wechselwirkungsnäherung verdeutlicht, und für ein Wechselwirkungsmodell in sphärischer Symmetrie präzisiert. Es ergibt sich nur eine schwache Quasiteilchenrenormierung. Besonders für röntgenspektroskopische Experimente wird eine gute Übereinstimmung erzielt.rnrnDie numerischen Ergebnisse für das Jz-Modell basieren auf Quanten-Monte-Carlo-Simulationen im Rahmen der dynamischen Molekularfeldtheorie (DMFT). Für alle anderen Systeme wird ein Mehrband-Algorithmus entwickelt und implementiert, welcher explizit nicht-diagonale Mehrbandprozesse berücksichtigt.rnrn
Resumo:
Basic concepts and definitions relative to Lagrangian Particle Dispersion Models (LPDMs)for the description of turbulent dispersion are introduced. The study focusses on LPDMs that use as input, for the large scale motion, fields produced by Eulerian models, with the small scale motions described by Lagrangian Stochastic Models (LSMs). The data of two different dynamical model have been used: a Large Eddy Simulation (LES) and a General Circulation Model (GCM). After reviewing the small scale closure adopted by the Eulerian model, the development and implementation of appropriate LSMs is outlined. The basic requirement of every LPDM used in this work is its fullfillment of the Well Mixed Condition (WMC). For the dispersion description in the GCM domain, a stochastic model of Markov order 0, consistent with the eddy-viscosity closure of the dynamical model, is implemented. A LSM of Markov order 1, more suitable for shorter timescales, has been implemented for the description of the unresolved motion of the LES fields. Different assumptions on the small scale correlation time are made. Tests of the LSM on GCM fields suggest that the use of an interpolation algorithm able to maintain an analytical consistency between the diffusion coefficient and its derivative is mandatory if the model has to satisfy the WMC. Also a dynamical time step selection scheme based on the diffusion coefficient shape is introduced, and the criteria for the integration step selection are discussed. Absolute and relative dispersion experiments are made with various unresolved motion settings for the LSM on LES data, and the results are compared with laboratory data. The study shows that the unresolved turbulence parameterization has a negligible influence on the absolute dispersion, while it affects the contribution of the relative dispersion and meandering to absolute dispersion, as well as the Lagrangian correlation.
Resumo:
In this thesis different approaches for the modeling and simulation of the blood protein fibrinogen are presented. The approaches are meant to systematically connect the multiple time and length scales involved in the dynamics of fibrinogen in solution and at inorganic surfaces. The first part of the thesis will cover simulations of fibrinogen on an all atom level. Simulations of the fibrinogen protomer and dimer are performed in explicit solvent to characterize the dynamics of fibrinogen in solution. These simulations reveal an unexpectedly large and fast bending motion that is facilitated by molecular hinges located in the coiled-coil region of fibrinogen. This behavior is characterized by a bending and a dihedral angle and the distribution of these angles is measured. As a consequence of the atomistic detail of the simulations it is possible to illuminate small scale behavior in the binding pockets of fibrinogen that hints at a previously unknown allosteric effect. In a second step atomistic simulations of the fibrinogen protomer are performed at graphite and mica surfaces to investigate initial adsorption stages. These simulations highlight the different adsorption mechanisms at the hydrophobic graphite surface and the charged, hydrophilic mica surface. It is found that the initial adsorption happens in a preferred orientation on mica. Many effects of practical interest involve aggregates of many fibrinogen molecules. To investigate such systems, time and length scales need to be simulated that are not attainable in atomistic simulations. It is therefore necessary to develop lower resolution models of fibrinogen. This is done in the second part of the thesis. First a systematically coarse grained model is derived and parametrized based on the atomistic simulations of the first part. In this model the fibrinogen molecule is represented by 45 beads instead of nearly 31,000 atoms. The intra-molecular interactions of the beads are modeled as a heterogeneous elastic network while inter-molecular interactions are assumed to be a combination of electrostatic and van der Waals interaction. A method is presented that determines the charges assigned to beads by matching the electrostatic potential in the atomistic simulation. Lastly a phenomenological model is developed that represents fibrinogen by five beads connected by rigid rods with two hinges. This model only captures the large scale dynamics in the atomistic simulations but can shed light on experimental observations of fibrinogen conformations at inorganic surfaces.
Resumo:
Thema dieser Arbeit ist die Entwicklung und Kombination verschiedener numerischer Methoden, sowie deren Anwendung auf Probleme stark korrelierter Elektronensysteme. Solche Materialien zeigen viele interessante physikalische Eigenschaften, wie z.B. Supraleitung und magnetische Ordnung und spielen eine bedeutende Rolle in technischen Anwendungen. Es werden zwei verschiedene Modelle behandelt: das Hubbard-Modell und das Kondo-Gitter-Modell (KLM). In den letzten Jahrzehnten konnten bereits viele Erkenntnisse durch die numerische Lösung dieser Modelle gewonnen werden. Dennoch bleibt der physikalische Ursprung vieler Effekte verborgen. Grund dafür ist die Beschränkung aktueller Methoden auf bestimmte Parameterbereiche. Eine der stärksten Einschränkungen ist das Fehlen effizienter Algorithmen für tiefe Temperaturen.rnrnBasierend auf dem Blankenbecler-Scalapino-Sugar Quanten-Monte-Carlo (BSS-QMC) Algorithmus präsentieren wir eine numerisch exakte Methode, die das Hubbard-Modell und das KLM effizient bei sehr tiefen Temperaturen löst. Diese Methode wird auf den Mott-Übergang im zweidimensionalen Hubbard-Modell angewendet. Im Gegensatz zu früheren Studien können wir einen Mott-Übergang bei endlichen Temperaturen und endlichen Wechselwirkungen klar ausschließen.rnrnAuf der Basis dieses exakten BSS-QMC Algorithmus, haben wir einen Störstellenlöser für die dynamische Molekularfeld Theorie (DMFT) sowie ihre Cluster Erweiterungen (CDMFT) entwickelt. Die DMFT ist die vorherrschende Theorie stark korrelierter Systeme, bei denen übliche Bandstrukturrechnungen versagen. Eine Hauptlimitation ist dabei die Verfügbarkeit effizienter Störstellenlöser für das intrinsische Quantenproblem. Der in dieser Arbeit entwickelte Algorithmus hat das gleiche überlegene Skalierungsverhalten mit der inversen Temperatur wie BSS-QMC. Wir untersuchen den Mott-Übergang im Rahmen der DMFT und analysieren den Einfluss von systematischen Fehlern auf diesen Übergang.rnrnEin weiteres prominentes Thema ist die Vernachlässigung von nicht-lokalen Wechselwirkungen in der DMFT. Hierzu kombinieren wir direkte BSS-QMC Gitterrechnungen mit CDMFT für das halb gefüllte zweidimensionale anisotrope Hubbard Modell, das dotierte Hubbard Modell und das KLM. Die Ergebnisse für die verschiedenen Modelle unterscheiden sich stark: während nicht-lokale Korrelationen eine wichtige Rolle im zweidimensionalen (anisotropen) Modell spielen, ist in der paramagnetischen Phase die Impulsabhängigkeit der Selbstenergie für stark dotierte Systeme und für das KLM deutlich schwächer. Eine bemerkenswerte Erkenntnis ist, dass die Selbstenergie sich durch die nicht-wechselwirkende Dispersion parametrisieren lässt. Die spezielle Struktur der Selbstenergie im Impulsraum kann sehr nützlich für die Klassifizierung von elektronischen Korrelationseffekten sein und öffnet den Weg für die Entwicklung neuer Schemata über die Grenzen der DMFT hinaus.
Resumo:
(11)C-ABP-688 is a selective tracer for the mGluR5 receptor. Its kinetics is fast and thus favourable for an equilibrium approach to determine receptor-related parameters. The purpose of this study was to test the hypothesis that the pattern of the (11)C-ABP688 uptake using a bolus-plus-infusion (B/I) protocol at early time points corresponds to the perfusion and at a later time point to the total distribution volume. METHODS: A bolus and a B/I study (1 h each) was performed in five healthy male volunteers. With the B/I protocol, early and late scans were normalized to gray matter, cerebellum and white matter. The same normalization was done on the maps of the total distribution volume (Vt) and K(1) which were calculated in the study with bolus only injection and the Logan method (Vt) and a two-tissue compartment model (K(1)). RESULTS: There was an excellent correlation close to the identity line between the pattern of the late uptake in the B/I study and Vt of the bolus-only study for all three normalizations. The pattern of the early uptake in the B/I study correlated well with the K(1) maps, but only when normalized to gray matter and cerebellum, not to white matter. CONCLUSION: It is demonstrated that with a B/I protocol the (11)C-ABP688 distribution in late scans reflects the pattern of the total distribution volume and is therefore a measure for the density pattern of mGluR5. The early scans following injection are related to blood flow, although not in a fully quantitative manner. The advantage of the B/I protocol is that no arterial blood sampling is required, which is advantageous in clinical studies.