1000 resultados para SCALE PHYSICS


Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the present work, a multi physics simulation of an innovative safety system for light water nuclear reactor is performed, with the aim to increase the reliability of its main decay heat removal system. The system studied, denoted by the acronym PERSEO (in Pool Energy Removal System for Emergency Operation) is able to remove the decay power from the primary side of the light water nuclear reactor through a heat suppression pool. The experimental facility, located at SIET laboratories (PIACENZA), is an evolution of the Thermal Valve concept where the triggering valve is installed liquid side, on a line connecting two pools at the bottom. During the normal operation, the valve is closed, while in emergency conditions it opens, the heat exchanger is flooded with consequent heat transfer from the primary side to the pool side. In order to verify the correct system behavior during long term accidental transient, two main experimental PERSEO tests are analyzed. For this purpose, a coupling between the mono dimensional system code CATHARE, which reproduces the system scale behavior, with a three-dimensional CFD code NEPTUNE CFD, allowing a full investigation of the pools and the injector, is implemented. The coupling between the two codes is realized through the boundary conditions. In a first analysis, the facility is simulated by the system code CATHARE V2.5 to validate the results with the experimental data. The comparison of the numerical results obtained shows a different void distribution during the boiling conditions inside the heat suppression pool for the two cases of single nodalization and three volume nodalization scheme of the pool. Finaly, to improve the investigation capability of the void distribution inside the pool and the temperature stratification phenomena below the injector, a two and three dimensional CFD models with a simplified geometry of the system are adopted.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this thesis, elemental research towards the implantation of a diamond-based molecular quantum computer is presented. The approach followed requires linear alignment of endohedral fullerenes on the diamond C(100) surface in the vicinity of subsurface NV-centers. From this, four fundamental experimental challenges arise: 1) The well-controlled deposition of endohedral fullerenes on a diamond surface. 2) The creation of NV-centers in diamond close to the surface. 3) Preparation and characterization of atomically-flat diamondsurfaces. 4) Assembly of linear chains of endohedral fullerenes. First steps to overcome all these challenges were taken in the framework of this thesis. Therefore, a so-called “pulse injection” technique was implemented and tested in a UHV chamber that was custom-designed for this and further tasks. Pulse injection in principle allows for the deposition of molecules from solution onto a substrate and can therefore be used to deposit molecular species that are not stable to sublimation under UHV conditions, such as the endohedral fullerenes needed for a quantum register. Regarding the targeted creation of NV-centers, FIB experiments were carried out in cooperation with the group of Prof. Schmidt-Kaler (AG Quantum, Physics Department, Johannes Gutenberg-Universität Mainz). As an entry into this challenging task, argon cations were implanted into (111) surface-oriented CaF2 crystals. The resulting implantation spots on the surface were imaged and characterized using AFM. In this context, general relations between the impact of the ions on the surface and their valency or kinetic energy, respectively, could be established. The main part of this thesis, however, is constituted by NCAFM studies on both, bare and hydrogen-terminated diamond C(100) surfaces. In cooperation with the group of Prof. Dujardin (Molecular Nanoscience Group, ISMO, Université de Paris XI), clean and atomically-flat diamond surfaces were prepared by exposure of the substrate to a microwave hydrogen plasma. Subsequently, both surface modifications were imaged in high resolution with NC-AFM. In the process, both hydrogen atoms in the unit cell of the hydrogenated surface were resolved individually, which was not achieved in previous STM studies of this surface. The NC-AFM images also reveal, for the first time, atomic-resolution contrast on the clean, insulating diamond surface and provide real-space experimental evidence for a (2×1) surface reconstruction. With regard to the quantum computing concept, high-resolution NC-AFM imaging was also used to study the adsorption and self-assembly potential of two different kinds of fullerenes (C60 and C60F48) on aforementioned diamond surfaces. In case of the hydrogenated surface, particular attention was paid to the influence of charge transfer doping on the fullerene-substrate interaction and the morphology emerging from self-assembly. Finally, self-assembled C60 islands on the hydrogen-terminated diamond surface were subject to active manipulation by an NC-AFM tip. Two different kinds of tip-induced island growth modes have been induced and were presented. In conclusion, the results obtained provide fundamental informations mandatory for the realization of a molecular quantum computer. In the process it was shown that NC-AFM is, under proper circumstances, a very capable tool for imaging diamond surfaces with highest resolution, surpassing even what has been achieved with STM up to now. Particular attention was paid to the influence of transfer doping on the morphology of fullerenes on the hydrogenated diamond surface, revealing new possibilities for tailoring the self-assembly of molecules that have a high electron affinity.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In dieser Arbeit wurden Simulation von Flüssigkeiten auf molekularer Ebene durchgeführt, wobei unterschiedliche Multi-Skalen Techniken verwendet wurden. Diese erlauben eine effektive Beschreibung der Flüssigkeit, die weniger Rechenzeit im Computer benötigt und somit Phänomene auf längeren Zeit- und Längenskalen beschreiben kann.rnrnEin wesentlicher Aspekt ist dabei ein vereinfachtes (“coarse-grained”) Modell, welches in einem systematischen Verfahren aus Simulationen des detaillierten Modells gewonnen wird. Dabei werden ausgewählte Eigenschaften des detaillierten Modells (z.B. Paar-Korrelationsfunktion, Druck, etc) reproduziert.rnrnEs wurden Algorithmen untersucht, die eine gleichzeitige Kopplung von detaillierten und vereinfachten Modell erlauben (“Adaptive Resolution Scheme”, AdResS). Dabei wird das detaillierte Modell in einem vordefinierten Teilvolumen der Flüssigkeit (z.B. nahe einer Oberfläche) verwendet, während der Rest mithilfe des vereinfachten Modells beschrieben wird.rnrnHierzu wurde eine Methode (“Thermodynamische Kraft”) entwickelt um die Kopplung auch dann zu ermöglichen, wenn die Modelle in verschiedenen thermodynamischen Zuständen befinden. Zudem wurde ein neuartiger Algorithmus der Kopplung beschrieben (H-AdResS) der die Kopplung mittels einer Hamilton-Funktion beschreibt. In diesem Algorithmus ist eine zur Thermodynamischen Kraft analoge Korrektur mit weniger Rechenaufwand möglich.rnrnAls Anwendung dieser grundlegenden Techniken wurden Pfadintegral Molekulardynamik (MD) Simulationen von Wasser untersucht. Mithilfe dieser Methode ist es möglich, quantenmechanische Effekte der Kerne (Delokalisation, Nullpunktsenergie) in die Simulation einzubeziehen. Hierbei wurde zuerst eine Multi-Skalen Technik (“Force-matching”) verwendet um eine effektive Wechselwirkung aus einer detaillierten Simulation auf Basis der Dichtefunktionaltheorie zu extrahieren. Die Pfadintegral MD Simulation verbessert die Beschreibung der intra-molekularen Struktur im Vergleich mit experimentellen Daten. Das Modell eignet sich auch zur gleichzeitigen Kopplung in einer Simulation, wobei ein Wassermolekül (beschrieben durch 48 Punktteilchen im Pfadintegral-MD Modell) mit einem vereinfachten Modell (ein Punktteilchen) gekoppelt wird. Auf diese Weise konnte eine Wasser-Vakuum Grenzfläche simuliert werden, wobei nur die Oberfläche im Pfadintegral Modell und der Rest im vereinfachten Modell beschrieben wird.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

PhEDEx, the CMS transfer management system, during the first LHC Run has moved about 150 PB and currently it is moving about 2.5 PB of data per week over the Worldwide LHC Computing Grid (WLGC). It was designed to complete each transfer required by users at the expense of the waiting time necessary for its completion. For this reason, after several years of operations, data regarding transfer latencies has been collected and stored into log files containing useful analyzable informations. Then, starting from the analysis of several typical CMS transfer workflows, a categorization of such latencies has been made with a focus on the different factors that contribute to the transfer completion time. The analysis presented in this thesis will provide the necessary information for equipping PhEDEx in the future with a set of new tools in order to proactively identify and fix any latency issues. PhEDEx, il sistema di gestione dei trasferimenti di CMS, durante il primo Run di LHC ha trasferito all’incirca 150 PB ed attualmente trasferisce circa 2.5 PB di dati alla settimana attraverso la Worldwide LHC Computing Grid (WLCG). Questo sistema è stato progettato per completare ogni trasferimento richiesto dall’utente a spese del tempo necessario per il suo completamento. Dopo svariati anni di operazioni con tale strumento, sono stati raccolti dati relativi alle latenze di trasferimento ed immagazzinati in log files contenenti informazioni utili per l’analisi. A questo punto, partendo dall’analisi di una ampia mole di trasferimenti in CMS, è stata effettuata una suddivisione di queste latenze ponendo particolare attenzione nei confronti dei fattori che contribuiscono al tempo di completamento del trasferimento. L’analisi presentata in questa tesi permetterà di equipaggiare PhEDEx con un insieme di utili strumenti in modo tale da identificare proattivamente queste latenze e adottare le opportune tattiche per minimizzare l’impatto sugli utenti finali.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Image-based modeling of tumor growth combines methods from cancer simulation and medical imaging. In this context, we present a novel approach to adapt a healthy brain atlas to MR images of tumor patients. In order to establish correspondence between a healthy atlas and a pathologic patient image, tumor growth modeling in combination with registration algorithms is employed. In a first step, the tumor is grown in the atlas based on a new multi-scale, multi-physics model including growth simulation from the cellular level up to the biomechanical level, accounting for cell proliferation and tissue deformations. Large-scale deformations are handled with an Eulerian approach for finite element computations, which can operate directly on the image voxel mesh. Subsequently, dense correspondence between the modified atlas and patient image is established using nonrigid registration. The method offers opportunities in atlasbased segmentation of tumor-bearing brain images as well as for improved patient-specific simulation and prognosis of tumor progression.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Previous studies have highlighted the severity of detrimental effects for life on earth after an assumed regionally limited nuclear war. These effects are caused by climatic, chemical and radiative changes persisting for up to one decade. However, so far only a very limited number of climate model simulations have been performed, giving rise to the question how realistic previous computations have been. This study uses the coupled chemistry climate model (CCM) SOCOL, which belongs to a different family of CCMs than previously used, to investigate the consequences of such a hypothetical nuclear conflict. In accordance with previous studies, the present work assumes a scenario of a nuclear conflict between India and Pakistan, each applying 50 warheads with an individual blasting power of 15 kt ("Hiroshima size") against the major population centers, resulting in the emission of tiny soot particles, which are generated in the firestorms expected in the aftermath of the detonations. Substantial uncertainties related to the calculation of likely soot emissions, particularly concerning assumptions of target fuel loading and targeting of weapons, have been addressed by simulating several scenarios, with soot emissions ranging from 1 to 12 Tg. Their high absorptivity with respect to solar radiation leads to a rapid self-lofting of the soot particles into the strato- and mesosphere within a few days after emission, where they remain for several years. Consequently, the model suggests earth's surface temperatures to drop by several degrees Celsius due to the shielding of solar irradiance by the soot, indicating a major global cooling. In addition, there is a substantial reduction of precipitation lasting 5 to 10 yr after the conflict, depending on the magnitude of the initial soot release. Extreme cold spells associated with an increase in sea ice formation are found during Northern Hemisphere winter, which expose the continental land masses of North America and Eurasia to a cooling of several degrees. In the stratosphere, the strong heating leads to an acceleration of catalytic ozone loss and, consequently, to enhancements of UV radiation at the ground. In contrast to surface temperature and precipitation changes, which show a linear dependence to the soot burden, there is a saturation effect with respect to stratospheric ozone chemistry. Soot emissions of 5 Tg lead to an ozone column reduction of almost 50% in northern high latitudes, while emitting 12 Tg only increases ozone loss by a further 10%. In summary, this study, though using a different chemistry climate model, corroborates the previous investigations with respect to the atmospheric impacts. In addition to these persistent effects, the present study draws attention to episodically cold phases, which would likely add to the severity of human harm worldwide. The best insurance against such a catastrophic development would be the delegitimization of nuclear weapons.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Asteroid 4Vesta seems to be a major intact protoplanet, with a surface composition similar to that of the HED (howardite-eucrite-diogenite) meteorites. The southern hemisphere is dominated by a giant impact scar, but previous impact models have failed to reproduce the observed topography. The recent discovery that Vesta's southern hemisphere is dominated by two overlapping basins provides an opportunity to model Vesta's topography more accurately. Here we report three-dimensional simulations of Vesta's global evolution under two overlapping planet-scale collisions. We closely reproduce its observed shape, and provide maps of impact excavation and ejecta deposition. Spiral patterns observed in the younger basin Rheasilvia, about one billion years old, are attributed to Coriolis forces during crater collapse. Surface materials exposed in the north come from a depth of about 20kilometres, according to our models, whereas materials exposed inside the southern double-excavation come from depths of about 60-100kilometres. If Vesta began as a layered, completely differentiated protoplanet, then our model predicts large areas of pure diogenites and olivine-rich rocks. These are not seen, possibly implying that the outer 100kilometres or so of Vesta is composed mainly of a basaltic crust (eucrites) with ultramafic intrusions (diogenites).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Understanding natural climate variability and its driving factors is crucial to assessing future climate change. Therefore, comparing proxy-based climate reconstructions with forcing factors as well as comparing these with paleoclimate model simulations is key to gaining insights into the relative roles of internal versus forced variability. A review of the state of modelling of the climate of the last millennium prior to the CMIP5–PMIP3 (Coupled Model Intercomparison Project Phase 5–Paleoclimate Modelling Intercomparison Project Phase 3) coordinated effort is presented and compared to the available temperature reconstructions. Simulations and reconstructions broadly agree on reproducing the major temperature changes and suggest an overall linear response to external forcing on multidecadal or longer timescales. Internal variability is found to have an important influence at hemispheric and global scales. The spatial distribution of simulated temperature changes during the transition from the Medieval Climate Anomaly to the Little Ice Age disagrees with that found in the reconstructions. Thus, either internal variability is a possible major player in shaping temperature changes through the millennium or the model simulations have problems realistically representing the response pattern to external forcing. A last millennium transient climate response (LMTCR) is defined to provide a quantitative framework for analysing the consistency between simulated and reconstructed climate. Beyond an overall agreement between simulated and reconstructed LMTCR ranges, this analysis is able to single out specific discrepancies between some reconstructions and the ensemble of simulations. The disagreement is found in the cases where the reconstructions show reduced covariability with external forcings or when they present high rates of temperature change.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Changes in Greenland accumulation and the stability in the relationship between accumulation variability and large-scale circulation are assessed by performing time-slice simulations for the present day, the preindustrial era, the early Holocene, and the Last Glacial Maximum (LGM) with a comprehensive climate model. The stability issue is an important prerequisite for reconstructions of Northern Hemisphere atmospheric circulation variability based on accumulation or precipitation proxy records from Greenland ice cores. The analysis reveals that the relationship between accumulation variability and large-scale circulation undergoes a significant seasonal cycle. As the contributions of the individual seasons to the annual signal change, annual mean accumulation variability is not necessarily related to the same atmospheric circulation patterns during the different climate states. Interestingly, within a season, local Greenland accumulation variability is indeed linked to a consistent circulation pattern, which is observed for all studied climate periods, even for the LGM. Hence, it would be possible to deduce a reliable reconstruction of seasonal atmospheric variability (e.g., for North Atlantic winters) if an accumulation or precipitation proxy were available that resolves single seasons. We further show that the simulated impacts of orbital forcing and changes in the ice sheet topography on Greenland accumulation exhibit strong spatial differences, emphasizing that accumulation records from different ice core sites regarding both interannual and long-term (centennial to millennial) variability cannot be expected to look alike since they include a distinct local signature. The only uniform signal to external forcing is the strong decrease in Greenland accumulation during glacial (LGM) conditions and an increase associated with the recent rise in greenhouse gas concentrations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The uncertainty on the calorimeter energy response to jets of particles is derived for the ATLAS experiment at the Large Hadron Collider (LHC). First, the calorimeter response to single isolated charged hadrons is measured and compared to the Monte Carlo simulation using proton-proton collisions at centre-of-mass energies of root s = 900 GeV and 7 TeV collected during 2009 and 2010. Then, using the decay of K-s and Lambda particles, the calorimeter response to specific types of particles (positively and negatively charged pions, protons, and anti-protons) is measured and compared to the Monte Carlo predictions. Finally, the jet energy scale uncertainty is determined by propagating the response uncertainty for single charged and neutral particles to jets. The response uncertainty is 2-5 % for central isolated hadrons and 1-3 % for the final calorimeter jet energy scale.