954 resultados para variable rate application


Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this treatise we consider finite systems of branching particles where the particles move independently of each other according to d-dimensional diffusions. Particles are killed at a position dependent rate, leaving at their death position a random number of descendants according to a position dependent reproduction law. In addition particles immigrate at constant rate (one immigrant per immigration time). A process with above properties is called a branching diffusion withimmigration (BDI). In the first part we present the model in detail and discuss the properties of the BDI under our basic assumptions. In the second part we consider the problem of reconstruction of the trajectory of a BDI from discrete observations. We observe positions of the particles at discrete times; in particular we assume that we have no information about the pedigree of the particles. A natural question arises if we want to apply statistical procedures on the discrete observations: How can we find couples of particle positions which belong to the same particle? We give an easy to implement 'reconstruction scheme' which allows us to redraw or 'reconstruct' parts of the trajectory of the BDI with high accuracy. Moreover asymptotically the whole path can be reconstructed. Further we present simulations which show that our partial reconstruction rule is tractable in practice. In the third part we study how the partial reconstruction rule fits into statistical applications. As an extensive example we present a nonparametric estimator for the diffusion coefficient of a BDI where the particles move according to one-dimensional diffusions. This estimator is based on the Nadaraya-Watson estimator for the diffusion coefficient of one-dimensional diffusions and it uses the partial reconstruction rule developed in the second part above. We are able to prove a rate of convergence of this estimator and finally we present simulations which show that the estimator works well even if we leave our set of assumptions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this thesis, we investigated the evaporation of sessile microdroplets on different solid substrates. Three major aspects were studied: the influence of surface hydrophilicity and heterogeneity on the evaporation dynamics for an insoluble solid substrate, the influence of external process parameters and intrinsic material properties on microstructuring of soluble polymer substrates and the influence of an increased area to volume ratio in a microfluidic capillary, when evaporation is hindered. In the first part, the evaporation dynamics of pure sessile water drops on smooth self-assembled monolayers (SAMs) of thiols or disulfides on gold on mica was studied. With increasing surface hydrophilicity the drop stayed pinned longer. Thus, the total evaporation time of a given initial drop volume was shorter, since the drop surface, through which the evaporation occurs, stays longer large. Usually, for a single drop the volume decreased linearly with t1.5, t being the evaporation time, for a diffusion-controlled evaporation process. However, when we measured the total evaporation time, ttot, for multiple droplets with different initial volumes, V0, we found a scaling of the form V0 = attotb. The more hydrophilic the substrate was, the more showed the scaling exponent a tendency to an increased value up to 1.6. This can be attributed to an increasing evaporation rate through a thin water layer in the vicinity of the drop. Under the assumption of a constant temperature at the substrate surface a cooling of the droplet and thus a decreased evaporation rate could be excluded as a reason for the different scaling exponent by simulations performed by F. Schönfeld at the IMM, Mainz. In contrast, for a hairy surface, made of dialkyldisulfide SAMs with different chain lengths and a 1:1 mixture of hydrophilic and hydrophobic end groups (hydroxy versus methyl group), the scaling exponent was found to be ~ 1.4. It increased to ~ 1.5 with increasing hydrophilicity. A reason for this observation can only be speculated: in the case of longer hydrophobic alkyl chains the formation of an air layer between substrate and surface might be favorable. Thus, the heat transport to the substrate might be reduced, leading to a stronger cooling and thus decreased evaporation rate. In the second part, the microstructuring of polystyrene surfaces by drops of toluene, a good solvent, was investigated. For this a novel deposition technique was developed, with which the drop can be deposited with a syringe. The polymer substrate is lying on a motorized table, which picks up the pendant drop by an upward motion until a liquid bridge is formed. A consecutive downward motion of the table after a variable delay, i.e. the contact time between drop and polymer, leads to the deposition of the droplet, which can evaporate. The resulting microstructure is investigated in dependence of the processes parameters, i.e. the approach and the retraction speed of the substrate and the delay between them, and in dependence of the intrinsic material properties, i.e. the molar mass and the type of the polymer/solvent system. The principal equivalence with the microstructuring by the ink-jet technique was demonstrated. For a high approach and retraction speed of 9 mm/s and no delay between them, a concave microtopology was observed. In agreement with the literature, this can be explained by a flow of solvent and the dissolved polymer to the rim of the pinned droplet, where polymer is accumulated. This effect is analogue to the well-known formation of ring-like stains after the evaporation of coffee drops (coffee-stain effect). With decreasing retraction speed down to 10 µm/s the resulting surface topology changes from concave to convex. This can be explained with the increasing dissolution of polymer into the solvent drop prior to the evaporation. If the polymer concentration is high enough, gelation occurs instead of a flow to the rim and the shape of the convex droplet is received. With increasing delay time from below 0 ms to 1s the depth of the concave microwells decreases from 4.6 µm to 3.2 µm. However, a convex surface topology could not be obtained, since for longer delay times the polymer sticks to the tip of the syringe. Thus, by changing the delay time a fine-tuning of the concave structure is accomplished, while by changing the retraction speed a principal change of the microtopolgy can be achieved. We attribute this to an additional flow inside the liquid bridge, which enhanced polymer dissolution. Even if the pendant drop is evaporating about 30 µm above the polymer surface without any contact (non-contact mode), concave structures were observed. Rim heights as high as 33 µm could be generated for exposure times of 20 min. The concave structure exclusively lay above the flat polymer surface outside the structure even after drying. This shows that toluene is taken up permanently. The increasing rim height, rh, with increasing exposure time to the solvent vapor obeys a diffusion law of rh = rh0  tn, with n in the range of 0.46 ~ 0.65. This hints at a non-Fickian swelling process. A detailed analysis showed that the rim height of the concave structure is modulated, unlike for the drop deposition. This is due to the local stress relaxation, which was initiated by the increasing toluene concentration in the extruded polymer surface. By altering the intrinsic material parameters i.e. the polymer molar mass and the polymer/solvent combination, several types of microstructures could be formed. With increasing molar mass from 20.9 kDa to 1.44 MDa the resulting microstructure changed from convex, to a structure with a dimple in the center, to concave, to finally an irregular structure. This observation can be explained if one assumes that the microstructuring is dominated by two opposing effects, a decreasing solubility with increasing polymer molar mass, but an increasing surface tension gradient leading to instabilities of Marangoni-type. Thus, a polymer with a low molar mass close or below the entanglement limit is subject to a high dissolution rate, which leads to fast gelation compared to the evaporation rate. This way a coffee-rim like effect is eliminated early and a convex structure results. For high molar masses the low dissolution rate and the low polymer diffusion might lead to increased surface tension gradients and a typical local pile-up of polymer is found. For intermediate polymer masses around 200 kDa, the dissolution and evaporation rate are comparable and the typical concave microtopology is found. This interpretation was supported by a quantitative estimation of the diffusion coefficient and the evaporation rate. For a different polymer/solvent system, polyethylmethacrylate (PEMA)/ethylacetate (EA), exclusively concave structures were found. Following the statements above this can be interpreted with a lower dissolution rate. At low molar masses the concentration of PEMA in EA most likely never reaches the gelation point. Thus, a concave instead of a convex structure occurs. At the end of this section, the optically properties of such microstructures for a potential application as microlenses are studied with laser scanning confocal microscopy. In the third part, the droplet was confined into a glass microcapillary to avoid evaporation. Since here, due to an increased area to volume ratio, the surface properties of the liquid and the solid walls became important, the influence of the surface hydrophilicity of the wall on the interfacial tension between two immiscible liquid slugs was investigated. For this a novel method for measuring the interfacial tension between the two liquids within the capillary was developed. This technique was demonstrated by measuring the interfacial tensions between slugs of pure water and standard solvents. For toluene, n-hexane and chloroform 36.2, 50.9 and 34.2 mN/m were measured at 20°C, which is in a good agreement with data from the literature. For a slug of hexane in contact with a slug of pure water containing ethanol in a concentration range between 0 and 70 (v/v %), a difference of up to 6 mN/m was found, when compared to commercial ring tensiometry. This discrepancy is still under debate.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

My PhD project has been focused on the study of the pulsating variable stars in two ultra-faint dwarf spheroidal satellites of the Milky Way, namely, Leo IV and Hercules; and in two fields of the Large Magellanic Cloud (namely, the Gaia South Ecliptic Pole calibration field, and the 30 Doradus region) that were repeatedly observed in the KS band by the VISTA Magellanic Cloud (VMC, PI M.R. Cioni) survey of the Magellanic System.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Speeding the VO2 kinetics results in a reduction of the O2 deficit. Two factors might determine VO2 kinetics: oxygen delivery to muscle (Tschakovsky and Hughson 1999) and a muscle 'metabolic inertia' (Grassi et al. 1996). Therefore, in study 1 we investigated VO2 kinetics and cardiovascular system adaptations during step exercise transitions in different regions of the moderate domain. In study 2 we investigated muscle oxygenation and cardio-pulmonary adaptations during step exercise tests before, after and over a period of training. Study 1 methods: Seven subjects (26 ± 8 yr; 176 ± 5 cm; 69 ± 6 kg) performed 4 types of step transition from rest (0-50W; 0-100W) or elevate baseline (25-75W; 25-125W). GET and VO2max were assessed before testing. O2 uptake and were measured during testing. Study 2 methods: 10 subjects (25 ± 4 yr; 175 ± 9 cm; 71 ± 12 kg) performed a step transition test (0 to 100 W) before, after and during 4 weeks of endurance training (ET). VO2max and GET were assessed before and after of ET (40 minutes, 3 times a week, 60% O2max). VO2 uptake, Q and deoxyheamoglobin were measured during testing. Study 1 results: VO2 τ and the functional gain were slower in the upper regions of the moderate domain. Q increased more abruptly during rest to work condition. Q τ was faster than VO2 τ for each exercise step. Study 2 results: VO2 τ became faster after ET (25%) and particularly after 1 training session (4%). Q kinetics changed after 4 training sessions nevertheless it was always faster than VO2 τ. An attenuation in ∆[HHb] /∆VO2 was detectible. Conclusion: these investigations suggest that muscle fibres recruitment exerts a influence on the VO2 response within the moderate domain either during different forms of step transition or following ET.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The purpose of this doctoral thesis is to prove existence for a mutually catalytic random walk with infinite branching rate on countably many sites. The process is defined as a weak limit of an approximating family of processes. An approximating process is constructed by adding jumps to a deterministic migration on an equidistant time grid. As law of jumps we need to choose the invariant probability measure of the mutually catalytic random walk with a finite branching rate in the recurrent regime. This model was introduced by Dawson and Perkins (1998) and this thesis relies heavily on their work. Due to the properties of this invariant distribution, which is in fact the exit distribution of planar Brownian motion from the first quadrant, it is possible to establish a martingale problem for the weak limit of any convergent sequence of approximating processes. We can prove a duality relation for the solution to the mentioned martingale problem, which goes back to Mytnik (1996) in the case of finite rate branching, and this duality gives rise to weak uniqueness for the solution to the martingale problem. Using standard arguments we can show that this solution is in fact a Feller process and it has the strong Markov property. For the case of only one site we prove that the model we have constructed is the limit of finite rate mutually catalytic branching processes as the branching rate approaches infinity. Therefore, it seems naturalto refer to the above model as an infinite rate branching process. However, a result for convergence on infinitely many sites remains open.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Von aromatischen Aldehyden abgeleitete α-Aminonitrile können ohne die Anwendung von Schutzgruppen in α-Position deprotoniert werden, wenn keine lithiumhaltigen Basen verwendet werden. Ziel der vorliegenden Arbeit war es, die Reaktionen deprotonierter α-Aminonitrile mit Elektrophilen zu untersuchen. Die Addition von α-Aminocarbanionen an Imine führt unter intramolekularer Eliminierung von HCN zu Endiaminen, die sich in einer Eintopfsynthese abhängig von der Aufarbeitung in 1,2-Diamine oder 1,2-Diimine umwandeln lassen. Die nach Oxidation durch Luftsauerstoff erhaltenen Diimine können mit dem Reduktionsmittel BH3·THF diastereoselektiv reduziert werden. Es hat sich hier gezeigt, dass durch Zugabe einer katalytischen Menge an NaBH4 hauptsächlich die syn-Diamine erhalten werden, der Zusatz von Phthalsäure wiederum liefert bevorzugt die anti-Produkte. In beiden Fällen wird das Produkt in quantitativer Ausbeute erhalten. So konnte also eine effektive diastereoselektive Reduktionsmethode entwickelt werden, die eine freie Wahl der syn- oder anti-Konfiguration ermöglicht. Um enantiomerenreine 1,2-Diamine zu erhalten, wurden verschiedene Methoden getestet. Sowohl auxiliargesteuerte Synthesen mit einem N-Glycosyl-Aminonitril oder mit chiralen Sulfinyliminen als auch die Reduktion durch chirale Borverbindungen (CBS-Katalysatoren, Triacyloxyborhydrid oder Diisopinocamphenylboran), Transferhydrierungen mit chiralen Difluortitanocen-, Noyori- oder Organophosphat-Katalysatoren sowie enantioselektive Hydrierungen mit chiralen Übergangsmetall-katalysatoren waren jedoch nicht erfolgreich. Die Umsetzung der 1,2-Diimine mit Chlormethylethern oder -estern liefert die entsprechenden unsymmetrischen Imidazoliumsalze. Diese konnten zu N-heterocyclischen Carbenen deprotoniert und erfolgreich als Liganden in Suzuki- und Heck-Reaktionen eingesetzt werden. Durch die 1,2-Addition α-deprotonierter Streckerprodukte und anschließende Reduktion im Eintopfverfahren konnten 1,2-Aminoalkohole in mäßigen bis guten Ausbeuten dargestellt werden. Die Umsetzung von α-Aminocarbanionen mit N-Acyliminen erlaubt zudem die Synthese tetrasubstituierter Imidazole und trisubstituierter Oxazole in drei beziehungsweise vier Stufen: Die zunächst gebildeten α-Amino-α-acylaminopropionitrile können isoliert und in Gegenwart von Base einer Retro-Strecker-Reaktion unterworfen werden. Abhängig vom Substitutionsmuster schließt sich in manchen Fällen nach der Eliminierung von HCN direkt die Cyclisierung zum Imidazol an. Nicht cyclisierte Intermediate lassen sich durch Dehydratisierung mit PCl5 zu Imidazolen umsetzen, aber auch unter sauren Bedingungen zu α-Acylaminoketonen hydrolysieren, welche wiederum durch Einwirkung von PCl5 in Oxazole überführt werden können. Auf diese Weise wurden Imidazole und Oxazole in moderaten bis hohen Gesamtausbeuten hergestellt.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Stickstoffmonoxid (NO) ist als reaktives Spurengas eine wichtige Komponente atmosphärenchemischer Prozesse und hat somit einen bedeutenden Einfluss auf die Zusammensetzung der Atmosphäre. Eine Hauptquelle des Spurengases stellen bodenmikrobiologische Prozesse dar, deren regionaler und globaler Anteil weiterhin mit größeren Unsicherheiten geschätzt wird. Ursache für die schwere Abschätzbarkeit der NO-Freisetzung aus Böden ist die hohe räumliche Variabilität der steuernden Faktoren. Als einer der wichtigsten Faktoren, die die Freisetzung von NO aus Böden regeln, gilt der Bodenwassergehalt. Ziel der vorliegenden Arbeit ist es, den Zusammenhang zwischen NO-Freisetzung, Bodenwassergehalt, den Bodeneigenschaften und den Standortbedingungen zu untersuchen und diesen möglichst zu quantifizieren. Dazu sind Bodenproben unterschiedlicher Landnutzungen in einem kleineren Wassereinzugsgebiet im Rheingau im Labor, unter kontrollierten Bedingungen, untersucht. Der charakteristische Zusammenhang zwischen Bodenfeuchte und NO-Freisetzung, die sogenannte Bodenfeuchtekurve, kann demnach weitestgehend auf die gemessenen Bodenmerkmale der untersuchten Proben zurückgeführt werden. Anhand der Bodenmerkmale kann die Bodenfeuchtekurve zufriedenstellend vorhergesagt werden. Dabei zeigt vor allem der Humusgehalt der Böden einen dominierenden Einfluss. Er ist die Variable, die die Unterschiede der Böden beim Zusammenhang zwischen Bodenfeuchte und NO-Freisetzung am besten und hinreichend erklären kann. Zur Konstruktion der Bodenfeuchtekurve müssen die optimale Bodenfeuchte und die dabei herrschende Freisetzung, sowie die obere Bodenfeuchte, bei der keine NO-Freisetzung mehr stattfindet, bekannt sein. Diese charakteristischen Punkte lassen sich durch lineare Regressionsmodelle gut aus den Bodeneigenschaften ableiten. Auf räumlicher Ebene werden die Bodeneigenschaften durch die standörtlichen Bedingungen geprägt, die wiederum Ausdruck der landschaftlichen Ausstattung sind. In der Kulturlandschaft kann der Mensch aufgrund seiner Landnutzungsansprüche als der dominierende Faktor angesehen werden. Die Landnutzung orientiert sich an den landschaftlichen Bedingungen und bestimmt in hohem Maße wichtige Bodeneigenschaften, die zu den erklärenden Merkmalen bei der Beziehung zwischen Bodenwassergehalt und NO-Freisetzung gehören. Die in erster Linie wirtschaftlich orientierten Kartenwerke Bodenschätzung, Weinbergsbodenkartierung und forstliche Standortkartierung sind dementsprechend geeignete Grundlagen, um eine Regionalisierung der Landschaft in - bezüglich der NO-Freisetzung - weitgehend homogene Flächen durchführen zu können. Eine hierauf beruhende Regionalisierung ist dazu geeignet, die räumliche Variabilität der NO-Freisetzung in räumlich sinnvoller Auflösung besser abschätzen zu können.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Within this thesis a new double laser pulse pumping scheme for plasma-based, transient collisionally excited soft x-ray lasers (SXRL) was developed, characterized and utilized for applications. SXRL operations from ~50 up to ~200 electron volt were demonstrated applying this concept. As a central technical tool, a special Mach-Zehnder interferometer in the chirped pulse amplification (CPA) laser front-end was developed for the generation of fully controllable double-pulses to optimally pump SXRLs.rnThis Mach-Zehnder device is fully controllable and enables the creation of two CPA pulses of different pulse duration and variable energy balance with an adjustable time delay. Besides the SXRL pumping, the double-pulse configuration was applied to determine the B-integral in the CPA laser system by amplifying short pulse replica in the system, followed by an analysis in the time domain. The measurement of B-integral values in the 0.1 to 1.5 radian range, only limited by the reachable laser parameters, proved to be a promising tool to characterize nonlinear effects in the CPA laser systems.rnContributing to the issue of SXRL pumping, the double-pulse was configured to optimally produce the gain medium of the SXRL amplification. The focusing geometry of the two collinear pulses under the same grazing incidence angle on the target, significantly improved the generation of the active plasma medium. On one hand the effect was induced by the intrinsically guaranteed exact overlap of the two pulses on the target, and on the other hand by the grazing incidence pre-pulse plasma generation, which allows for a SXRL operation at higher electron densities, enabling higher gain in longer wavelength SXRLs and higher efficiency at shorter wavelength SXRLs. The observation of gain enhancement was confirmed by plasma hydrodynamic simulations.rnThe first introduction of double short-pulse single-beam grazing incidence pumping for SXRL pumping below 20 nanometer at the laser facility PHELIX in Darmstadt (Germany), resulted in a reliable operation of a nickel-like palladium SXRL at 14.7 nanometer with a pump energy threshold strongly reduced to less than 500 millijoule. With the adaptation of the concept, namely double-pulse single-beam grazing incidence pumping (DGRIP) and the transfer of this technology to the laser facility LASERIX in Palaiseau (France), improved efficiency and stability of table-top high-repetition soft x-ray lasers in the wavelength region below 20 nanometer was demonstrated. With a total pump laser energy below 1 joule the target, 2 mircojoule of nickel-like molybdenum soft x-ray laser emission at 18.9 nanometer was obtained at 10 hertz repetition rate, proving the attractiveness for high average power operation. An easy and rapid alignment procedure fulfilled the requirements for a sophisticated installation, and the highly stable output satisfied the need for a reliable strong SXRL source. The qualities of the DGRIP scheme were confirmed in an irradiation operation on user samples with over 50.000 shots corresponding to a deposited energy of ~ 50 millijoule.rnThe generation of double-pulses with high energies up to ~120 joule enabled the transfer to shorter wavelength SXRL operation at the laser facility PHELIX. The application of DGRIP proved to be a simple and efficient method for the generation of soft x-ray lasers below 10 nanometer. Nickel-like samarium soft x-ray lasing at 7.3 nanometer was achieved at a low total pump energy threshold of 36 joule, which confirmed the suitability of the applied pumping scheme. A reliable and stable SXRL operation was demonstrated, due to the single-beam pumping geometry despite the large optical apertures. The soft x-ray lasing of nickel-like samarium was an important milestone for the feasibility of applying the pumping scheme also for higher pumping pulse energies, which are necessary to obtain soft x-ray laser wavelengths in the water window. The reduction of the total pump energy below 40 joule for 7.3 nanometer short wavelength lasing now fulfilled the requirement for the installation at the high-repetition rate operation laser facility LASERIX.rn

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of the dissertation was to test the feasibility of a new psychotherapeutic protocol for treating children and adolescents with mood and anxiety disorders: Child-Well-Being Therapy (CWBT). It originates from adult Well-Being Therapy protocol (WBT) and represents a conceptual innovation for treating affective disorders. WBT is based on the multidimensional model of well-being postulated by Ryff (eudaimonic perspective), in sequential combination with cognitive-behavioral therapy (CBT). Results showed that eudaimonic well-being was impaired in children with affective disorders in comparison with matched healthy students. A first open investigation aimed at exploring the feasibility of a 8-session CWBT protocol in a group of children with emotional and behavioural disorders has been implemented. Data showed how CWBT resulted associated to symptoms reduction, together with the decrease of externalizing problems, maintained at 1-year follow-up. CWBT triggered also an improvement in psychological well-being as well as an increasing flourishing trajectory over time. Subsequently, a modified and extended version of CWBT (12-sessions) has been developed and then tested in a controlled study with 34 patients (8 to 16 years) affected by mood and anxiety disorders. They were consecutively randomized into 3 different groups: CWBT, CBT, 6-month waiting list (WL). Both treatments resulted effective in decreasing distress and in improving well-being. Moreover, CWBT was associated with higher improvement in anxiety and showed a greater recovery rate (83%) than CBT (54%). Both groups maintained beneficial effects and CWBT group displayed a lower level of distress as well as a higher positive trend in well-being scores over time. Findings need to be interpret with caution, because of study limitations, however important clinical implications emerged. Further investigations should determine whether the sequential integration of well-being and symptom-oriented strategies could play an important role in children and adolescents’ psychotherapeutic options, fostering a successful adaptation to adversities during the growth process.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In recent years the advances in genomics allowed to understand the importance of Transposable Elements (TE) in the evolution of eukaryotic genomes. In this thesis I face two aspects of the TE impact on the in the animal kingdom. The first part is a comparison of the dynamics of the TE dynamics in three species of stick-insects of the Genus Bacillus. I produced three random genomic libraries of 200 Kbps for the three parental species of the taxon: a gonochoric population of Bacillus rossius (facultative parthenogenetic), Bacillus grandii (gonochoric) and Bacillus atticus (obligate parthenogenetic). The unisexual taxon Bacillus atticus does not shows dramatic differences in TE total content and activity with respect to Bacillus grandii and Bacillus rossius. This datum does not confirm the trend observed in other animal models in which unisexual taxa tend to repress the activity of TE to escape the extinction by accumulation of harmful mutations. In the second part I tried to add a contribute to the debate initiated in recent years about the possibility that a high TE content is linked to a high rate of speciation. I designed an evolutionary framework to establish the different rate of speciation among two or more taxa, then I compared TE dynamics considering the different rates of speciation. The species dataset comprises: 29 mammals, four birds, two fish and two insects. On the whole the majority of comparisons confirms the expected trend. In particular the amount of species analyzed in Mammalia allowed me to get a statistical support (p<0,05) of the fact that the TE activity of recently mobilized elements is positively related with the rate of speciation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

I present a new experimental method called Total Internal Reflection Fluorescence Cross-Correlation Spectroscopy (TIR-FCCS). It is a method that can probe hydrodynamic flows near solid surfaces, on length scales of tens of nanometres. Fluorescent tracers flowing with the liquid are excited by evanescent light, produced by epi-illumination through the periphery of a high NA oil-immersion objective. Due to the fast decay of the evanescent wave, fluorescence only occurs for tracers in the ~100 nm proximity of the surface, thus resulting in very high normal resolution. The time-resolved fluorescence intensity signals from two laterally shifted (in flow direction) observation volumes, created by two confocal pinholes are independently measured and recorded. The cross-correlation of these signals provides important information for the tracers’ motion and thus their flow velocity. Due to the high sensitivity of the method, fluorescent species with different size, down to single dye molecules can be used as tracers. The aim of my work was to build an experimental setup for TIR-FCCS and use it to experimentally measure the shear rate and slip length of water flowing on hydrophilic and hydrophobic surfaces. However, in order to extract these parameters from the measured correlation curves a quantitative data analysis is needed. This is not straightforward task due to the complexity of the problem, which makes the derivation of analytical expressions for the correlation functions needed to fit the experimental data, impossible. Therefore in order to process and interpret the experimental results I also describe a new numerical method of data analysis of the acquired auto- and cross-correlation curves – Brownian Dynamics techniques are used to produce simulated auto- and cross-correlation functions and to fit the corresponding experimental data. I show how to combine detailed and fairly realistic theoretical modelling of the phenomena with accurate measurements of the correlation functions, in order to establish a fully quantitative method to retrieve the flow properties from the experiments. An importance-sampling Monte Carlo procedure is employed in order to fit the experiments. This provides the optimum parameter values together with their statistical error bars. The approach is well suited for both modern desktop PC machines and massively parallel computers. The latter allows making the data analysis within short computing times. I applied this method to study flow of aqueous electrolyte solution near smooth hydrophilic and hydrophobic surfaces. Generally on hydrophilic surface slip is not expected, while on hydrophobic surface some slippage may exists. Our results show that on both hydrophilic and moderately hydrophobic (contact angle ~85°) surfaces the slip length is ~10-15nm or lower, and within the limitations of the experiments and the model, indistinguishable from zero.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Due to its practical importance and inherent complexity, the optimisation of distribution networks for supplying drinking water has been the subject of extensive study for the past 30 years. The optimization is governed by sizing the pipes in the water distribution network (WDN) and / or optimises specific parts of the network such as pumps, tanks etc. or try to analyse and optimise the reliability of a WDN. In this thesis, the author has analysed two different WDNs (Anytown City and Cabrera city networks), trying to solve and optimise a multi-objective optimisation problem (MOOP). The main two objectives in both cases were the minimisation of Energy Cost (€) or Energy consumption (kWh), along with the total Number of pump switches (TNps) during a day. For this purpose, a decision support system generator for Multi-objective optimisation used. Its name is GANetXL and has been developed by the Center of Water System in the University of Exeter. GANetXL, works by calling the EPANET hydraulic solver, each time a hydraulic analysis has been fulfilled. The main algorithm used, was a second-generation algorithm for multi-objective optimisation called NSGA_II that gave us the Pareto fronts of each configuration. The first experiment that has been carried out was the network of Anytown city. It is a big network with a pump station of four fixed speed parallel pumps that are boosting the water dynamics. The main intervention was to change these pumps to new Variable speed driven pumps (VSDPs), by installing inverters capable to diverse their velocity during the day. Hence, it’s been achieved great Energy and cost savings along with minimisation in the number of pump switches. The results of the research are thoroughly illustrated in chapter 7, with comments and a variety of graphs and different configurations. The second experiment was about the network of Cabrera city. The smaller WDN had a unique FS pump in the system. The problem was the same as far as the optimisation process was concerned, thus, the minimisation of the energy consumption and in parallel the minimisation of TNps. The same optimisation tool has been used (GANetXL).The main scope was to carry out several and different experiments regarding a vast variety of configurations, using different pump (but this time keeping the FS mode), different tank levels, different pipe diameters and different emitters coefficient. All these different modes came up with a large number of results that were compared in the chapter 8. Concluding, it should be said that the optimisation of WDNs is a very interested field that has a vast space of options to deal with. This includes a large number of algorithms to choose from, different techniques and configurations to be made and different support system generators. The researcher has to be ready to “roam” between these choices, till a satisfactory result will convince him/her that has reached a good optimisation point.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The cone penetration test (CPT), together with its recent variation (CPTU), has become the most widely used in-situ testing technique for soil profiling and geotechnical characterization. The knowledge gained over the last decades on the interpretation procedures in sands and clays is certainly wide, whilst very few contributions can be found as regards the analysis of CPT(u) data in intermediate soils. Indeed, it is widely accepted that at the standard rate of penetration (v = 20 mm/s), drained penetration occurs in sands while undrained penetration occurs in clays. However, a problem arise when the available interpretation approaches are applied to cone measurements in silts, sandy silts, silty or clayey sands, since such intermediate geomaterials are often characterized by permeability values within the range in which partial drainage is very likely to occur. Hence, the application of the available and well-established interpretation procedures, developed for ‘standard’ clays and sands, may result in invalid estimates of soil parameters. This study aims at providing a better understanding on the interpretation of CPTU data in natural sand and silt mixtures, by taking into account two main aspects, as specified below: 1)Investigating the effect of penetration rate on piezocone measurements, with the aim of identifying drainage conditions when cone penetration is performed at a standard rate. This part of the thesis has been carried out with reference to a specific CPTU database recently collected in a liquefaction-prone area (Emilia-Romagna Region, Italy). 2)Providing a better insight into the interpretation of piezocone tests in the widely studied silty sediments of the Venetian lagoon (Italy). Research has focused on the calibration and verification of some site-specific correlations, with special reference to the estimate of compressibility parameters for the assessment of long-term settlements of the Venetian coastal defences.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Zahnverlust zu Lebzeiten („antemortem tooth loss“, AMTL) kann als Folge von Zahnerkrankungen, Traumata, Zahnextraktionen oder extremer kontinuierlicher Eruption sowie als Begleiterscheinung fortgeschrittener Stadien von Skorbut oder Lepra auftreten. Nach dem Zahnverlust setzt die Wundheilung als Sekundärheilung ein, während der sich die Alveole mit Blut füllt und sich ein Koagulum bildet. Anschließend erfolgt dessen Umwandlung in Knochengewebe und schließlich verstreicht die Alveole derart, dass sie makroskopisch nicht mehr erkannt werden kann. Der Zeitrahmen der knöchernen Konsolidierung des Kieferkammes ist im Detail wenig erforscht. Aufgrund des gehäuften Auftretens von AMTL in menschlichen Populationen, ist die Erarbeitung eines Zeitfensters, mit dessen Hilfe durch makroskopische Beobachtung des Knochens die Zeitspanne seit dem Zahnverlust („time since tooth loss“, TSL) ermittelt werden kann, insbesondere im archäologischen Kontext äußerst wertvoll. Solch ein Zeitschema mit Angaben über die Variabilität der zeitlichen Abläufe bei den Heilungsvorgängen kann nicht nur in der Osteologie, sondern auch in der Forensik, der allgemeinen Zahnheilkunde und der Implantologie nutzbringend angewandt werden. rnrnNach dem Verlust eines Zahnes wird das Zahnfach in der Regel durch ein Koagulum aufgefüllt. Das sich bildende Gewebe wird rasch in noch unreifen Knochen umgewandelt, welcher den Kieferknochen und auch die angrenzenden Zähne stabilisiert. Nach seiner Ausreifung passt sich das Gewebe schließlich dem umgebenden Knochen an. Das Erscheinungsbild des Zahnfaches während dieses Vorgangs durchläuft verschiedene Stadien, welche in der vorliegenden Studie anhand von klinischen Röntgenaufnahmen rezenter Patienten sowie durch Untersuchungen an archäologischen Skelettserien identifiziert wurden. Die Heilungsvorgänge im Zahnfach können in eine prä-ossale Phase (innerhalb einer Woche nach Zahnverlust), eine Verknöcherungsphase (etwa 14 Wochen nach Zahnverlust) und eine ossifizierte bzw. komplett verheilte Phase (mindestens 29 Wochen nach Zahnverlust) eingeteilt werden. Etliche Faktoren – wie etwa die Resorption des Interdentalseptums, der Zustand des Alveolarknochens oder das Individualgeschlecht – können den normalen Heilungsprozess signifikant beschleunigen oder hemmen und so Unterschiede von bis zu 19 Wochen verursachen. Weitere Variablen wirkten sich nicht signifikant auf den zeitlichen Rahmen des Heilungsprozesse aus. Relevante Abhängigkeiten zwischen verschiedenen Variabeln wurden ungeachtet der Alveolenauffüllung ebenfalls getestet. Gruppen von unabhängigen Variabeln wurden im Hinblick auf Auffüllungsgrad und TSL in multivariablen Modellen untersucht. Mit Hilfe dieser Ergebnisse ist eine grobe Einschätzung der Zeitspanne nach einem Zahnverlust in Wochen möglich, wobei die Einbeziehung weiterer Parameter eine höhere Präzision ermöglicht. rnrnObwohl verschiedene dentale Pathologien in dieser Studie berücksichtigt wurden, sollten zukünftige Untersuchungen genauer auf deren potenzielle Einflussnahme auf den alveolaren Heilungsprozess eingehen. Der kausale Zusammenhang einiger Variablen (wie z. B. Anwesenheit von Nachbarzähnen oder zahnmedizinische Behandlungen), welche die Geschwindigkeit der Heilungsrate beeinflussen, wäre von Bedeutung für zukünftige Untersuchungen des oralen Knochengewebes. Klinische Vergleichsstudien an forensischen Serien mit bekannter TSL oder an einer sich am Anfang des Heilungsprozesses befindlichen klinischen Serie könnten eine Bekräftigung dieser Ergebnisse liefern.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In der vorliegenden Arbeit wurden die durch Training induzierten motorischen Gedächtnisleistungen der Taufliege Drosophila melanogaster beim Überklettern von acht symmetrisch verteilten Lücken auf einem rotierenden Ring untersucht. Durch den auf sie einwirkenden optischen Fluss der vorbeiziehenden äußeren Umgebung wurden die Fliegen angeregt, diesem optomotorischen Reiz entgegenzuwirken und die Lücken laufend zu überqueren. Durch Training verbessert und langfristig gelernt wird die kompensatorische Lückenüberquerung X+ gegen die Rotation. In der aus diesem Training erhaltenen Lernkurve war eine überdurchschnittlich hohe Leistungsverbesserung nach einem einzigen Trainingslauf mit einem zeitlichen Bestand von ca. 40 Minuten abzulesen, um danach vom motorischen Gedächtnisspeicher trainierter Fliegen nicht mehr abgerufen werden zu können. Nach einer Ruhephase von einem bis mehreren Tagen wurden die Fliegen auf mögliche Langzeitlernleistungen untersucht und diese für verschiedene Intervalle nachgewiesen. Sowohl die Leistungsverbesserung während des Trainings, als auch der Lerneffekt nach 24h bleiben in mutanten rutabaga2080 sowie rut1 Fliegen aus. Betroffen ist das Gen der Adenylylzyklase I, ein Schlüsselprotein der cAMP-Signalkaskade, die u.a. im olfaktorischen und visuellen Lernen gebraucht wird. Damit ergab sich die Möglichkeit die motorischen Gedächtnisformen durch partielle Rettung zu kartieren. Die motorische Gedächtniskonsolidierung ist schlafabhängig. Wie sich herausstellte, benötigen WTB Fliegen nur eine Dunkelphase von 10h zwischen einem ersten Trainingslauf und einem Testlauf um signifikante Leistungssteigerungen zu erzielen. In weiterführenden Versuchen wurden die Fliegen nachts sowie tagsüber mit einer LED-Lampe oder in einer Dunkelkammer, mit einem Kreisschüttler oder einer Laborwippe depriviert, mit dem Ergebnis, dass nur jene Fliegen ihre Leistung signifikant gegenüber einem ersten Trainingslauf verbessern konnten, welche entweder ausschließlich der Dunkelheit ausgesetzt waren oder welchen die Möglichkeit gegeben wurde, ein Gedächtnis zunächst in einer natürlichen Schlafphase zu konsolidieren (21Uhr bis 7Uhr MEZ). In weiteren Experimenten wurden die experimentellen Bedingungen entweder während des Trainings oder des Tests auf eine Fliege und damit verbunden auf eine erst durch das Training mögliche motorische Gedächtniskonsolidierung einwirken zu können, untersucht. Dazu wurden die Experimentparameter Lückenweite, Rotationsrichtung des Lückenringes, Geschwindigkeit des Lückenringes sowie die Verteilung der acht Lücken auf dem Ring (symmetrisch, asymmetrisch) im Training oder beim Gedächtnisabruf im Testlauf verändert. Aus den Ergebnissen kann geschlussfolgert werden, dass die Lückenweite langzeitkonsolidiert wird, die Rotationsrichtung kurzzeitig abgespeichert wird und die Drehgeschwindigkeit motivierend auf die Fliegen wirkt. Die symmetrische Verteilung der Lücken auf dem Ring dient der Langzeitkonsolidierung und ist als Trainingseingang von hoher Wichtigkeit. Mit Hilfe verschiedener Paradigmen konnten die Leistungsverbesserungen der Fliegen bei Abruf eines Kurz- bzw. Langzeitgedächtnisses hochauflösend betrachtet werden (Transfer). Die Konzentration, mit der eine WTB Fliege eine motorische Aufgabe - die Überquerung von Lücken entgegengesetzt der Rotationsrichtung - durchführt, konnte mit Hilfe von Distraktoreizen bestimmt werden. Wie sich herausstellte, haben Distraktoren einen Einfluss auf die Erfolgsquote einer Überquerung, d.h. mit zunehmender Distraktionsstärke nahm die Wahrscheinlichkeit einer Lückenüberquerung ab. Die Ablenkungsreize wirkten sich weiterhin auf die Vermessung einer Lücke aus, in dem entweder "peering"-artigen Bewegungen im Training durchgeführt wurden oder je nach Reizstärke ausschließlich nur jene Lücken vermessen wurden, welche auch überquert werden sollten.