958 resultados para MASSIVE NUCLEI


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Thrombin is a multifunctional protease, which has a central role in the development and progression of coronary atherosclerotic lesions and it is a possible mediator of myocardial ischemia-reperfusion injury. Its generation and procoagulant activity are greatly upregulated during cardiopulmonary bypass (CPB). On the other hand, activated protein C, a physiologic anticoagulant that is activated by thrombomodulin-bound thrombin, has been beneficial in various models of ischemia-reperfusion. Therefore, our aim in this study was to test whether thrombin generation or protein C activation during coronary artery bypass grafting (CABG) associate with postoperative myocardial damage or hemodynamic changes. To further investigate the regulation of thrombin during CABG, we tested whether preoperative thrombophilic factors associate with increased CPB-related generation of thrombin or its procoagulant activity. We also measured the anticoagulant effects of heparin during CPB with a novel coagulation test, prothrombinase-induced clotting time (PiCT), and compared the performance of this test with the present standard of laboratory-based anticoagulation monitoring. One hundred patients undergoing elective on-pump CABG were studied prospectively. A progressive increase in markers of thrombin generation (F1+2), fibrinolysis (D-dimer), and fibrin formation (soluble fibrin monomer complexes) was observed during CPB, which was further distinctly propagated by reperfusion after myocardial ischemia, and continued to peak after the neutralization of heparin with protamine. Thrombin generation during reperfusion after CABG associated with postoperative myocardial damage and increased pulmonary vascular resistance. Activated protein C levels increased only slightly during CPB before the release of the aortic clamp, but reperfusion and more significantly heparin neutralization caused a massive increase in activated protein C levels. Protein C activation was clearly delayed in relation to both thrombin generation and fibrin formation. Even though activated protein C associated dynamically with postoperative hemodynamic performance, it did not associate with postoperative myocardial damage. Preoperative thrombophilic variables did not associate with perioperative thrombin generation or its procoagulant activity. Therefore, our results do not favor routine thrombophilia screening before CABG. There was poor agreement between PiCT and other measurements of heparin effects in the setting of CPB. However, lower heparin levels during CPB associated with inferior thrombin control and high heparin levels during CPB associated with fewer perioperative transfusions of blood products. Overall, our results suggest that hypercoagulation after CABG, especially during reperfusion, might be clinically important.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Technology is increasingly infiltrating all aspects of our lives and the rapid uptake of devices that live near, on or in our bodies are facilitating radical new ways of working, relating and socialising. This distribution of technology into the very fabric of our everyday life creates new possibilities, but also raises questions regarding our future relationship with data and the quantified self. By embedding technology into the fabric of our clothes and accessories, it becomes ‘wearable’. Such ‘wearables’ enable the acquisition of and the connection to vast amounts of data about people and environments in order to provide life-augmenting levels of interactivity. Wearable sensors for example, offer the potential for significant benefits in the future management of our wellbeing. Fitness trackers such as ‘Fitbit’ and ‘Garmen’ provide wearers with the ability to monitor their personal fitness indicators while other wearables provide healthcare professionals with information that improves diagnosis. While the rapid uptake of wearables may offer unique and innovative opportunities, there are also concerns surrounding the high levels of data sharing that come as a consequence of these technologies. As more ‘smart’ devices connect to the Internet, and as technology becomes increasingly available (e.g. via Wi-Fi, Bluetooth), more products, artefacts and things are becoming interconnected. This digital connection of devices is called The ‘Internet of Things’ (IoT). IoT is spreading rapidly, with many traditionally non-online devices becoming increasingly connected; products such as mobile phones, fridges, pedometers, coffee machines, video cameras, cars and clothing. The IoT is growing at a rapid rate with estimates indicating that by 2020 there will be over 25 billion connected things globally. As the number of devices connected to the Internet increases, so too does the amount of data collected and type of information that is stored and potentially shared. The ability to collect massive amounts of data - known as ‘big data’ - can be used to better understand and predict behaviours across all areas of research from societal and economic to environmental and biological. With this kind of information at our disposal, we have a more powerful lens with which to perceive the world, and the resulting insights can be used to design more appropriate products, services and systems. It can however, also be used as a method of surveillance, suppression and coercion by governments or large organisations. This is becoming particularly apparent in advertising that targets audiences based on the individual preferences revealed by the data collected from social media and online devices such as GPS systems or pedometers. This type of technology also provides fertile ground for public debates around future fashion, identity and broader social issues such as culture, politics and the environment. The potential implications of these type of technological interactions via wearables, through and with the IoT, have never been more real or more accessible. But, as highlighted, this interconnectedness also brings with it complex technical, ethical and moral challenges. Data security and the protection of privacy and personal information will become ever more present in current and future ethical and moral debates of the 21st century. This type of technology is also a stepping-stone to a future that includes implantable technology, biotechnologies, interspecies communication and augmented humans (cyborgs). Technologies that live symbiotically and perpetually in our bodies, the built environment and the natural environment are no longer the stuff of science fiction; it is in fact a reality. So, where next?... The works exhibited in Wear Next_ provide a snapshot into the broad spectrum of wearables in design and in development internationally. This exhibition has been curated to serve as a platform for enhanced broader debate around future technology, our mediated future-selves and the evolution of human interactions. As you explore the exhibition, may we ask that you pause and think to yourself, what might we... Wear Next_? WEARNEXT ONLINE LISTINGS AND MEDIA COVERAGE: http://indulgemagazine.net/wear-next/ http://www.weekendnotes.com/wear-next-exhibition-gallery-artisan/ http://concreteplayground.com/brisbane/event/wear-next_/ http://www.nationalcraftinitiative.com.au/news_and_events/event/48/wear-next http://bneart.com/whats-on/wear-next_/ http://creativelysould.tumblr.com/post/124899079611/creative-weekend-art-edition http://www.abc.net.au/radionational/programs/breakfast/smartly-dressed-the-future-of-wearable-technology/6744374 http://couriermail.newspaperdirect.com/epaper/viewer.aspx RADIO COVERAGE http://www.abc.net.au/radionational/programs/breakfast/wear-next-exhibition-whats-next-for-wearable-technology/6745986 TELEVISION COVERAGE http://www.abc.net.au/radionational/programs/breakfast/wear-next-exhibition-whats-next-for-wearable-technology/6745986 https://au.news.yahoo.com/video/watch/29439742/how-you-could-soon-be-wearing-smart-clothes/#page1

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The molecular mechanism of helix nucleation in peptides and proteins is not yet understood and the question of whether sharp turns in the polypeptide backbone serve as nuclei for protein folding has evoked controversy1,2. A recent study of the conformation of a tetrapeptide containing the stereochemically constrained residue alpha-aminoisobutyric acid, both in solution and the solid state, yielded a structure consisting of two consecutive beta-turns, leading to an incipient 310 helical conformation3,4. This led us to speculate that specific tri- and tetra-peptide sequences may indeed provide a helical twist to the amino-terminal segment of helical regions in proteins and provide a nucleation site for further propagation. The transformation from a 310 helical structure to an alpha-helix should be facile and requires only small changes in the phi and psi conformational angles and a rearrangement of the hydrogen bonding pattern5. If such a mechanism is involved then it should be possible to isolate an incipient 310 helical conformation in a tripeptide amide or tetrapeptide sequence, based purely on the driving force derived from short-range interactions. We have synthesised and studied the model peptide pivaloyl-Pro-Pro-Ala-NHMe (compound I) and provide here spectroscopic evidence for a 310 helical conformation in compound I.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Gastrointestinal infections with Salmonella enterica serovars have different clinical outcomes that range from localized inflammation to a life-threatening systemic disease in the case of typhoid fever. Using a mouse model of systemic salmonellosis, we investigated the contribution of neutrophils to the innate immune defense against Salmonella after oral infection. Neutrophil infiltration was dependent on the bacterial burden in various infected organs (Peyer's patches, mesenteric lymph nodes, spleen, and liver). However, the massive infiltration of neutrophils did not allow clearance of an infection with wild-type Salmonella, presumably due to protection of intracellular Salmonella against neutrophil activities. A Salmonella mutant strain deficient in Salmonella pathogenicity island 2 (SPI2) was able to infect systemic sites, but its replication was highly restricted and it did not cause detectable attraction of neutrophils. Neutrophil depletion by antibody treatment of mice did not restore the virulence of SPI2 or auxotrophic mutant strains, supporting the hypothesis that attenuation of the strains is not due to greater susceptibility to neutrophil killing. Our observations reveal that neutrophils have completely different roles during systemic salmonellosis and localized gastrointestinal infections. In the latter conditions, rapid neutrophil attraction efficiently prevents the spread of the pathogen, whereas the neutrophil influx is delayed during systemic infections and cannot protect against lethal bacteremia.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We investigate viscous two-temperature accretion disc flows around rotating black holes. We describe the global solution of accretion flows with a sub-Keplerian angular momentum profile, by solving the underlying conservation equations including explicit cooling processes self-consistently. Bremsstrahlung, synchrotron and inverse Comptonization of soft photons are considered as possible cooling mechanisms. We focus on the set of solutions for sub-Eddington, Eddington and super-Eddington mass accretion rates around Schwarzschild and Kerr black holes with a Kerr parameter of 0.998. It is found that the flow, during its infall from the Keplerian to sub-Kepleria transition region to the black hole event horizon, passes through various phases of advection: the general advective paradigm to the radiatively inefficient phase, and vice versa. Hence, the flow governs a much lower electron temperature similar to 10(8)-10(9.5) K, in the range of accretion rate in Eddington units 0.01 less than or similar to (M) over dot less than or similar to 100, compared to the hot protons of temperature similar to 10(10.2)-10(11.8) K. Therefore, the solution may potentially explain the hard X-rays and gamma-rays emitted from active galactic nuclei (AGNs) and X-ray binaries. We then compare the solutions for two different regimes of viscosity. We conclude that a weakly viscous flow is expected to be cooling dominated, particularly at the inner region of the disc, compared to its highly viscous counterpart, which is radiatively inefficient. With all the solutions in hand, we finally reproduce the observed luminosities of the underfed AGNs and quasars (e. g. Sgr A*) to ultraluminous X-ray sources (e. g. SS433), at different combinations of input parameters, such as the mass accretion rate and the ratio of specific heats. The set of solutions also predicts appropriately the luminosity observed in highly luminous AGNs and ultraluminous quasars (e. g. PKS 0743-67).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Television is a massive industry in China, yet fewer people are watching television screens. This ground-breaking study explores how television content is changing, how the Chinese government is responding to the challenges presented by digital media, and how businesses are brokering alliances in both traditional and new media sectors. Table of Contents Acknowledgments p. vi Introduction p. 1 1 Television in Transition p. 8 2 Nation Building p. 34 3 Soft Power p. 56 4 Formats p. 85 5 Channels and Content p. 111 6 Convergence p. 141 7 Rethinking Chinese Television Research p. 164 Bibliography p. 173 Index p. 184

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We present a new method for establishing correlation between deuterium and its attached carbon in a deuterated liquid crystal. The method is based on transfer of polarization using the DAPT pulse sequence proposed originally for two spin half nuclei, now extended to a spin-1 and a spin-1/2 nuclei. DAPT utilizes the evolution of magnetization of the spin pair under two blocks of phase shifted BLEW-12 pulses on one of the spins separated by a 90 degree pulse on the other spin. The method is easy to implement and does not need to satisfy matching conditions unlike the Hartmann-Hahn cross-polarization. Experimental results presented demonstrate the efficacy of the method.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Stroke, ischemic or hemorrhagic, belongs among the foremost causes of death and disability worldwide. Massive brain swelling is the leading cause of death in large hemispheric strokes and is only modestly alleviated by available treatment. Thrombolysis with tissue plasminogen activator (TPA) is the only approved therapy in acute ischemic stroke, but fear of TPA-mediated hemorrhage is often a reason for withholding this otherwise beneficial treatment. In addition, recanalization of the occluded artery (spontaneously or with thrombolysis) may cause reperfusion injury by promoting brain edema, hemorrhage, and inflammatory cell infiltration. A dominant event underlying these phenomena seems to be disruption of the blood-brain barrier (BBB). In contrast to ischemic stroke, no widely approved clinical therapy exists for intracerebral hemorrhage (ICH), which is associated with poor outcome mainly due to the mass effect of enlarging hematoma and associated brain swelling. Mast cells (MCs) are perivascularly located resident inflammatory cells which contain potent vasoactive, proteolytic, and fibrinolytic substances in their cytoplasmic granules. Experiments from our laboratory showed MC density and their state of granulation to be altered early following focal transient cerebral ischemia, and degranulating MCs were associated with perivascular edema and hemorrhage. (I) Pharmacological MC stabilization led to significantly reduced ischemic brain swelling (40%) and BBB leakage (50%), whereas pharmacological MC degranulation raised these by 90% and 50%, respectively. Pharmacological MC stabilization also revealed a 40% reduction in neutrophil infiltration. Moreover, genetic MC deficiency was associated with an almost 60% reduction in brain swelling, 50% reduction in BBB leakage, and 50% less neutrophil infiltration, compared with controls. (II) TPA induced MC degranulation in vitro. In vivo experiments with post-ischemic TPA administration demonstrated 70- to 100-fold increases in hemorrhage formation (HF) compared with controls HF. HF was significantly reduced by pharmacological MC stabilization at 3 (95%), 6 (75%), and 24 hours (95%) of follow-up. Genetic MC deficiency again supported the role of MCs, leading to 90% reduction in HF at 6 and 24 hours. Pharmacological MC stabilization and genetic MC deficiency were also associated with significant reduction in brain swelling and in neutrophil infiltration. Importantly, these effects translated into a significantly better neurological outcome and lower mortality after 24 hours. (III) Finally, in ICH experiments, pharmacological MC stabilization resulted in significantly less brain swelling, diminished growth in hematoma volume, better neurological scores, and decreased mortality. Pharmacological MC degranulation produced the opposite effects. Genetic MC deficiency revealed a beneficial effect similar to that found with pharmacological MC stabilization. In sum, the role of MCs in these clinically relevant scenarios is supported by a series of experiments performed both in vitro and in vivo. That not only genetic MC deficiency but also drugs targeting MCs could modulate these parameters (translated into better outcome and decreased mortality), suggests a potential therapeutic approach in a number of highly prevalent cerebral insults in which extensive tissue injury is followed by dangerous brain swelling and inflammatory cell infiltration. Furthermore, these experiments could hint at a novel therapy to improve the safety of thrombolytics, and a potential cellular target for those seeking novel forms of treatment for ICH.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Conventional invasive coronary angiography is the clinical gold standard for detecting of coronary artery stenoses. Noninvasive multidetector computed tomography (MDCT) in combination with retrospective ECG gating has recently been shown to permit visualization of the coronary artery lumen and detection of coronary artery stenoses. Single photon emission tomography (SPECT) perfusion imaging has been considered the reference method for evaluation of nonviable myocardium, but magnetic resonance imaging (MRI) can accurately depict structure, function, effusion, and myocardial viability, with an overall capacity unmatched by any other single imaging modality. Magnetocardiography (MCG) provides noninvasively information about myocardial excitation propagation and repolarization without the use of electrodes. This evolving technique may be considered the magnetic equivalent to electrocardiography. The aim of the present series of studies was to evaluate changes in the myocardium assessed with SPECT and MRI caused by coronary artery disease, examine the capability of multidetector computed tomography coronary angiography (MDCT-CA) to detect significant stenoses in the coronary arteries, and MCG to assess remote myocardial infarctions. Our study showed that in severe, progressing coronary artery disease laser treatment does not improve global left ventricular function or myocardial perfusion, but it does preserve systolic wall thickening in fixed defects (scar). It also prevents changes from ischemic myocardial regions to scar. The MCG repolarization variables are informative in remote myocardial infarction, and may perform as well as the conventional QRS criteria in detection of healed myocardial infarction. These STT abnormalities are more pronounced in patients with Q-wave infarction than in patients with non-Q-wave infarctions. MDCT-CA had a sensitivity of 82%, a specificity of 94%, a positive predictive value of 79%, and a negative predictive value of 95% for stenoses over 50% in the main coronary arteries as compared with conventional coronary angiography in patients with known coronary artery disease. Left ventricular wall dysfunction, perfusion defects, and infarctions were detected in 50-78% of sectors assigned to calcifications or stenoses, but also in sectors supplied by normally perfused coronary arteries. Our study showed a low sensitivity (sensitivity 63%) in detecting obstructive coronary artery disease assessed by MDCT in patients with severe aortic stenosis. Massive calcifications complicated correct assessment of the lumen of coronary arteries.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Cavernomas are rare neurovascular lesions, encountered in up to 10% of patients harboring vascular abnormalities of the CNS. Cavernomas consist of dilated thin-walled sinusoids or caverns covered by a single layer of endothelium. Due to advancements in neuroradiology, the number of cavernoma patients coming to be evaluated in neurosurgical practice is increasing. In the present work, we summarized our results on the treatment of cavernomas. Particular attention was paid to uncommon locations or insufficiently investigated cavernomas, including 1. Intraventricular cavernomas; 2. Multiple cavernomas; 3. Spinal cavernomas; and 4. Temporal lobe cavernomas. After analyzing the patient series with these lesions, we concluded that: 1. IVCs are characterized by a high tendency to cause repetitive hemorrhages in a short period of time after the first event. In most patients, hemorrhages were not life-threatening. Surgery is indicated when re-bleedings are frequent and the mass-effect causes progressive neurological deterioration. Modern microsurgical techniques allow safe removal of the IVC, but surgery on fourth ventricle cavernomas carries increased risk of postoperative cranial nerve deficits. 2. In MC cases, when the cavernoma bleeds or generates drug-resistant epilepsy, microsurgical removal of the symptomatic lesion is beneficial to patients. In our series, surgical removal of the most active cavernoma usually the biggest lesion with signs of recent hemorrhage - was safe and prevented further bleedings. Epilepsy outcome showed the effectiveness of active treatment of MCs. However, due to the remaining cavernomas, epileptogenic activity can persist postoperatively, frequently necessitating long-term use of antiepileptic drugs. 3. Spinal cavernomas can cause severe neurological deterioration due to low tolerance of the spinal cord to mass-effect with progressive myelopathy. When aggravated by extralesional massive hemorrhage, neurological decline is usually acute and requires immediate treatment. Microsurgical removal of a cavernoma is effective and safe, improving neurological deficits. Sensorimotor deficits and pain improved postoperatively at a high rate, whereas bladder dysfunction remained essentially unchanged, causing social discomfort to patients. 4. Microsurgical removal of temporal lobe cavernomas is beneficial for patents suffering from drug-resistant epilepsy. In our series, 69% of patients with this condition became seizure-free postoperatively. Duration of epilepsy did not correlate with seizure prognosis. The most frequent disabling symptom at follow-up was memory disorder, considered to be the result of a complex interplay between chronic epilepsy and possible damage to the temporal lobe during surgery.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Mobile applications are being increasingly deployed on a massive scale in various mobile sensor grid database systems. With limited resources from the mobile devices, how to process the huge number of queries from mobile users with distributed sensor grid databases becomes a critical problem for such mobile systems. While the fundamental semantic cache technique has been investigated for query optimization in sensor grid database systems, the problem is still difficult due to the fact that more realistic multi-dimensional constraints have not been considered in existing methods. To solve the problem, a new semantic cache scheme is presented in this paper for location-dependent data queries in distributed sensor grid database systems. It considers multi-dimensional constraints or factors in a unified cost model architecture, determines the parameters of the cost model in the scheme by using the concept of Nash equilibrium from game theory, and makes semantic cache decisions from the established cost model. The scenarios of three factors of semantic, time and locations are investigated as special cases, which improve existing methods. Experiments are conducted to demonstrate the semantic cache scheme presented in this paper for distributed sensor grid database systems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Some naturally occurring strains of fungi cease growing through successive subculturing, i.e., they senesce. In Neurospora, senescing strains usually contain intramitochondrial linear or circular plasmids. An entire plasmid or its part(s) integrates into the mtDNA, causing insertional mutagenesis. The functionally defective mitochondria replicate faster than the wild-type mitochondria and spread through interconnected hyphal cells. Senescence could also be due to spontaneous lethal nuclear gene mutations arising in the multinucleated mycelium. However, their phenotypic effects remain masked until the nuclei segregate into a homokaryotic spore, and the spore germinates to form a mycelium that is incapable of extended culturing. Ultimately the growth of a fungal colony ceases due to dysfunctional oxidative phosphorylation. Results with senescing nuclear mutants or growth-impaired cytoplasmic mutants suggest that mtDNA is inherently unstable, requiring protection by as yet unidentified nuclear-gene-encoded factors for normal functioning. Interestingly, these results are in accord with the endosymbiotic theory of origin of eukaryotic cells.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

It is widely accepted that the global climate is heating up due to human activities, such as burning of fossil fuels. Therefore we find ourselves forced to make decisions on what measures, if any, need to be taken to decrease our warming effect on the planet before any irrevocable damage occurs. Research is being conducted in a variety of fields to better understand all relevant processes governing Earth s climate, and to assess the relative roles of anthropogenic and biogenic emissions into the atmosphere. One of the least well quantified problems is the impact of small aerosol particles (both of anthropogenic and biogenic origin) on climate, through reflecting solar radiation and their ability to act as condensation nuclei for cloud droplets. In this thesis, the compounds driving the biogenic formation of new particles in the atmosphere have been examined through detailed measurements. As directly measuring the composition of these newly formed particles is extremely difficult, the approach was to indirectly study their different characteristics by measuring the hygroscopicity (water uptake) and volatility (evaporation) of particles between 10 and 50 nm. To study the first steps of the formation process in the sub-3 nm range, the nucleation of gaseous precursors to small clusters, the chemical composition of ambient naturally charged ions were measured. The ion measurements were performed with a newly developed mass spectrometer, which was first characterized in the laboratory before being deployed at a boreal forest measurement site. It was also successfully compared to similar, low-resolution instruments. The ambient measurements showed that sulfuric acid clusters dominate the negative ion spectrum during new particle formation events. Sulfuric acid/ammonia clusters were detected in ambient air for the first time in this work. Even though sulfuric acid is believed to be the most important gas phase precursor driving the initial cluster formation, measurements of the hygroscopicity and volatility of growing 10-50 nm particles in Hyytiälä showed an increasing role of organic vapors of a variety of oxidation levels. This work has provided additional insights into the compounds participating both in the initial formation and subsequent growth of atmospheric new aerosol particles. It will hopefully prove an important step in understanding atmospheric gas-to-particle conversion, which, by influencing cloud properties, can have important climate impacts. All available knowledge needs to be constantly updated, summarized, and brought to the attention of our decision-makers. Only by increasing our understanding of all the relevant processes can we build reliable models to predict the long-term effects of decisions made today.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Aerosol particles in the atmosphere are known to significantly influence ecosystems, to change air quality and to exert negative health effects. Atmospheric aerosols influence climate through cooling of the atmosphere and the underlying surface by scattering of sunlight, through warming of the atmosphere by absorbing sun light and thermal radiation emitted by the Earth surface and through their acting as cloud condensation nuclei. Aerosols are emitted from both natural and anthropogenic sources. Depending on their size, they can be transported over significant distances, while undergoing considerable changes in their composition and physical properties. Their lifetime in the atmosphere varies from a few hours to a week. New particle formation is a result of gas-to-particle conversion. Once formed, atmospheric aerosol particles may grow due to condensation or coagulation, or be removed by deposition processes. In this thesis we describe analyses of air masses, meteorological parameters and synoptic situations to reveal conditions favourable for new particle formation in the atmosphere. We studied the concentration of ultrafine particles in different types of air masses, and the role of atmospheric fronts and cloudiness in the formation of atmospheric aerosol particles. The dominant role of Arctic and Polar air masses causing new particle formation was clearly observed at Hyytiälä, Southern Finland, during all seasons, as well as at other measurement stations in Scandinavia. In all seasons and on multi-year average, Arctic and North Atlantic areas were the sources of nucleation mode particles. In contrast, concentrations of accumulation mode particles and condensation sink values in Hyytiälä were highest in continental air masses, arriving at Hyytiälä from Eastern Europe and Central Russia. The most favourable situation for new particle formation during all seasons was cold air advection after cold-front passages. Such a period could last a few days until the next front reached Hyytiälä. The frequency of aerosol particle formation relates to the frequency of low-cloud-amount days in Hyytiälä. Cloudiness of less than 5 octas is one of the factors favouring new particle formation. Cloudiness above 4 octas appears to be an important factor that prevents particle growth, due to the decrease of solar radiation, which is one of the important meteorological parameters in atmospheric particle formation and growth. Keywords: Atmospheric aerosols, particle formation, air mass, atmospheric front, cloudiness

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Fusion energy is a clean and safe solution for the intricate question of how to produce non-polluting and sustainable energy for the constantly growing population. The fusion process does not result in any harmful waste or green-house gases, since small amounts of helium is the only bi-product that is produced when using the hydrogen isotopes deuterium and tritium as fuel. Moreover, deuterium is abundant in seawater and tritium can be bred from lithium, a common metal in the Earth's crust, rendering the fuel reservoirs practically bottomless. Due to its enormous mass, the Sun has been able to utilize fusion as its main energy source ever since it was born. But here on Earth, we must find other means to achieve the same. Inertial fusion involving powerful lasers and thermonuclear fusion employing extreme temperatures are examples of successful methods. However, these have yet to produce more energy than they consume. In thermonuclear fusion, the fuel is held inside a tokamak, which is a doughnut-shaped chamber with strong magnets wrapped around it. Once the fuel is heated up, it is controlled with the help of these magnets, since the required temperatures (over 100 million degrees C) will separate the electrons from the nuclei, forming a plasma. Once the fusion reactions occur, excess binding energy is released as energetic neutrons, which are absorbed in water in order to produce steam that runs turbines. Keeping the power losses from the plasma low, thus allowing for a high number of reactions, is a challenge. Another challenge is related to the reactor materials, since the confinement of the plasma particles is not perfect, resulting in particle bombardment of the reactor walls and structures. Material erosion and activation as well as plasma contamination are expected. Adding to this, the high energy neutrons will cause radiation damage in the materials, causing, for instance, swelling and embrittlement. In this thesis, the behaviour of a material situated in a fusion reactor was studied using molecular dynamics simulations. Simulations of processes in the next generation fusion reactor ITER include the reactor materials beryllium, carbon and tungsten as well as the plasma hydrogen isotopes. This means that interaction models, {\it i.e. interatomic potentials}, for this complicated quaternary system are needed. The task of finding such potentials is nonetheless nearly at its end, since models for the beryllium-carbon-hydrogen interactions were constructed in this thesis and as a continuation of that work, a beryllium-tungsten model is under development. These potentials are combinable with the earlier tungsten-carbon-hydrogen ones. The potentials were used to explain the chemical sputtering of beryllium due to deuterium plasma exposure. During experiments, a large fraction of the sputtered beryllium atoms were observed to be released as BeD molecules, and the simulations identified the swift chemical sputtering mechanism, previously not believed to be important in metals, as the underlying mechanism. Radiation damage in the reactor structural materials vanadium, iron and iron chromium, as well as in the wall material tungsten and the mixed alloy tungsten carbide, was also studied in this thesis. Interatomic potentials for vanadium, tungsten and iron were modified to be better suited for simulating collision cascades that are formed during particle irradiation, and the potential features affecting the resulting primary damage were identified. Including the often neglected electronic effects in the simulations was also shown to have an impact on the damage. With proper tuning of the electron-phonon interaction strength, experimentally measured quantities related to ion-beam mixing in iron could be reproduced. The damage in tungsten carbide alloys showed elemental asymmetry, as the major part of the damage consisted of carbon defects. On the other hand, modelling the damage in the iron chromium alloy, essentially representing steel, showed that small additions of chromium do not noticeably affect the primary damage in iron. Since a complete assessment of the response of a material in a future full-scale fusion reactor is not achievable using only experimental techniques, molecular dynamics simulations are of vital help. This thesis has not only provided insight into complicated reactor processes and improved current methods, but also offered tools for further simulations. It is therefore an important step towards making fusion energy more than a future goal.