86 resultados para Instrumental Methods
Resumo:
During the last 10-15 years interest in mouse behavioural analysis has evolved considerably. The driving force is development in molecular biological techniques that allow manipulation of the mouse genome by changing the expression of genes. Therefore, with some limitations it is possible to study how genes participate in regulation of physiological functions and to create models explaining genetic contribution to various pathological conditions. The first aim of our study was to establish a framework for behavioural phenotyping of genetically modified mice. We established comprehensive battery of tests for the initial screening of mutant mice. These included tests for exploratory and locomotor activity, emotional behaviour, sensory functions, and cognitive performance. Our interest was in the behavioural patterns of common background strains used for genetic manipulations in mice. Additionally we studied the behavioural effect of sex differences, test history, and individual housing. Our findings highlight the importance of careful consideration of genetic background for analysis of mutant mice. It was evident that some backgrounds may mask or modify the behavioural phenotype of mutants and thereby lead to false positive or negative findings. Moreover, there is no universal strain that is equally suitable for all tests, and using different backgrounds allows one to address possible phenotype modifying factors. We discovered that previous experience affected performance in several tasks. The most sensitive traits were the exploratory and emotional behaviour, as well as motor and nociceptive functions. Therefore, it may be essential to repeat some of the tests in naïve animals for assuring the phenotype. Social isolation for a long time period had strong effects on exploratory behaviour, but also on learning and memory. All experiments revealed significant interactions between strain and environmental factors (test history or housing condition) indicating genotype-dependent effects of environmental manipulations. Several mutant line analyses utilize this information. For example, we studied mice overexpressing as well as those lacking extracellular matrix protein heparin-binding growth-associated molecule (HB-GAM), and mice lacking N-syndecan (a receptor for HB-GAM). All mutant mice appeared to be fertile and healthy, without any apparent neurological or sensory defects. The lack of HB-GAM and N-syndecan, however, significantly reduced the learning capacity of the mice. On the other hand, overexpression of HB-GAM resulted in facilitated learning. Moreover, HB-GAM knockout mice displayed higher anxiety-like behaviour, whereas anxiety was reduced in HB-GAM overexpressing mice. Changes in hippocampal plasticity accompanied the behavioural phenotypes. We conclude that HB-GAM and N-syndecan are involved in the modulation of synaptic plasticity in hippocampus and play a role in regulation of anxiety- and learning-related behaviour.
Resumo:
Composting refers to aerobic degradation of organic material and is one of the main waste treatment methods used in Finland for treating separated organic waste. The composting process allows converting organic waste to a humus-like end product which can be used to increase the organic matter in agricultural soils, in gardening, or in landscaping. Microbes play a key role as degraders during the composting-process, and the microbiology of composting has been studied for decades, but there are still open questions regarding the microbiota in industrial composting processes. It is known that with the traditional, culturing-based methods only a small fraction, below 1%, of the species in a sample is normally detected. In recent years an immense diversity of bacteria, fungi and archaea has been found to occupy many different environments. Therefore the methods of characterising microbes constantly need to be developed further. In this thesis the presence of fungi and bacteria in full-scale and pilot-scale composting processes was characterised with cloning and sequencing. Several clone libraries were constructed and altogether nearly 6000 clones were sequenced. The microbial communities detected in this study were found to differ from the compost microbes observed in previous research with cultivation based methods or with molecular methods from processes of smaller scale, although there were similarities as well. The bacterial diversity was high. Based on the non-parametric coverage estimations, the number of bacterial operational taxonomic units (OTU) in certain stages of composting was over 500. Sequences similar to Lactobacillus and Acetobacteria were frequently detected in the early stages of drum composting. In tunnel stages of composting the bacterial community comprised of Bacillus, Thermoactinomyces, Actinobacteria and Lactobacillus. The fungal diversity was found to be high and phylotypes similar to yeasts were abundantly found in the full-scale drum and tunnel processes. In addition to phylotypes similar to Candida, Pichia and Geotrichum moulds from genus Thermomyces and Penicillium were observed in tunnel stages of composting. Zygomycetes were detected in the pilot-scale composting processes and in the compost piles. In some of the samples there were a few abundant phylotypes present in the clone libraries that masked the rare ones. The rare phylotypes were of interest and a method for collecting them from clone libraries for sequencing was developed. With negative selection of the abundant phylotyps the rare ones were picked from the clone libraries. Thus 41% of the clones in the studied clone libraries were sequenced. Since microbes play a central role in composting and in many other biotechnological processes, rapid methods for characterization of microbial diversity would be of value, both scientifically and commercially. Current methods, however, lack sensitivity and specificity and are therefore under development. Microarrays have been used in microbial ecology for a decade to study the presence or absence of certain microbes of interest in a multiplex manner. The sequence database collected in this thesis was used as basis for probe design and microarray development. The enzyme assisted detection method, ligation-detection-reaction (LDR) based microarray, was adapted for species-level detection of microbes characteristic of each stage of the composting process. With the use of a specially designed control probe it was established that a species specific probe can detect target DNA representing as little as 0.04% of total DNA in a sample. The developed microarray can be used to monitor composting processes or the hygienisation of the compost end product. A large compost microbe sequence dataset was collected and analysed in this thesis. The results provide valuable information on microbial community composition during industrial scale composting processes. The microarray method was developed based on the sequence database collected in this study. The method can be utilised in following the fate of interesting microbes during composting process in an extremely sensitive and specific manner. The platform for the microarray is universal and the method can easily be adapted for studying microbes from environments other than compost.
Resumo:
The risk is obvious for soft tissue complications after operative treatment of the Achilles tendon, calcaneal bone or after ankle arthroplasty. Such complications after malleolar fractures are, however, seldom seen. The reason behind these complications is that the soft tissue in this region is tight and does not allow much tension to the wound area after surgery. Furthermore the area of operation may be damaged by swelling after the injury, or can be affected by peripheral vascular disease. While complications in this area are unavoidable, they can be diminished. This study attempts to highlight the possible predisposing factors leading to complications in these operations and on the other hand, to determine the solutions to solve soft tissue problems in this region. The study consists of five papers. The first article is a reprint on the soft tissue reconstruction of 25 patients after their complicated Achilles tendon surgeries were analysed. The second study reviews a series of 126 patients after having undergone an operative treatment of calcaneal bone fractures and analyses the complications and possible reasons behind them. The third part analyses a series of corrections of 35 soft tissue complications after calcaneal fracture operations. The fourth part reviews a series of 7 patients who had undergone complicated ankle arthroplasties. The last article presents a series of post operative lateral defects of the ankle treated with a less frequently used distally based peroneus brevis muscle flap and analyses the results. What can be conducted from these studies is that in general, the results after the correction of even severe soft tissue complications in the ankle region are good. For the small defects around the Achilles tendon, the local flaps are useful, but the larger defects are best treated with a free flap. We found that a long delay from trauma to surgery and a long operating time were predisposing factors that lead to soft tissue complications after operatively treated calcaneal bone fractures. The more severe the injury, the greater the risk for wound complication. Surprisingly, the long-term results after infected calcaneal osteosyntheses were acceptable and the calcaneal bone seems to tolerate chronic infections very well if the soft tissue is reconstructed successfully. Behind the complicated ankle arthroplasties, unexpectedly high number of cases experiencing arteriosclerosis of the lower extremity was found. These complications lead to ankle fusion but can be solved with a free flap if the vascularity is intact or can be reconstructed. For this reason a vascular examination of the lower extremity arteries of the patients going to ankle arthroplasty is strongly recommended. Moreover postoperative lateral malleolar wound infections which typically create lateral ankle defects can successfully be treated with a peroneus brevis muscle flap covered with a free skin graft.
Resumo:
Välikorvaleikkauksiin usein liittyvän välikorvan ja kuuloluuketjun kirurgisen rekonstruktion tavoitteena on luoda olosuhteet, jotka mahdollistavat hyvän kuulon sekä välikorvan säilymisen tulehduksettomana ja ilmapitoisena. Välikorvan rekonstruktiossa on käytetty implanttimateriaaleina perinteisesti potilaan omia kudoksia sekä tarvittaessa erilaisia hajoamattomia biomateriaaleja, mm. titaania ja silikonia. Ongelmana biomateriaalien käytössä voi olla bakteerien adherenssi eli tarttuminen vieraan materiaalin pintaan, mikä saattaa johtaa biofilmin muodostumiseen. Tämä voi aiheuttaa kroonisen, huonosti antibiootteihin reagoivan infektion kudoksessa, mikä usein käytännössä johtaa uusintaleikkaukseen ja implantin poistoon. Maitohappo- ja glykolihappopohjaiset biologisesti hajoavat polymeerit ovat olleet kliinisessä käytössä jo vuosikymmeniä. Niitä on käytetty erityisesti tukimateriaaleina mm. ortopediassa sekä kasvo- ja leukakirurgiassa. Niitä ei ole toistaiseksi käytetty välikorvakirurgiassa. Korvan kuvantamiseen käytetään ensisijaisesti tietokonetomografiaa (TT). TT-tutkimuksen ongelmana on potilaan altistuminen suhteellisen korkealle sädeannokselle, joka kasvaa kumulatiivisesti, jos kuvaus joudutaan toistamaan. Väitöskirjatyö selvittää uuden, aiemmin kliinisessä työssä rutiinisti lähinnä hampaiston ja kasvojen alueen kuvantamiseen käytetyn rajoitetun kartiokeila-TT:n soveltuvuutta korvan alueen kuvantamiseen. Väitöskirjan kahdessa ensimmäisessä osatyössä tutkittiin ja verrattiin kahden kroonisia ja postoperatiivisia korvainfektioita aiheuttavan bakteerin, Staphylococcus aureuksen ja Pseudomonas aeruginosan, in vitro adherenssia titaanin, silikonin ja kahden eri biohajoavan polymeerin (PLGA) pintaan. Lisäksi tutkittiin materiaalien albumiinipinnoituksen vaikutusta adherenssiin. Kolmannessa osatyössä tutkittiin eläinmallissa PLGA:n biokompatibiliteettia eli kudosyhteensopivuutta kokeellisessa välikorvakirurgiassa. Chinchillojen välikorviin istutettiin PLGA-materiaalia, eläimiä seurattiin, ja ne lopetettiin 6 kk:n kuluttua operaatiosta. Biokompatibiliteetin arviointi perustui kliinisiin havaintoihin sekä kudosnäytteisiin. Neljännessä osatyössä tutkittiin kartiokeila-TT:n soveltuvuutta korvan alueen kuvantamiseen vertaamalla sen tarkkuutta perinteisen spiraali-TT:n tarkkuuteen. Molemmilla laitteilla kuvattiin ohimo- eli temporaaliluita korvan alueen kliinisesti ja kirurgisesti tärkeiden rakenteiden kuvantumisen tarkkuuden arvioimiseksi. Viidennessä osatyössä arvioitiin myös operoitujen temporaaliluiden kuvantumista kartiokeila-TT:ssa. Bakteeritutkimuksissa PLGA-materiaalin pintaan tarttui keskimäärin korkeintaan saman verran tai vähemmän bakteereita kuin silikonin tai titaanin. Albumiinipinnoitus vähensi bakteeriadherenssia merkitsevästi kaikilla materiaaleilla. Eläinkokeiden perusteella PLGA todettiin hyvin siedetyksi välikorvassa. Korvakäytävissä tai välikorvissa ei todettu infektioita, tärykalvon perforaatioita tai materiaalin esiin työntymistä. Kudosnäytteissä näkyi lievää tulehdusreaktiota ja fibroosia implantin ympärillä. Temporaaliluutöissä rajoitettu kartiokeila-TT todettiin vähintään yhtä tarkaksi menetelmäksi kuin spiraali-TT välikorvan ja sisäkorvan rakenteiden kuvantamisessa, ja sen aiheuttama kertasäderasitus todettiin spiraali-TT:n vastaavaa huomattavasti vähäisemmäksi. Kartiokeila-TT soveltui hyvin välikorvaimplanttien ja postoperatiivisen korvan kuvantamiseen. Tulokset osoittavat, että PLGA on välikorvakirurgiaan soveltuva, turvallinen ja kudosyhteensopiva biomateriaali. Biomateriaalien pinnoittaminen albumiinilla vähentää merkittävästi bakteeriadherenssia niihin, mikä puoltaa pinnoituksen soveltamista implanttikirurgiassa. Kartiokeila-TT soveltuu korvan alueen kuvantamiseen. Sen tarkkuus kliinisesti tärkeiden rakenteiden osoittamisessa on vähintään yhtä hyvä ja sen potilaalle aiheuttama sädeannos pienempi kuin nykyisen korva-spiraali-TT:n. Tämä tekee menetelmästä spiraali-TT:aa potilasturvallisemman vaihtoehdon erityisesti, jos potilaan tilanne vaatii seurantaa ja useampia kuvauksia, ja jos halutaan kuvata rajoitettuja alueita uni- tai bilateraalisesti.
Resumo:
Idiopathic pulmonary fibrosis (IPF) is an interstitial lung disease with unknown aetiology and poor prognosis. IPF is characterized by alveolar epithelial damage that leads tissue remodelling and ultimately to the loss of normal lung architecture and function. Treatment has been focused on anti-inflammatory therapies, but due to their poor efficacy new therapeutic modalities are being sought. There is a need for early diagnosis and also for differential diagnostic markers for IPF and other interstitial lung diseases. The study utilized patient material obtained from bronchoalveolar lavage (BAL), diagnostic biopsies or lung transplantation. Human pulmonary fibroblast cell cultures were propagated and asbestos-induced pulmonary fibrosis in mice was used as an experimental animal model of IPF. The possible markers for IPF were scanned by immunohistochemistry, RT-PCR, ELISA and western blot. Matrix metalloproteinases (MMPs) are proteolytic enzymes that participate in tissue remodelling. Microarray studies have introduced potential markers that could serve as additional tools for the assessment of IPF and one of the most promising was MMP 7. MMP-7 protein levels were measured in the BAL fluid of patients with idiopathic interstitial lung diseases or idiopathic cough. MMP-7 was however similarly elevated in the BAL fluid of all these disorders and thus cannot be used as a differential diagnostic marker for IPF. Activation of transforming growth factor (TGF)-ß is considered to be a key element in the progression of IPF. Bone morphogenetic proteins (BMP) are negative regulators of intracellular TGF-ß signalling and BMP-4 signalling is in turn negatively regulated by gremlin. Gremlin was found to be highly upregulated in the IPF lungs and IPF fibroblasts. Gremlin was detected in the thickened IPF parenchyma and endothelium of small capillaries, whereas in non-specific interstitial pneumonia it localized predominantly in the alveolar epithelium. Parenchymal gremlin immunoreactivity might indicate IPF-type interstitial pneumonia. Gremlin mRNA levels were higher in patients with end-stage fibrosis suggesting that gremlin might be a marker for more advanced disease. Characterization of the fibroblastic foci in the IPF lungs showed that immunoreactivity to platelet-derived growth factor (PDGF) receptor-α and PDGF receptor-β was elevated in IPF parenchyma, but the fibroblastic foci showed only minor immunoreactivity to the PDGF receptors or the antioxidant peroxiredoxin II. Ki67 positive cells were also observed predominantly outside the fibroblastic foci, suggesting that the fibroblastic foci may not be composed of actively proliferating cells. When inhibition of profibrotic PDGF-signalling by imatinib mesylate was assessed, imatinib mesylate reduced asbestos-induced pulmonary fibrosis in mice as well as human pulmonary fibroblast migration in vitro but it had no effect on the lung inflammation.
Resumo:
Numerical weather prediction (NWP) models provide the basis for weather forecasting by simulating the evolution of the atmospheric state. A good forecast requires that the initial state of the atmosphere is known accurately, and that the NWP model is a realistic representation of the atmosphere. Data assimilation methods are used to produce initial conditions for NWP models. The NWP model background field, typically a short-range forecast, is updated with observations in a statistically optimal way. The objective in this thesis has been to develope methods in order to allow data assimilation of Doppler radar radial wind observations. The work has been carried out in the High Resolution Limited Area Model (HIRLAM) 3-dimensional variational data assimilation framework. Observation modelling is a key element in exploiting indirect observations of the model variables. In the radar radial wind observation modelling, the vertical model wind profile is interpolated to the observation location, and the projection of the model wind vector on the radar pulse path is calculated. The vertical broadening of the radar pulse volume, and the bending of the radar pulse path due to atmospheric conditions are taken into account. Radar radial wind observations are modelled within observation errors which consist of instrumental, modelling, and representativeness errors. Systematic and random modelling errors can be minimized by accurate observation modelling. The impact of the random part of the instrumental and representativeness errors can be decreased by calculating spatial averages from the raw observations. Model experiments indicate that the spatial averaging clearly improves the fit of the radial wind observations to the model in terms of observation minus model background (OmB) standard deviation. Monitoring the quality of the observations is an important aspect, especially when a new observation type is introduced into a data assimilation system. Calculating the bias for radial wind observations in a conventional way can result in zero even in case there are systematic differences in the wind speed and/or direction. A bias estimation method designed for this observation type is introduced in the thesis. Doppler radar radial wind observation modelling, together with the bias estimation method, enables the exploitation of the radial wind observations also for NWP model validation. The one-month model experiments performed with the HIRLAM model versions differing only in a surface stress parameterization detail indicate that the use of radar wind observations in NWP model validation is very beneficial.
Resumo:
Modern-day weather forecasting is highly dependent on Numerical Weather Prediction (NWP) models as the main data source. The evolving state of the atmosphere with time can be numerically predicted by solving a set of hydrodynamic equations, if the initial state is known. However, such a modelling approach always contains approximations that by and large depend on the purpose of use and resolution of the models. Present-day NWP systems operate with horizontal model resolutions in the range from about 40 km to 10 km. Recently, the aim has been to reach operationally to scales of 1 4 km. This requires less approximations in the model equations, more complex treatment of physical processes and, furthermore, more computing power. This thesis concentrates on the physical parameterization methods used in high-resolution NWP models. The main emphasis is on the validation of the grid-size-dependent convection parameterization in the High Resolution Limited Area Model (HIRLAM) and on a comprehensive intercomparison of radiative-flux parameterizations. In addition, the problems related to wind prediction near the coastline are addressed with high-resolution meso-scale models. The grid-size-dependent convection parameterization is clearly beneficial for NWP models operating with a dense grid. Results show that the current convection scheme in HIRLAM is still applicable down to a 5.6 km grid size. However, with further improved model resolution, the tendency of the model to overestimate strong precipitation intensities increases in all the experiment runs. For the clear-sky longwave radiation parameterization, schemes used in NWP-models provide much better results in comparison with simple empirical schemes. On the other hand, for the shortwave part of the spectrum, the empirical schemes are more competitive for producing fairly accurate surface fluxes. Overall, even the complex radiation parameterization schemes used in NWP-models seem to be slightly too transparent for both long- and shortwave radiation in clear-sky conditions. For cloudy conditions, simple cloud correction functions are tested. In case of longwave radiation, the empirical cloud correction methods provide rather accurate results, whereas for shortwave radiation the benefit is only marginal. Idealised high-resolution two-dimensional meso-scale model experiments suggest that the reason for the observed formation of the afternoon low level jet (LLJ) over the Gulf of Finland is an inertial oscillation mechanism, when the large-scale flow is from the south-east or west directions. The LLJ is further enhanced by the sea-breeze circulation. A three-dimensional HIRLAM experiment, with a 7.7 km grid size, is able to generate a similar LLJ flow structure as suggested by the 2D-experiments and observations. It is also pointed out that improved model resolution does not necessary lead to better wind forecasts in the statistical sense. In nested systems, the quality of the large-scale host model is really important, especially if the inner meso-scale model domain is small.
Resumo:
This work focuses on the role of macroseismology in the assessment of seismicity and probabilistic seismic hazard in Northern Europe. The main type of data under consideration is a set of macroseismic observations available for a given earthquake. The macroseismic questionnaires used to collect earthquake observations from local residents since the late 1800s constitute a special part of the seismological heritage in the region. Information of the earthquakes felt on the coasts of the Gulf of Bothnia between 31 March and 2 April 1883 and on 28 July 1888 was retrieved from the contemporary Finnish and Swedish newspapers, while the earthquake of 4 November 1898 GMT is an example of an early systematic macroseismic survey in the region. A data set of more than 1200 macroseismic questionnaires is available for the earthquake in Central Finland on 16 November 1931. Basic macroseismic investigations including preparation of new intensity data point (IDP) maps were conducted for these earthquakes. Previously disregarded usable observations were found in the press. The improved collection of IDPs of the 1888 earthquake shows that this event was a rare occurrence in the area. In contrast to earlier notions it was felt on both sides of the Gulf of Bothnia. The data on the earthquake of 4 November 1898 GMT were augmented with historical background information discovered in various archives and libraries. This earthquake was of some concern to the authorities, because extra fire inspections were conducted in three towns at least, i.e. Tornio, Haparanda and Piteå, located in the centre of the area of perceptibility. This event posed the indirect hazard of fire, although its magnitude around 4.6 was minor on the global scale. The distribution of slightly damaging intensities was larger than previously outlined. This may have resulted from the amplification of the ground shaking in the soft soil of the coast and river valleys where most of the population was found. The large data set of the 1931 earthquake provided an opportunity to apply statistical methods and assess methodologies that can be used when dealing with macroseismic intensity. It was evaluated using correspondence analysis. Different approaches such as gridding were tested to estimate the macroseismic field from the intensity values distributed irregularly in space. In general, the characteristics of intensity warrant careful consideration. A more pervasive perception of intensity as an ordinal quantity affected by uncertainties is advocated. A parametric earthquake catalogue comprising entries from both the macroseismic and instrumental era was used for probabilistic seismic hazard assessment. The parametric-historic methodology was applied to estimate seismic hazard at a given site in Finland and to prepare a seismic hazard map for Northern Europe. The interpretation of these results is an important issue, because the recurrence times of damaging earthquakes may well exceed thousands of years in an intraplate setting such as Northern Europe. This application may therefore be seen as an example of short-term hazard assessment.
Resumo:
An efficient and statistically robust solution for the identification of asteroids among numerous sets of astrometry is presented. In particular, numerical methods have been developed for the short-term identification of asteroids at discovery, and for the long-term identification of scarcely observed asteroids over apparitions, a task which has been lacking a robust method until now. The methods are based on the solid foundation of statistical orbital inversion properly taking into account the observational uncertainties, which allows for the detection of practically all correct identifications. Through the use of dimensionality-reduction techniques and efficient data structures, the exact methods have a loglinear, that is, O(nlog(n)), computational complexity, where n is the number of included observation sets. The methods developed are thus suitable for future large-scale surveys which anticipate a substantial increase in the astrometric data rate. Due to the discontinuous nature of asteroid astrometry, separate sets of astrometry must be linked to a common asteroid from the very first discovery detections onwards. The reason for the discontinuity in the observed positions is the rotation of the observer with the Earth as well as the motion of the asteroid and the observer about the Sun. Therefore, the aim of identification is to find a set of orbital elements that reproduce the observed positions with residuals similar to the inevitable observational uncertainty. Unless the astrometric observation sets are linked, the corresponding asteroid is eventually lost as the uncertainty of the predicted positions grows too large to allow successful follow-up. Whereas the presented identification theory and the numerical comparison algorithm are generally applicable, that is, also in fields other than astronomy (e.g., in the identification of space debris), the numerical methods developed for asteroid identification can immediately be applied to all objects on heliocentric orbits with negligible effects due to non-gravitational forces in the time frame of the analysis. The methods developed have been successfully applied to various identification problems. Simulations have shown that the methods developed are able to find virtually all correct linkages despite challenges such as numerous scarce observation sets, astrometric uncertainty, numerous objects confined to a limited region on the celestial sphere, long linking intervals, and substantial parallaxes. Tens of previously unknown main-belt asteroids have been identified with the short-term method in a preliminary study to locate asteroids among numerous unidentified sets of single-night astrometry of moving objects, and scarce astrometry obtained nearly simultaneously with Earth-based and space-based telescopes has been successfully linked despite a substantial parallax. Using the long-term method, thousands of realistic 3-linkages typically spanning several apparitions have so far been found among designated observation sets each spanning less than 48 hours.
Resumo:
Atmospheric aerosol particles have a strong impact on the global climate. A deep understanding of the physical and chemical processes affecting the atmospheric aerosol climate system is crucial in order to describe those processes properly in global climate models. Besides the climatic effects, aerosol particles can deteriorate e.g. visibility and human health. Nucleation is a fundamental step in atmospheric new particle formation. However, details of the atmospheric nucleation mechanisms have remained unresolved. The main reason for that has been the non-existence of instruments capable of measuring neutral newly formed particles in the size range below 3 nm in diameter. This thesis aims to extend the detectable particle size range towards close-to-molecular sizes (~1nm) of freshly nucleated clusters, and by direct measurement obtain the concentrations of sub-3 nm particles in atmospheric environment and in well defined laboratory conditions. In the work presented in this thesis, new methods and instruments for the sub-3 nm particle detection were developed and tested. The selected approach comprises four different condensation based techniques and one electrical detection scheme. All of them are capable to detect particles with diameters well below 3 nm, some even down to ~1 nm. The developed techniques and instruments were deployed in the field measurements as well as in laboratory nucleation experiments. Ambient air studies showed that in a boreal forest environment a persistent population of 1-2 nm particles or clusters exists. The observation was done using 4 different instruments showing a consistent capability for the direct measurement of the atmospheric nucleation. The results from the laboratory experiments showed that sulphuric acid is a key species in the atmospheric nucleation. The mismatch between the earlier laboratory data and ambient observations on the dependency of nucleation rate on sulphuric acid concentration was explained. The reason was shown to be associated in the inefficient growth of the nucleated clusters and in the insufficient detection efficiency of particle counters used in the previous experiments. Even though the exact molecular steps of nucleation still remain an open question, the instrumental techniques developed in this work as well as their application in laboratory and ambient studies opened a new view into atmospheric nucleation and prepared the way for investigating the nucleation processes with more suitable tools.