10 resultados para computer-based tools
em Helda - Digital Repository of University of Helsinki
Resumo:
Finite-state methods have been adopted widely in computational morphology and related linguistic applications. To enable efficient development of finite-state based linguistic descriptions, these methods should be a freely available resource for academic language research and the language technology industry. The following needs can be identified: (i) a registry that maps the existing approaches, implementations and descriptions, (ii) managing the incompatibilities of the existing tools, (iii) increasing synergy and complementary functionality of the tools, (iv) persistent availability of the tools used to manipulate the archived descriptions, (v) an archive for free finite-state based tools and linguistic descriptions. Addressing these challenges contributes to building a common research infrastructure for advanced language technology.
Resumo:
Strategies of scientific, question-driven inquiry are stated to be important cultural practices that should be educated in schools and universities. The present study focuses on investigating multiple efforts to implement a model of Progressive Inquiry and related Web-based tools in primary, secondary and university level education, to develop guidelines for educators in promoting students collaborative inquiry practices with technology. The research consists of four studies. In Study I, the aims were to investigate how a human tutor contributed to the university students collaborative inquiry process through virtual forums, and how the influence of the tutoring activities is demonstrated in the students inquiry discourse. Study II examined an effort to implement technology-enhanced progressive inquiry as a distance working project in a middle school context. Study III examined multiple teachers' methods of organizing progressive inquiry projects in primary and secondary classrooms through a generic analysis framework. In Study IV, a design-based research effort consisting of four consecutive university courses, applying progressive inquiry pedagogy, was retrospectively re-analyzed in order to develop the generic design framework. The results indicate that appropriate teacher support for students collaborative inquiry efforts appears to include interplay between spontaneity and structure. Careful consideration should be given to content mastery, critical working strategies or essential knowledge practices that the inquiry approach is intended to promote. In particular, those elements in students activities should be structured and directed, which are central to the aim of Progressive Inquiry, but which the students do not recognize or demonstrate spontaneously, and which are usually not taken into account in existing pedagogical methods or educational conventions. Such elements are, e.g., productive co-construction activities; sustained engagement in improving produced ideas and explanations; critical reflection of the adopted inquiry practices, and sophisticated use of modern technology for knowledge work. Concerning the scaling-up of inquiry pedagogy, it was concluded that one individual teacher can also apply the principles of Progressive Inquiry in his or her own teaching in many innovative ways, even under various institutional constraints. The developed Pedagogical Infrastructure Framework enabled recognizing and examining some central features and their interplay in the designs of examined inquiry units. The framework may help to recognize and critically evaluate the invisible learning-cultural conventions in various educational settings and can mediate discussions about how to overcome or change them.
Resumo:
A small fraction of the energy absorbed in the light reactions of photosynthesis is re-emitted as chlorophyll-a fluorescence. Chlorophyll-a fluorescence and photochemistry compete for excitation energy in photosystem II (PSII). Therefore, changes in the photochemical capacity can be detected through analysis of chlorophyll fluorescence. Chlorophyll fluorescence techniques have been widely used to follow the diurnal (fast), and the seasonal (slow) acclimation in the energy partitioning between photochemical and non-photochemical processes in PSII. Energy partitioning in PSII estimated through chlorophyll fluorescence can be used as a proxy of the plant physiological status, and measured at different spatial and temporal scales. However, a number of technical and theoretical limitations still limit the use of chlorophyll fluorescence data for the study of the acclimation of PSII. The aim of this Thesis was to study the diurnal and seasonal acclimation of PSII in field conditions through the development and testing of new chlorophyll fluorescence-based tools, overcoming these limitations. A new model capable of following the fast acclimation of PSII to rapid fluctuations in light intensity was developed. The model was used to study the rapid acclimation in the electron transport rate under fluctuating light. Additionally, new chlorophyll fluorescence parameters were developed for estimating the seasonal acclimation in the sustained rate constant of thermal energy dissipation and photochemistry. The parameters were used to quantitatively evaluate the effect of light and temperature on the seasonal acclimation of PSII. The results indicated that light environment not only affected the degree but also the kinetics of response of the acclimation to temperature, which was attributed to differences in the structural organization of PSII during seasonal acclimation. Furthermore, zeaxanthin-facilitated thermal dissipation appeared to be the main mechanisms modulating the fraction of absorbed energy being dissipated thermally during winter in field Scots pine. Finally, the integration between diurnal and seasonal acclimation mechanisms was studied using a recently developed instrument MONI-PAM (Walz GmbH, Germany) capable of continuously monitoring the energy partitioning in PSII.
Resumo:
Light scattering, or scattering and absorption of electromagnetic waves, is an important tool in all remote-sensing observations. In astronomy, the light scattered or absorbed by a distant object can be the only source of information. In Solar-system studies, the light-scattering methods are employed when interpreting observations of atmosphereless bodies such as asteroids, atmospheres of planets, and cometary or interplanetary dust. Our Earth is constantly monitored from artificial satellites at different wavelengths. With remote sensing of Earth the light-scattering methods are not the only source of information: there is always the possibility to make in situ measurements. The satellite-based remote sensing is, however, superior in the sense of speed and coverage if only the scattered signal can be reliably interpreted. The optical properties of many industrial products play a key role in their quality. Especially for products such as paint and paper, the ability to obscure the background and to reflect light is of utmost importance. High-grade papers are evaluated based on their brightness, opacity, color, and gloss. In product development, there is a need for computer-based simulation methods that could predict the optical properties and, therefore, could be used in optimizing the quality while reducing the material costs. With paper, for instance, pilot experiments with an actual paper machine can be very time- and resource-consuming. The light-scattering methods presented in this thesis solve rigorously the interaction of light and material with wavelength-scale structures. These methods are computationally demanding, thus the speed and accuracy of the methods play a key role. Different implementations of the discrete-dipole approximation are compared in the thesis and the results provide practical guidelines in choosing a suitable code. In addition, a novel method is presented for the numerical computations of orientation-averaged light-scattering properties of a particle, and the method is compared against existing techniques. Simulation of light scattering for various targets and the possible problems arising from the finite size of the model target are discussed in the thesis. Scattering by single particles and small clusters is considered, as well as scattering in particulate media, and scattering in continuous media with porosity or surface roughness. Various techniques for modeling the scattering media are presented and the results are applied to optimizing the structure of paper. However, the same methods can be applied in light-scattering studies of Solar-system regoliths or cometary dust, or in any remote-sensing problem involving light scattering in random media with wavelength-scale structures.
Resumo:
Epidemiological studies have shown an elevation in the incidence of asthma, allergic symptoms and respiratory infections among people living or working in buildings with moisture and mould problems. Microbial growth is suspected to have a key role, since the severity of microbial contamination and symptoms show a positive correlation, while the removal of contaminated materials relieves the symptoms. However, the cause-and-effect relationship has not been well established and knowledge of the causative agents is incomplete. The present consensus of indoor microbes relies on culture-based methods. Microbial cultivation and identification is known to provide qualitatively and quantitatively biased results, which is suspected to be one of the reasons behind the often inconsistent findings between objectively measured microbiological attributes and health. In the present study the indoor microbial communities were assessed using culture-independent, DNA based methods. Fungal and bacterial diversity was determined by amplifying and sequencing the nucITS- and16S-gene regions, correspondingly. In addition, the cell equivalent numbers of 69 mould species or groups were determined by quantitative PCR (qPCR). The results from molecular analyses were compared with results obtained using traditional plate cultivation for fungi. Using DNA-based tools, the indoor microbial diversity was found to be consistently higher and taxonomically wider than viable diversity. The dominant sequence types of fungi, and also of bacteria were mainly affiliated with well-known microbial species. However, in each building they were accompanied by various rare, uncultivable and unknown species. In both moisture-damaged and undamaged buildings the dominant fungal sequence phylotypes were affiliated with the classes Dothideomycetes (mould-like filamentous ascomycetes); Agaricomycetes (mushroom- and polypore-like filamentous basidiomycetes); Urediniomycetes (rust-like basidiomycetes); Tremellomycetes and the family Malasseziales (both yeast-like basidiomycetes). The most probable source for the majority of fungal types was the outdoor environment. In contrast, the dominant bacterial phylotypes in both damaged and undamaged buildings were affiliated with human-associated members within the phyla Actinobacteria and Firmicutes. Indications of elevated fungal diversity within potentially moisture-damage-associated fungal groups were recorded in two of the damaged buildings, while one of the buildings was characterized by an abundance of members of the Penicillium chrysogenum and P. commune species complexes. However, due to the small sample number and strong normal variation firm conclusions concerning the effect of moisture damage on the species diversity could not be made. The fungal communities in dust samples showed seasonal variation, which reflected the seasonal fluctuation of outdoor fungi. Seasonal variation of bacterial communities was less clear but to some extent attributable to the outdoor sources as well. The comparison of methods showed that clone library sequencing was a feasible method for describing the total microbial diversity, indicated a moderate quantitative correlation between sequencing and qPCR results and confirmed that culture based methods give both a qualitative and quantitative underestimate of microbial diversity in the indoor environment. However, certain important indoor fungi such as Penicillium spp. were clearly underrepresented in the sequence material, probably due to their physiological and genetic properties. Species specific qPCR was a more efficient and sensitive method for detecting and quantitating individual species than sequencing, but in order to exploit the full advantage of the method in building investigations more information is needed about the microbial species growing on damaged materials. In the present study, a new method was also developed for enhanced screening of the marker gene clone libraries. The suitability of the screening method to different kinds of microbial environments including biowaste compost material and indoor settled dusts was evaluated. The usability was found to be restricted to environments that support the growth and subsequent dominance of a small number microbial species, such as compost material.
Resumo:
This thesis studies human gene expression space using high throughput gene expression data from DNA microarrays. In molecular biology, high throughput techniques allow numerical measurements of expression of tens of thousands of genes simultaneously. In a single study, this data is traditionally obtained from a limited number of sample types with a small number of replicates. For organism-wide analysis, this data has been largely unavailable and the global structure of human transcriptome has remained unknown. This thesis introduces a human transcriptome map of different biological entities and analysis of its general structure. The map is constructed from gene expression data from the two largest public microarray data repositories, GEO and ArrayExpress. The creation of this map contributed to the development of ArrayExpress by identifying and retrofitting the previously unusable and missing data and by improving the access to its data. It also contributed to creation of several new tools for microarray data manipulation and establishment of data exchange between GEO and ArrayExpress. The data integration for the global map required creation of a new large ontology of human cell types, disease states, organism parts and cell lines. The ontology was used in a new text mining and decision tree based method for automatic conversion of human readable free text microarray data annotations into categorised format. The data comparability and minimisation of the systematic measurement errors that are characteristic to each lab- oratory in this large cross-laboratories integrated dataset, was ensured by computation of a range of microarray data quality metrics and exclusion of incomparable data. The structure of a global map of human gene expression was then explored by principal component analysis and hierarchical clustering using heuristics and help from another purpose built sample ontology. A preface and motivation to the construction and analysis of a global map of human gene expression is given by analysis of two microarray datasets of human malignant melanoma. The analysis of these sets incorporate indirect comparison of statistical methods for finding differentially expressed genes and point to the need to study gene expression on a global level.
Resumo:
Ubiquitous computing is about making computers and computerized artefacts a pervasive part of our everyday lifes, bringing more and more activities into the realm of information. The computationalization, informationalization of everyday activities increases not only our reach, efficiency and capabilities but also the amount and kinds of data gathered about us and our activities. In this thesis, I explore how information systems can be constructed so that they handle this personal data in a reasonable manner. The thesis provides two kinds of results: on one hand, tools and methods for both the construction as well as the evaluation of ubiquitous and mobile systems---on the other hand an evaluation of the privacy aspects of a ubiquitous social awareness system. The work emphasises real-world experiments as the most important way to study privacy. Additionally, the state of current information systems as regards data protection is studied. The tools and methods in this thesis consist of three distinct contributions. An algorithm for locationing in cellular networks is proposed that does not require the location information to be revealed beyond the user's terminal. A prototyping platform for the creation of context-aware ubiquitous applications called ContextPhone is described and released as open source. Finally, a set of methodological findings for the use of smartphones in social scientific field research is reported. A central contribution of this thesis are the pragmatic tools that allow other researchers to carry out experiments. The evaluation of the ubiquitous social awareness application ContextContacts covers both the usage of the system in general as well as an analysis of privacy implications. The usage of the system is analyzed in the light of how users make inferences of others based on real-time contextual cues mediated by the system, based on several long-term field studies. The analysis of privacy implications draws together the social psychological theory of self-presentation and research in privacy for ubiquitous computing, deriving a set of design guidelines for such systems. The main findings from these studies can be summarized as follows: The fact that ubiquitous computing systems gather more data about users can be used to not only study the use of such systems in an effort to create better systems but in general to study phenomena previously unstudied, such as the dynamic change of social networks. Systems that let people create new ways of presenting themselves to others can be fun for the users---but the self-presentation requires several thoughtful design decisions that allow the manipulation of the image mediated by the system. Finally, the growing amount of computational resources available to the users can be used to allow them to use the data themselves, rather than just being passive subjects of data gathering.
Resumo:
The analysis of lipid compositions from biological samples has become increasingly important. Lipids have a role in cardiovascular disease, metabolic syndrome and diabetes. They also participate in cellular processes such as signalling, inflammatory response, aging and apoptosis. Also, the mechanisms of regulation of cell membrane lipid compositions are poorly understood, partially because a lack of good analytical methods. Mass spectrometry has opened up new possibilities for lipid analysis due to its high resolving power, sensitivity and the possibility to do structural identification by fragment analysis. The introduction of Electrospray ionization (ESI) and the advances in instrumentation revolutionized the analysis of lipid compositions. ESI is a soft ionization method, i.e. it avoids unwanted fragmentation the lipids. Mass spectrometric analysis of lipid compositions is complicated by incomplete separation of the signals, the differences in the instrument response of different lipids and the large amount of data generated by the measurements. These factors necessitate the use of computer software for the analysis of the data. The topic of the thesis is the development of methods for mass spectrometric analysis of lipids. The work includes both computational and experimental aspects of lipid analysis. The first article explores the practical aspects of quantitative mass spectrometric analysis of complex lipid samples and describes how the properties of phospholipids and their concentration affect the response of the mass spectrometer. The second article describes a new algorithm for computing the theoretical mass spectrometric peak distribution, given the elemental isotope composition and the molecular formula of a compound. The third article introduces programs aimed specifically for the analysis of complex lipid samples and discusses different computational methods for separating the overlapping mass spectrometric peaks of closely related lipids. The fourth article applies the methods developed by simultaneously measuring the progress curve of enzymatic hydrolysis for a large number of phospholipids, which are used to determine the substrate specificity of various A-type phospholipases. The data provides evidence that the substrate efflux from bilayer is the key determining factor for the rate of hydrolysis.
Resumo:
Gene mapping is a systematic search for genes that affect observable characteristics of an organism. In this thesis we offer computational tools to improve the efficiency of (disease) gene-mapping efforts. In the first part of the thesis we propose an efficient simulation procedure for generating realistic genetical data from isolated populations. Simulated data is useful for evaluating hypothesised gene-mapping study designs and computational analysis tools. As an example of such evaluation, we demonstrate how a population-based study design can be a powerful alternative to traditional family-based designs in association-based gene-mapping projects. In the second part of the thesis we consider a prioritisation of a (typically large) set of putative disease-associated genes acquired from an initial gene-mapping analysis. Prioritisation is necessary to be able to focus on the most promising candidates. We show how to harness the current biomedical knowledge for the prioritisation task by integrating various publicly available biological databases into a weighted biological graph. We then demonstrate how to find and evaluate connections between entities, such as genes and diseases, from this unified schema by graph mining techniques. Finally, in the last part of the thesis, we define the concept of reliable subgraph and the corresponding subgraph extraction problem. Reliable subgraphs concisely describe strong and independent connections between two given vertices in a random graph, and hence they are especially useful for visualising such connections. We propose novel algorithms for extracting reliable subgraphs from large random graphs. The efficiency and scalability of the proposed graph mining methods are backed by extensive experiments on real data. While our application focus is in genetics, the concepts and algorithms can be applied to other domains as well. We demonstrate this generality by considering coauthor graphs in addition to biological graphs in the experiments.