214 resultados para tiedon jakaminen


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Department of Forest Resource Management in the University of Helsinki has in years 2004?2007 carried out so-called SIMO -project to develop a new generation planning system for forest management. Project parties are organisations doing most of Finnish forest planning in government, industry and private owned forests. Aim of this study was to find out the needs and requirements for new forest planning system and to clarify how parties see targets and processes in today's forest planning. Representatives responsible for forest planning in each organisation were interviewed one by one. According to study the stand-based system for managing and treating forests continues in the future. Because of variable data acquisition methods with different accuracy and sources, and development of single tree interpretation, more and more forest data is collected without field work. The benefits of using more specific forest data also calls for use of information units smaller than tree stand. In Finland the traditional way to arrange forest planning computation is divided in two elements. After updating the forest data to present situation every stand unit's growth is simulated with different alternative treatment schedule. After simulation, optimisation selects for every stand one treatment schedule so that the management program satisfies the owner's goals in the best possible way. This arrangement will be maintained in the future system. The parties' requirements to add multi-criteria problem solving, group decision support methods as well as heuristic and spatial optimisation into system make the programming work more challenging. Generally the new system is expected to be adjustable and transparent. Strict documentation and free source code helps to bring these expectations into effect. Variable growing models and treatment schedules with different source information, accuracy, methods and the speed of processing are supposed to work easily in system. Also possibilities to calibrate models regionally and to set local parameters changing in time are required. In future the forest planning system will be integrated in comprehensive data management systems together with geographic, economic and work supervision information. This requires a modular method of implementing the system and the use of a simple data transmission interface between modules and together with other systems. No major differences in parties' view of the systems requirements were noticed in this study. Rather the interviews completed the full picture from slightly different angles. In organisation the forest management is considered quite inflexible and it only draws the strategic lines. It does not yet have a role in operative activity, although the need and benefits of team level forest planning are admitted. Demands and opportunities of variable forest data, new planning goals and development of information technology are known. Party organisations want to keep on track with development. One example is the engagement in extensive SIMO-project which connects the whole field of forest planning in Finland.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In Finland, suckler cow production is carried out in circumstances characterized by a long winter period and a short grazing period. The traditional winter housing system for suckler cows has been insulated or uninsulated buildings, but there is a demand for developing less expensive housing systems. In addition, more information is needed on new winter feeding strategies, carried out in inexpensive winter facilities with conventional (hay, grass silage, straw) or alternative (treated straw, industrial by-product, whole-crop silage) feeds. The new feeding techniques should not have any detrimental effects on animal welfare in order to be acceptable to both farmers and consumers. Furthermore, no official feeding recommendations for suckler cows are available in Finland and, thus, recommendations for dairy cows have been used. However, this may lead to over- or underfeeding of suckler cows and, finally, to decreased economic output. In Experiment I, second-calf beef-dairy suckler cows were used to compare the effects of diets based on hay (H) or urea-treated straw (US) at two feeding levels (Moderate; M vs. Low; L) on the performance of cows and calves. Live weight (LW) gain during the indoor feeding was lower for cows on level L than on level M. Cows on diet US lost more LW indoors than those on diet H. The cows replenished the LW losses on good pasture. Calf LW gain and cow milk production were unaffected by the treatments. Conception rate was unaffected by the treatments but was only 69%. Urea-treated straw proved to be a suitable winter feed for spring-calving suckler cows. Experiment II studied the effects of feeding accuracy on the performance of first- and second-calf beef-dairy cows and calves. In II-1, the day-to-day variation in the roughage offered ranged up to ± 40%. In II-2, the same variation was used in two-week periods. Variation of the roughages offered had minor effects on cow performance. Reproduction was unaffected by the feeding accuracy. Accurate feeding is not necessary for young beef-dairy crosses, if the total amount of energy offered over a period of a few weeks fulfills the energy requirements. Effects of feeding strategies with alternative feeds on the performance of mature beef-dairy and beef cows and calves were evaluated in Experiment III. Two studies consisted of two feeding strategies (Step-up vs. Flat-rate) and two diets (Control vs. Alternative). There were no differences between treatments in the cow LW, body condition score (BCS), calf pre-weaning LW gain and cow reproduction. A flat-rate strategy can be practised in the nutrition of mature suckler cows. Oat hull based flour-mill by product can partly replace grass silage and straw in the winter diet. Whole-crop barley silage can be offered as a sole feed to suckler cows. Experiment IV evaluated during the winter feeding period the effects of replacing grass silage with whole-crop barley or oat silage on mature beef cow and calf performance. Both whole-crop silages were suitable winter feeds for suckler cows in cold outdoor winter conditions. Experiment V aimed at assessing the effects of daily feeding vs. feeding every third day on the performance of mature beef cows and calves. No differences between the treatments were observed in cow LW, BCS, milk production and calf LW. The serum concentrations of urea and long-chain fatty acids were increased on the third day after feeding in the cows fed every third day. Despite of that the feeding every third day is an acceptable feeding strategy for mature suckler cows. Experiment VI studied the effects of feeding levels and long-term cold climatic conditions on mature beef cows and calves. The cows were overwintered in outdoor facilities or in an uninsulated indoor facility. Whole-crop barley silage was offered either ad libitum or restricted. All the facilities offered adequate shelter for the cows. The restricted offering of whole-crop barley silage provided enough energy for the cows. The Finnish energy recommendations for dairy cows were too high for mature beef breed suckler cows in good body condition at housing, even in cold conditions. Therefore, there is need to determine feeding recommendations for suckler cows in Finland. The results showed that the required amount of energy can be offered to the cows using conventional or alternative feeds provided at a lower feeding level, with an inaccurate feeding, flat-rate feeding or feeding every third day strategy. The cows must have an opportunity to replenish the LW and BCS losses at pasture before the next winter. Production in cold conditions can be practised in inexpensive facilities when shelter against rain and wind, a dry resting place, adequate amounts of feed suitable for cold conditions and water are provided for the animals as was done in the present study.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Rhizoremediation is the use of microbial populations present in the rhizosphere of plants for environmental cleanup. The idea of this work was that bacteria living in the rhizosphere of a nitrogen-fixing leguminous plant, goat's rue (Galega orientalis), could take part in the degradation of harmful monoaromatic hydrocarbons, such as benzene, toluene and xylene (BTEX), from oil-contaminated soils. In addition to chemical (e.g. pollutant concentration) and physical (e.g. soil structure) information, the knowledge of biological aspects (e.g. bacteria and their catabolic genes) is essential when developing the rhizoremediation into controlled and effective bioremediation practice. Therefore, the need for reliable biomonitoring methods is obvious. The main aims of this thesis were to evaluate the symbiotic G. orientalis - Rhizobium galegae system for rhizoremediation of oil-contaminated soils, to develop molecular methods for biomonitoring, and to apply these methods for studying the microbiology of rhizoremediation. In vitro, Galega plants and rhizobia remained viable in m-toluate concentrations up to 3000 mg/l. Plant growth and nodulation were inhibited in 500 mg/l m-toluate, but were restored when plants were transferred to clean medium. In the greenhouse, Galega showed good growth, nodulation and nitrogen fixation, and developed a strong rhizosphere in soils contaminated with oil or spiked with 2000 mg/l m-toluate. The high aromatic tolerance of R. galegae and the viability of Galega plants in oil-polluted soils proved this legume system to be a promising method for the rhizoremediation of oil-contaminated soils. Molecular biomonitoring methods were designed and/or developed further for bacteria and their degradation genes. A combination of genomic fingerprinting ((GTG)5-PCR), taxonomic ribotyping of 16S rRNA genes and partial 16S rRNA gene sequencing were chosen for molecular grouping of culturable, heterogeneous rhizosphere bacteria. PCR primers specific for the xylE gene were designed for TOL plasmid detection. Amplified enzyme-coding DNA restriction analysis (AEDRA) with AluI was used to profile both TOL plasmids (xylE primers) and, in general, aromatics-degrading plasmids (C230 primers). The sensitivity of the direct monitoring of TOL plasmids in soil was enhanced by nested C23O-xylE-PCR. Rhizosphere bacteria were isolated from the greenhouse and field lysimeter experiments. High genetic diversity was observed among the 50 isolated, m-toluate tolerating rhizosphere bacteria in the form of five major lineages of the Bacteria domain. Gram-positive Rhodococcus, Bacillus and Arthrobacter and gram-negative Pseudomonas were the most abundant genera. The inoculum Pseudomonas putida PaW85/pWW0 was not found in the rhizosphere samples. Even if there were no ecological niches available for the bioaugmentation bacterium itself, its conjugative catabolic plasmid might have had some additional value for other bacterial species and thus, for rhizoremediation. Only 10 to 20% of the isolated, m-toluate tolerating bacterial strains were also able to degrade m-toluate. TOL plasmids were a major group of catabolic plasmids among these bacteria. The ability to degrade m-toluate by using enzymes encoded by a TOL plasmid was detected only in species of the genus Pseudomonas, and the best m-toluate degraders were these Pseudomonas species. Strain-specific differences in degradation abilities were found for P.oryzihabitans and P. migulae: some of these strains harbored a TOL plasmid - a new finding observed in this work, indicating putative horizontal plasmid transfer in the rhizosphere. One P. oryzihabitans strain harbored the pWW0 plasmid that had probably conjugated from the bioaugmentation Pseudomonas. Some P. migulae and P. oryzihabitans strains seemed to harbor both the pWW0- and the pDK1-type TOL plasmid. Alternatively, they might have harbored a TOL plasmid with both the pWW0- and the pDK1-type xylE gene. The breakdown of m-toluate by gram-negative bacteria was not restricted to the TOL pathway. Also some gram-positive Rhodococcus erythropolis and Arthrobacter aurescens strains were able to degrade m-toluate in the absence of a TOL plasmid. Three aspects of the rhizosphere effect of G. orientalis were manifested in oil-contaminated soil in the field: 1) G. orientalis and Pseudomonas bioaugmentation increased the amount of rhizosphere bacteria. G. orientalis especially together with Pseudomonas bioaugmentation increased the numbers of m-toluate utilizing and catechol positive bacteria indicating an increase in degradation potential. 2) Also the bacterial diversity, when measured as the amount of ribotypes, was increased in the Galega rhizosphere with or without Pseudomonas bioaugmentation. However, the diversity of m-toluate utilizing bacteria did not significantly increase. At the community level, by using the 16S rRNA gene PCR-DGGE method, the highest diversity of species was also observed in vegetated soils compared with non-vegetated soils. Diversified communities may best guarantee the overall success in rhizoremediation by offering various genetic machineries for catabolic processes. 3) At the end of the experiment, no TOL plasmid could be detected by direct DNA analysis in soil treated with both G. orientalis and Pseudomonas. The detection limit for TOL plasmids was encountered indicating decreased amount of degradation plasmids and thus, the success of rhizoremediation. The use of G. orientalis for rhizoremediation is unique. In this thesis new information was obtained about the rhizosphere effect of Galega orientalis in BTEX contaminated soils. The molecular biomonitoring methods can be applied for several purposes within environmental biotechnology, such as for evaluating the intrinsic biodegradation potential, monitoring the enhanced bioremediation, and estimating the success of bioremediation. Environmental protection by using nature's own resources and thus, acting according to the principle of sustainable development, would be both economically and environmentally beneficial for society. Keywords: molecular biomonitoring, genetic fingerprinting, soil bacteria, bacterial diversity, TOL plasmid, catabolic genes, horizontal gene transfer, rhizoremediation, rhizosphere effect, Galega orientalis, aerobic biodegradation, petroleum hydrocarbons, BTEX

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The present study focuses on the translational strategies of Cocksfoot mottle virus (CfMV, genus Sobemovirus), which infects monocotyledonous plants. CfMV RNA lacks the 5'cap and the 3'poly(A) tail that ensure efficient translation of cellular messenger RNAs (mRNAs). Instead, CfMV RNA is covalently linked to a viral protein VPg (viral protein, genome-linked). This indicates that the viral untranslated regions (UTRs) must functionally compensate for the lack of the cap and poly(A) tail. We examined the efficacy of translation initiation in CfMV by comparing it to well-studied viral translational enhancers. Although insertion of the CfMV 5'UTR (CfMVe) into plant expression vectors improved gene expression in barley more than the other translational enhancers examined, studies at the RNA level showed that CfMVe alone or in combination with the CfMV 3'UTR did not provide the RNAs translational advantage. Mutation analysis revealed that translation initiation from CfMVe involved scanning. Interestingly, CfMVe also promoted translation initiation from an intercistronic position of dicistronic mRNAs in vitro. Furthermore, internal initiation occurred with similar efficacy in translation lysates that had reduced concentrations of eukaryotic initiation factor (eIF) 4E, suggesting that initiation was independent of the eIF4E. In contrast, reduced translation in the eIF4G-depleted lysates indicated that translation from internally positioned CfMVe was eIF4G-dependent. After successful translation initiation, leaky scanning brings the ribosomes to the second open reading frame (ORF). The CfMV polyprotein is produced from this and the following overlapping ORF via programmed -1 ribosomal frameshift (-1 PRF). Two signals in the mRNA at the beginning of the overlap program approximately every fifth ribosome to slip one nucleotide backwards and continue translation in the new -1 frame. This leads to the production of C-terminally extended polyprotein, which encodes the viral RNA-dependent RNA polymerase (RdRp). The -1 PRF event in CfMV was very efficient, even though it was programmed by a simple stem-loop structure instead of a pseudoknot, which is usually required for high -1 PRF frequencies. Interestingly, regions surrounding the -1 PRF signals improved the -1 PRF frequencies. Viral protein P27 inhibited the -1 PRF event in vivo, putatively by binding to the -1 PRF site. This suggested that P27 could regulate the occurrence of -1 PRF. Initiation of viral replication requires that viral proteins are released from the polyprotein. This is catalyzed by viral serine protease, which is also encoded from the polyprotein. N-terminal amino acid sequencing of CfMV VPg revealed that the junction of the protease and VPg was cleaved between glutamate (E) and asparagine (N) residues. This suggested that the processing sites used in CfMV differ from the glutamate and serine (S) or threonine (T) sites utilized in other sobemoviruses. However, further analysis revealed that the E/S and E/T sites may be used to cleave out some of the CfMV proteins.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The purpose of this study is to describe the development of application of mass spectrometry for the structural analyses of non-coding ribonucleic acids during past decade. Mass spectrometric methods are compared of traditional gel electrophoretic methods, the characteristics of performance of mass spectrometric, analyses are studied and the future trends of mass spectrometry of ribonucleic acids are discussed. Non-coding ribonucleic acids are short polymeric biomolecules which are not translated to proteins, but which may affect the gene expression in all organisms. Regulatory ribonucleic acids act through transient interactions with key molecules in signal transduction pathways. Interactions are mediated through specific secondary and tertiary structures. Posttranscriptional modifications in the structures of molecules may introduce new properties to the organism, such as adaptation to environmental changes or development of resistance to antibiotics. In the scope of this study, the structural studies include i) determination of the sequence of nucleobases in the polymer chain, ii) characterisation and localisation of posttranscriptional modifications in nucleobases and in the backbone structure, iii) identification of ribonucleic acid-binding molecules and iv) probing of higher order structures in the ribonucleic acid molecule. Bacteria, archaea, viruses and HeLa cancer cells have been used as target organisms. Synthesised ribonucleic acids consisting of structural regions of interest have been frequently used. Electrospray ionisation (ESI) and matrix-assisted laser desorption ionisation (MALDI) have been used for ionisation of ribonucleic analytes. Ammonium acetate and 2-propanol are common solvents for ESI. Trihydroxyacetophenone is the optimal MALDI matrix for ionisation of ribonucleic acids and peptides. Ammonium salts are used in ESI buffers and MALDI matrices as additives to remove cation adducts. Reverse phase high performance liquid chromatography has been used for desalting and fractionation of analytes either off-line of on-line, coupled with ESI source. Triethylamine and triethylammonium bicarbonate are used as ion pair reagents almost exclusively. Fourier transform ion cyclotron resonance analyser using ESI coupled with liquid chromatography is the platform of choice for all forms of structural analyses. Time-of-flight (TOF) analyser using MALDI may offer sensitive, easy-to-use and economical solution for simple sequencing of longer oligonucleotides and analyses of analyte mixtures without prior fractionation. Special analysis software is used for computer-aided interpretation of mass spectra. With mass spectrometry, sequences of 20-30 nucleotides of length may be determined unambiguously. Sequencing may be applied to quality control of short synthetic oligomers for analytical purposes. Sequencing in conjunction with other structural studies enables accurate localisation and characterisation of posttranscriptional modifications and identification of nucleobases and amino acids at the sites of interaction. High throughput screening methods for RNA-binding ligands have been developed. Probing of the higher order structures has provided supportive data for computer-generated three dimensional models of viral pseudoknots. In conclusion. mass spectrometric methods are well suited for structural analyses of small species of ribonucleic acids, such as short non-coding ribonucleic acids in the molecular size region of 20-30 nucleotides. Structural information not attainable with other methods of analyses, such as nuclear magnetic resonance and X-ray crystallography, may be obtained with the use of mass spectrometry. Sequencing may be applied to quality control of short synthetic oligomers for analytical purposes. Ligand screening may be used in the search of possible new therapeutic agents. Demanding assay design and challenging interpretation of data requires multidisclipinary knowledge. The implement of mass spectrometry to structural studies of ribonucleic acids is probably most efficiently conducted in specialist groups consisting of researchers from various fields of science.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The present study examines citizen participation in local government and municipal democracy. Previous research has shown that the prerequisite for active citizenship lies in the opportunities available for local residents to determine which perspectives and planning needs are relevant. This research looks at whether the conception of knowledge employed in municipal planning allows for this kind of active role for local citizens. Methodologically the study employs an hermeneutic approach. The aim has been to identify various approaches steering the practice of municipal democracy. The theory behind the study comes from the assumption of the intersubjectivity of reality. Construing the rationality of one s own behaviour is seen as a prerequisite for meaningful action. In this context, criteria for the functionality of municipal democracy and the purpose of strengthening citizen participation are defined. The study is divided into two parts. Firstly, the purpose of participation and the opportunities for local residents to contribute is examined theoretically with reference to previous studies. The intention is to provide an overview of the Finnish cross-disciplinary debate on resident participation. This debate is reflected onto the prevailing views on changes in the municipal operating environment and modes of operation. In conclusion, a theoretical model is constructed to explain how the various modes of operation in regional municipalities affect the purpose of resident participation and the utilisation of information received through this participation. The second part of the study discusses the utilisation of this information and knowledge acquired through the participation of local residents and all those involved in political and administrative processes in municipalities. These first-hand reports are analysed using the model constructed earlier in the study. The goal is to understand how political and administrative practice affects opportunities for local residents to participate and contribute. The core of this analysis is based on the pragmatic conception of knowledge employed in municipal administration. The study argues that the normal practice of municipal administration does not support the systematic utilisation of local residents experience. This is caused by two interlinked factors: firstly, knowledge constructed through these practices requires that the knowledge is apolitical; secondly, arising from this there is confusion with regard to when during a planning process does information obtained from the public become relevant; in other words, what are the politics of knowledge? The study suggests that the solution is in the complementary concept of knowledge, which implicitly acknowledges the politics of knowledge. The complementary concept of knowledge would serve the politicisation of issues on the level of interpretations linked with social reality, an indispensable requirement for functional democracy. Keywords: participation, municipal democracy, knowledge base for planning, experiential knowledge

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Many problems in analysis have been solved using the theory of Hodge structures. P. Deligne started to treat these structures in a categorical way. Following him, we introduce the categories of mixed real and complex Hodge structures. Category of mixed Hodge structures over the field of real or complex numbers is a rigid abelian tensor category, and in fact, a neutral Tannakian category. Therefore it is equivalent to the category of representations of an affine group scheme. The direct sums of pure Hodge structures of different weights over real or complex numbers can be realized as a representation of the torus group, whose complex points is the Cartesian product of two punctured complex planes. Mixed Hodge structures turn out to consist of information of a direct sum of pure Hodge structures of different weights and a nilpotent automorphism. Therefore mixed Hodge structures correspond to the representations of certain semidirect product of a nilpotent group and the torus group acting on it.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Metabolism is the cellular subsystem responsible for generation of energy from nutrients and production of building blocks for larger macromolecules. Computational and statistical modeling of metabolism is vital to many disciplines including bioengineering, the study of diseases, drug target identification, and understanding the evolution of metabolism. In this thesis, we propose efficient computational methods for metabolic modeling. The techniques presented are targeted particularly at the analysis of large metabolic models encompassing the whole metabolism of one or several organisms. We concentrate on three major themes of metabolic modeling: metabolic pathway analysis, metabolic reconstruction and the study of evolution of metabolism. In the first part of this thesis, we study metabolic pathway analysis. We propose a novel modeling framework called gapless modeling to study biochemically viable metabolic networks and pathways. In addition, we investigate the utilization of atom-level information on metabolism to improve the quality of pathway analyses. We describe efficient algorithms for discovering both gapless and atom-level metabolic pathways, and conduct experiments with large-scale metabolic networks. The presented gapless approach offers a compromise in terms of complexity and feasibility between the previous graph-theoretic and stoichiometric approaches to metabolic modeling. Gapless pathway analysis shows that microbial metabolic networks are not as robust to random damage as suggested by previous studies. Furthermore the amino acid biosynthesis pathways of the fungal species Trichoderma reesei discovered from atom-level data are shown to closely correspond to those of Saccharomyces cerevisiae. In the second part, we propose computational methods for metabolic reconstruction in the gapless modeling framework. We study the task of reconstructing a metabolic network that does not suffer from connectivity problems. Such problems often limit the usability of reconstructed models, and typically require a significant amount of manual postprocessing. We formulate gapless metabolic reconstruction as an optimization problem and propose an efficient divide-and-conquer strategy to solve it with real-world instances. We also describe computational techniques for solving problems stemming from ambiguities in metabolite naming. These techniques have been implemented in a web-based sofware ReMatch intended for reconstruction of models for 13C metabolic flux analysis. In the third part, we extend our scope from single to multiple metabolic networks and propose an algorithm for inferring gapless metabolic networks of ancestral species from phylogenetic data. Experimenting with 16 fungal species, we show that the method is able to generate results that are easily interpretable and that provide hypotheses about the evolution of metabolism.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Event-based systems are seen as good candidates for supporting distributed applications in dynamic and ubiquitous environments because they support decoupled and asynchronous many-to-many information dissemination. Event systems are widely used, because asynchronous messaging provides a flexible alternative to RPC (Remote Procedure Call). They are typically implemented using an overlay network of routers. A content-based router forwards event messages based on filters that are installed by subscribers and other routers. The filters are organized into a routing table in order to forward incoming events to proper subscribers and neighbouring routers. This thesis addresses the optimization of content-based routing tables organized using the covering relation and presents novel data structures and configurations for improving local and distributed operation. Data structures are needed for organizing filters into a routing table that supports efficient matching and runtime operation. We present novel results on dynamic filter merging and the integration of filter merging with content-based routing tables. In addition, the thesis examines the cost of client mobility using different protocols and routing topologies. We also present a new matching technique called temporal subspace matching. The technique combines two new features. The first feature, temporal operation, supports notifications, or content profiles, that persist in time. The second feature, subspace matching, allows more expressive semantics, because notifications may contain intervals and be defined as subspaces of the content space. We also present an application of temporal subspace matching pertaining to metadata-based continuous collection and object tracking.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The ever expanding growth of the wireless access to the Internet in recent years has led to the proliferation of wireless and mobile devices to connect to the Internet. This has created the possibility of mobile devices equipped with multiple radio interfaces to connect to the Internet using any of several wireless access network technologies such as GPRS, WLAN and WiMAX in order to get the connectivity best suited for the application. These access networks are highly heterogeneous and they vary widely in their characteristics such as bandwidth, propagation delay and geographical coverage. The mechanism by which a mobile device switches between these access networks during an ongoing connection is referred to as vertical handoff and it often results in an abrupt and significant change in the access link characteristics. The most common Internet applications such as Web browsing and e-mail make use of the Transmission Control Protocol (TCP) as their transport protocol and the behaviour of TCP depends on the end-to-end path characteristics such as bandwidth and round-trip time (RTT). As the wireless access link is most likely the bottleneck of a TCP end-to-end path, the abrupt changes in the link characteristics due to a vertical handoff may affect TCP behaviour adversely degrading the performance of the application. The focus of this thesis is to study the effect of a vertical handoff on TCP behaviour and to propose algorithms that improve the handoff behaviour of TCP using cross-layer information about the changes in the access link characteristics. We begin this study by identifying the various problems of TCP due to a vertical handoff based on extensive simulation experiments. We use this study as a basis to develop cross-layer assisted TCP algorithms in handoff scenarios involving GPRS and WLAN access networks. We then extend the scope of the study by developing cross-layer assisted TCP algorithms in a broader context applicable to a wide range of bandwidth and delay changes during a handoff. And finally, the algorithms developed here are shown to be easily extendable to the multiple-TCP flow scenario. We evaluate the proposed algorithms by comparison with standard TCP (TCP SACK) and show that the proposed algorithms are effective in improving TCP behavior in vertical handoff involving a wide range of bandwidth and delay of the access networks. Our algorithms are easy to implement in real systems and they involve modifications to the TCP sender algorithm only. The proposed algorithms are conservative in nature and they do not adversely affect the performance of TCP in the absence of cross-layer information.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Segmentation is a data mining technique yielding simplified representations of sequences of ordered points. A sequence is divided into some number of homogeneous blocks, and all points within a segment are described by a single value. The focus in this thesis is on piecewise-constant segments, where the most likely description for each segment and the most likely segmentation into some number of blocks can be computed efficiently. Representing sequences as segmentations is useful in, e.g., storage and indexing tasks in sequence databases, and segmentation can be used as a tool in learning about the structure of a given sequence. The discussion in this thesis begins with basic questions related to segmentation analysis, such as choosing the number of segments, and evaluating the obtained segmentations. Standard model selection techniques are shown to perform well for the sequence segmentation task. Segmentation evaluation is proposed with respect to a known segmentation structure. Applying segmentation on certain features of a sequence is shown to yield segmentations that are significantly close to the known underlying structure. Two extensions to the basic segmentation framework are introduced: unimodal segmentation and basis segmentation. The former is concerned with segmentations where the segment descriptions first increase and then decrease, and the latter with the interplay between different dimensions and segments in the sequence. These problems are formally defined and algorithms for solving them are provided and analyzed. Practical applications for segmentation techniques include time series and data stream analysis, text analysis, and biological sequence analysis. In this thesis segmentation applications are demonstrated in analyzing genomic sequences.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

XML documents are becoming more and more common in various environments. In particular, enterprise-scale document management is commonly centred around XML, and desktop applications as well as online document collections are soon to follow. The growing number of XML documents increases the importance of appropriate indexing methods and search tools in keeping the information accessible. Therefore, we focus on content that is stored in XML format as we develop such indexing methods. Because XML is used for different kinds of content ranging all the way from records of data fields to narrative full-texts, the methods for Information Retrieval are facing a new challenge in identifying which content is subject to data queries and which should be indexed for full-text search. In response to this challenge, we analyse the relation of character content and XML tags in XML documents in order to separate the full-text from data. As a result, we are able to both reduce the size of the index by 5-6\% and improve the retrieval precision as we select the XML fragments to be indexed. Besides being challenging, XML comes with many unexplored opportunities which are not paid much attention in the literature. For example, authors often tag the content they want to emphasise by using a typeface that stands out. The tagged content constitutes phrases that are descriptive of the content and useful for full-text search. They are simple to detect in XML documents, but also possible to confuse with other inline-level text. Nonetheless, the search results seem to improve when the detected phrases are given additional weight in the index. Similar improvements are reported when related content is associated with the indexed full-text including titles, captions, and references. Experimental results show that for certain types of document collections, at least, the proposed methods help us find the relevant answers. Even when we know nothing about the document structure but the XML syntax, we are able to take advantage of the XML structure when the content is indexed for full-text search.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this thesis we present and evaluate two pattern matching based methods for answer extraction in textual question answering systems. A textual question answering system is a system that seeks answers to natural language questions from unstructured text. Textual question answering systems are an important research problem because as the amount of natural language text in digital format grows all the time, the need for novel methods for pinpointing important knowledge from the vast textual databases becomes more and more urgent. We concentrate on developing methods for the automatic creation of answer extraction patterns. A new type of extraction pattern is developed also. The pattern matching based approach chosen is interesting because of its language and application independence. The answer extraction methods are developed in the framework of our own question answering system. Publicly available datasets in English are used as training and evaluation data for the methods. The techniques developed are based on the well known methods of sequence alignment and hierarchical clustering. The similarity metric used is based on edit distance. The main conclusions of the research are that answer extraction patterns consisting of the most important words of the question and of the following information extracted from the answer context: plain words, part-of-speech tags, punctuation marks and capitalization patterns, can be used in the answer extraction module of a question answering system. This type of patterns and the two new methods for generating answer extraction patterns provide average results when compared to those produced by other systems using the same dataset. However, most answer extraction methods in the question answering systems tested with the same dataset are both hand crafted and based on a system-specific and fine-grained question classification. The the new methods developed in this thesis require no manual creation of answer extraction patterns. As a source of knowledge, they require a dataset of sample questions and answers, as well as a set of text documents that contain answers to most of the questions. The question classification used in the training data is a standard one and provided already in the publicly available data.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This thesis studies human gene expression space using high throughput gene expression data from DNA microarrays. In molecular biology, high throughput techniques allow numerical measurements of expression of tens of thousands of genes simultaneously. In a single study, this data is traditionally obtained from a limited number of sample types with a small number of replicates. For organism-wide analysis, this data has been largely unavailable and the global structure of human transcriptome has remained unknown. This thesis introduces a human transcriptome map of different biological entities and analysis of its general structure. The map is constructed from gene expression data from the two largest public microarray data repositories, GEO and ArrayExpress. The creation of this map contributed to the development of ArrayExpress by identifying and retrofitting the previously unusable and missing data and by improving the access to its data. It also contributed to creation of several new tools for microarray data manipulation and establishment of data exchange between GEO and ArrayExpress. The data integration for the global map required creation of a new large ontology of human cell types, disease states, organism parts and cell lines. The ontology was used in a new text mining and decision tree based method for automatic conversion of human readable free text microarray data annotations into categorised format. The data comparability and minimisation of the systematic measurement errors that are characteristic to each lab- oratory in this large cross-laboratories integrated dataset, was ensured by computation of a range of microarray data quality metrics and exclusion of incomparable data. The structure of a global map of human gene expression was then explored by principal component analysis and hierarchical clustering using heuristics and help from another purpose built sample ontology. A preface and motivation to the construction and analysis of a global map of human gene expression is given by analysis of two microarray datasets of human malignant melanoma. The analysis of these sets incorporate indirect comparison of statistical methods for finding differentially expressed genes and point to the need to study gene expression on a global level.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Matrix decompositions, where a given matrix is represented as a product of two other matrices, are regularly used in data mining. Most matrix decompositions have their roots in linear algebra, but the needs of data mining are not always those of linear algebra. In data mining one needs to have results that are interpretable -- and what is considered interpretable in data mining can be very different to what is considered interpretable in linear algebra. --- The purpose of this thesis is to study matrix decompositions that directly address the issue of interpretability. An example is a decomposition of binary matrices where the factor matrices are assumed to be binary and the matrix multiplication is Boolean. The restriction to binary factor matrices increases interpretability -- factor matrices are of the same type as the original matrix -- and allows the use of Boolean matrix multiplication, which is often more intuitive than normal matrix multiplication with binary matrices. Also several other decomposition methods are described, and the computational complexity of computing them is studied together with the hardness of approximating the related optimization problems. Based on these studies, algorithms for constructing the decompositions are proposed. Constructing the decompositions turns out to be computationally hard, and the proposed algorithms are mostly based on various heuristics. Nevertheless, the algorithms are shown to be capable of finding good results in empirical experiments conducted with both synthetic and real-world data.