926 resultados para Genomic data integration


Relevância:

40.00% 40.00%

Publicador:

Resumo:

Although the citriculture is one of the most important economic activities in Brazil, it is based on a small number of varieties. This fact has contributed for the vulnerability of the culture regarding the phytosanitary problems. A higher number of varieties/genotypes with potential for commercial growing, either for the industry or fresh market, has been one of the main objectives of citrus breeding programs. The genetic breeding of citrus has improved, in the last decades, due to the possibility of an association between biotechnological tools and classical methods of breeding. The use of molecular markers for early selection of zygotic seedlings from controlled crosses resulted in the possibility of selection of a high number of new combination and, as a consequence, the establishment of a great number of hybrids in field experiments. The faster new tools are incorporated in the program, the faster is possibility to reach new genotypes that can be tested as a new variety. Good traits should be kept or incorporate, whereas bad traits have to be excluded or minimized in the new genotype. Scion and rootstock can not be considered separately, and graft compatibility, fruit quality and productivity are essential traits to be evaluated in the last stages of the program. The mapping of QTLs has favored breeding programs of several perennial species and in citrus it was possible to map several characteristics with qualitative and quantitative inheritance. The existence of linkage maps and QTLs already mapped, the development of EST and BAC library and the sequencing of the Citrus complete genome altogether make very demanding and urgent the exploration of such data to launch a wider genetic study of citrus. The rising of information on genome of several organisms has opened new approaches looking for integration between breeding, genetic and genome. Genome assisted selection (GAS) involves more than gene or complete genome sequencing and is becoming an import support in breeding programs of annual and perennial species. An huge information amount can be derivate from genome analysis. The use and benefit of such informations will depend on the genetic basis of the breeding program.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This thesis work describes the creation of a pipework data structure for design system integration. Work is completed in pulp and paper plant delivery company with global engineering network operations in mind. User case of process design to 3D pipework design is introduced with influence of subcontracting engineering offices. Company data element list is gathered by using key person interviews and results are processed into a pipework data element list. Inter-company co-operation is completed in standardization association and common standard for pipework data elements is found. As result inter-company created pipework data element list is introduced. Further list usage, development and relations to design software vendors are evaluated.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The aim of this master’s thesis is to provide a real life example of how marketing research data is used by different functions in the NPD process. In order to achieve this goal, a case study in a company was implemented where gathering, analysis, distribution and synthesis of marketing research data in NPD were studied. The main research question was formulated as follows: How is marketing research data integrated and used by different company functions in the NPD process? The theory part of the master’s thesis was focused on the discussion of the marketing function role in NPD, use of marketing research particularly in the food industry, as well as issues related to the marketing/R&D interface during the NPD process. The empirical part of the master’s thesis was based on qualitative explanatory case study research. Individual in-depth interviews with company representatives, company documents and online research were used for data collection and analyzed through triangulation method. The empirical findings advocate that the most important marketing data sources at the concept generation stage of NPD are: global trends monitoring, retailing audit and consumers insights. These data sets are crucial for establishing the potential of the product on the market and defining the desired features for the new product to be developed. The findings also suggest the example of successful crossfunctional communication during the NPD process with formal and informal communication patterns. General managerial recommendations are given on the integration in NPD of a strategy, process, continuous improvement, and motivated cross-functional product development teams.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

There has been a clear lack of common data exchange semantics for inter-organisational workflow management systems where the research has mainly focused on technical issues rather than language constructs. This paper presents the neutral data exchanges semantics required for the workflow integration within the AXAEDIS framework and presents the mechanism for object discovery from the object repository where little or no knowledge about the object is available. The paper also presents workflow independent integration architecture with the AXAEDIS Framework.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper describes the implementation of a 3D variational (3D-Var) data assimilation scheme for a morphodynamic model applied to Morecambe Bay, UK. A simple decoupled hydrodynamic and sediment transport model is combined with a data assimilation scheme to investigate the ability of such methods to improve the accuracy of the predicted bathymetry. The inverse forecast error covariance matrix is modelled using a Laplacian approximation which is calibrated for the length scale parameter required. Calibration is also performed for the Soulsby-van Rijn sediment transport equations. The data used for assimilation purposes comprises waterlines derived from SAR imagery covering the entire period of the model run, and swath bathymetry data collected by a ship-borne survey for one date towards the end of the model run. A LiDAR survey of the entire bay carried out in November 2005 is used for validation purposes. The comparison of the predictive ability of the model alone with the model-forecast-assimilation system demonstrates that using data assimilation significantly improves the forecast skill. An investigation of the assimilation of the swath bathymetry as well as the waterlines demonstrates that the overall improvement is initially large, but decreases over time as the bathymetry evolves away from that observed by the survey. The result of combining the calibration runs into a pseudo-ensemble provides a higher skill score than for a single optimized model run. A brief comparison of the Optimal Interpolation assimilation method with the 3D-Var method shows that the two schemes give similar results.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Brain activity can be measured with several non-invasive neuroimaging modalities, but each modality has inherent limitations with respect to resolution, contrast and interpretability. It is hoped that multimodal integration will address these limitations by using the complementary features of already available data. However, purely statistical integration can prove problematic owing to the disparate signal sources. As an alternative, we propose here an advanced neural population model implemented on an anatomically sound cortical mesh with freely adjustable connectivity, which features proper signal expression through a realistic head model for the electroencephalogram (EEG), as well as a haemodynamic model for functional magnetic resonance imaging based on blood oxygen level dependent contrast (fMRI BOLD). It hence allows simultaneous and realistic predictions of EEG and fMRI BOLD from the same underlying model of neural activity. As proof of principle, we investigate here the influence on simulated brain activity of strengthening visual connectivity. In the future we plan to fit multimodal data with this neural population model. This promises novel, model-based insights into the brain's activity in sleep, rest and task conditions.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Background: Affymetrix GeneChip arrays are widely used for transcriptomic studies in a diverse range of species. Each gene is represented on a GeneChip array by a probe- set, consisting of up to 16 probe-pairs. Signal intensities across probe- pairs within a probe-set vary in part due to different physical hybridisation characteristics of individual probes with their target labelled transcripts. We have previously developed a technique to study the transcriptomes of heterologous species based on hybridising genomic DNA (gDNA) to a GeneChip array designed for a different species, and subsequently using only those probes with good homology. Results: Here we have investigated the effects of hybridising homologous species gDNA to study the transcriptomes of species for which the arrays have been designed. Genomic DNA from Arabidopsis thaliana and rice (Oryza sativa) were hybridised to the Affymetrix Arabidopsis ATH1 and Rice Genome GeneChip arrays respectively. Probe selection based on gDNA hybridisation intensity increased the number of genes identified as significantly differentially expressed in two published studies of Arabidopsis development, and optimised the analysis of technical replicates obtained from pooled samples of RNA from rice. Conclusion: This mixed physical and bioinformatics approach can be used to optimise estimates of gene expression when using GeneChip arrays.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The past years have shown an enormous advancement in sequencing and array-based technologies, producing supplementary or alternative views of the genome stored in various formats and databases. Their sheer volume and different data scope pose a challenge to jointly visualize and integrate diverse data types. We present AmalgamScope a new interactive software tool focusing on assisting scientists with the annotation of the human genome and particularly the integration of the annotation files from multiple data types, using gene identifiers and genomic coordinates. Supported platforms include next-generation sequencing and microarray technologies. The available features of AmalgamScope range from the annotation of diverse data types across the human genome to integration of the data based on the annotational information and visualization of the merged files within chromosomal regions or the whole genome. Additionally, users can define custom transcriptome library files for any species and use the file exchanging distant server options of the tool.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

With the aim of determining the genetic basis of metabolic regulation in tomato fruit, we constructed a detailed physical map of genomic regions spanning previously described metabolic quantitative trait loci of a Solanum pennellii introgression line population. Two genomic libraries from S. pennellii were screened with 104 colocated markers from five selected genomic regions, and a total of 614 bacterial artificial chromosome (BAC)/cosmids were identified as seed clones. Integration of sequence data with the genetic and physical maps of Solanum lycopersicum facilitated the anchoring of 374 of these BAC/cosmid clones. The analysis of this information resulted in a genome-wide map of a nondomesticated plant species and covers 10% of the physical distance of the selected regions corresponding to approximately 1% of the wild tomato genome. Comparative analyses revealed that S. pennellii and domesticated tomato genomes can be considered as largely colinear. A total of 1,238,705 bp from both BAC/cosmid ends and nine large insert clones were sequenced, annotated, and functionally categorized. The sequence data allowed the evaluation of the level of polymorphism between the wild and cultivated tomato species. An exhaustive microsynteny analysis allowed us to estimate the divergence date of S. pennellii and S. lycopersicum at 2.7 million years ago. The combined results serve as a reference for comparative studies both at the macrosyntenic and microsyntenic levels. They also provide a valuable tool for fine-mapping of quantitative trait loci in tomato. Furthermore, they will contribute to a deeper understanding of the regulatory factors underpinning metabolism and hence defining crop chemical composition.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Online geographic-databases have been growing increasingly as they have become a crucial source of information for both social networks and safety-critical systems. Since the quality of such applications is largely related to the richness and completeness of their data, it becomes imperative to develop adaptable and persistent storage systems, able to make use of several sources of information as well as enabling the fastest possible response from them. This work will create a shared and extensible geographic model, able to retrieve and store information from the major spatial sources available. A geographic-based system also has very high requirements in terms of scalability, computational power and domain complexity, causing several difficulties for a traditional relational database as the number of results increases. NoSQL systems provide valuable advantages for this scenario, in particular graph databases which are capable of modeling vast amounts of inter-connected data while providing a very substantial increase of performance for several spatial requests, such as finding shortestpath routes and performing relationship lookups with high concurrency. In this work, we will analyze the current state of geographic information systems and develop a unified geographic model, named GeoPlace Explorer (GE). GE is able to import and store spatial data from several online sources at a symbolic level in both a relational and a graph databases, where several stress tests were performed in order to find the advantages and disadvantages of each database paradigm.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Abstract Background The study and analysis of gene expression measurements is the primary focus of functional genomics. Once expression data is available, biologists are faced with the task of extracting (new) knowledge associated to the underlying biological phenomenon. Most often, in order to perform this task, biologists execute a number of analysis activities on the available gene expression dataset rather than a single analysis activity. The integration of heteregeneous tools and data sources to create an integrated analysis environment represents a challenging and error-prone task. Semantic integration enables the assignment of unambiguous meanings to data shared among different applications in an integrated environment, allowing the exchange of data in a semantically consistent and meaningful way. This work aims at developing an ontology-based methodology for the semantic integration of gene expression analysis tools and data sources. The proposed methodology relies on software connectors to support not only the access to heterogeneous data sources but also the definition of transformation rules on exchanged data. Results We have studied the different challenges involved in the integration of computer systems and the role software connectors play in this task. We have also studied a number of gene expression technologies, analysis tools and related ontologies in order to devise basic integration scenarios and propose a reference ontology for the gene expression domain. Then, we have defined a number of activities and associated guidelines to prescribe how the development of connectors should be carried out. Finally, we have applied the proposed methodology in the construction of three different integration scenarios involving the use of different tools for the analysis of different types of gene expression data. Conclusions The proposed methodology facilitates the development of connectors capable of semantically integrating different gene expression analysis tools and data sources. The methodology can be used in the development of connectors supporting both simple and nontrivial processing requirements, thus assuring accurate data exchange and information interpretation from exchanged data.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Pig meat and carcass quality is a complex concept determined by environmental and genetic factors concurring to the phenotypic variation in qualitative characteristics of meat (fat content, tenderness, juiciness, flavor,etc). This thesis shows the results of different investigations to study and to analyze pig meat and carcass quality focusing mainly on genomic; moreover proteomic approach has been also used. The aim was to analyze data from association studies between genes considered as candidate and meat and carcass quality in different pig breeds. The approach was used to detect new SNP in genes functionally associated to the studied traits and to confirm as candidate other genes already known. Five polymorphisms (one new SNP in Calponin 1 gene and four additional polymorphism already known in other genes) were considered on chromosome 2 (SSC2). Calponin 1 (CNN1) was associated to the studied traits and furthermore the results reported confirmed the data already known for Lactate dehydrogenase A (LDHA), Low density lipoprotein receptor (LDLR), Myogenic differentiation 1 (MYOD1) e Ubiquitin-like 5 (UBL5), in Italian Large White pigs. Using an in silico search it was possible to detect on SSC2 a new SNP of Deoxyhypusine synthase (DHPS) gene partially overlapping with WD repeat domain 83 (WDR83) gene and significant for the meat pH variation in Italian Large White (ILW) pigs. Perilipin 1 (PLIN1) mapping on chromosome 7 and Perilipin 2 (PLIN2) mapping on chromosome 1 were studied and the results obtained in Duroc breed have shown significant associations with carcass traits. Moreover a study of protein composition of porcine LD muscle, indicated an effect of temperature treatment of carcass, on proteins of the sarcoplasmic fraction and in particular on PGM1 phosphorylation. Future studies on pig meat quality should be based on the integration of different experimental approaches (genomics, proteomics, transcriptomics, etc).