888 resultados para Controlled vocabulary


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Expression data contribute significantly to the biological value of the sequenced human genome, providing extensive information about gene structure and the pattern of gene expression. ESTs, together with SAGE libraries and microarray experiment information, provide a broad and rich view of the transcriptome. However, it is difficult to perform large-scale expression mining of the data generated by these diverse experimental approaches. Not only is the data stored in disparate locations, but there is frequent ambiguity in the meaning of terms used to describe the source of the material used in the experiment. Untangling semantic differences between the data provided by different resources is therefore largely reliant on the domain knowledge of a human expert. We present here eVOC, a system which associates labelled target cDNAs for microarray experiments, or cDNA libraries and their associated transcripts with controlled terms in a set of hierarchical vocabularies. eVOC consists of four orthogonal controlled vocabularies suitable for describing the domains of human gene expression data including Anatomical System, Cell Type, Pathology and Developmental Stage. We have curated and annotated 7016 cDNA libraries represented in dbEST, as well as 104 SAGE libraries,with expression information,and provide this as an integrated, public resource that allows the linking of transcripts and libraries with expression terms. Both the vocabularies and the vocabulary-annotated libraries can be retrieved from http://www.sanbi.ac.za/evoc/. Several groups are involved in developing this resource with the aim of unifying transcript expression information.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Using free text and controlled vocabulary in Medline and CINAHL

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Introduction: the statistical record used in the Field Academic Programs (PAC for it’s initials in Spanish) of Rehabilitation denotes generalities in the data conceptualization, which complicates the reliable guidance in making decisions and provides a low support for research in rehabilitation and disability. In response, the Research Group in Rehabilitation and Social Integration of Persons with Disabilities has worked on the creation of a registry to characterize the population seen by Rehabilitation PAC. This registry includes the use of the International Classification of Functioning, Disability and Health (ICF) of the WHO. Methodology: the proposed methodology includes two phases: the first one is a descriptive study and the second one involves performing methodology Methontology, which integrates the identification and development of ontology knowledge. This article contextualizes the progress made in the second phase. Results: the development of the registry in 2008, as an information system, included documentary review and the analysis of possible use scenarios to help guide the design and development of the SIDUR system. The system uses the ICF given that it is a terminology standardization that allows the reduction of ambiguity and that makes easier the transformation of health facts into data translatable to information systems. The record raises three categories and a total of 129 variables Conclusions: SIDUR facilitates accessibility to accurate and updated information, useful for decision making and research.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

There are three key components for developing a metadata system: a container structure laying out the key semantic issues of interest and their relationships; an extensible controlled vocabulary providing possible content; and tools to create and manipulate that content. While metadata systems must allow users to enter their own information, the use of a controlled vocabulary both imposes consistency of definition and ensures comparability of the objects described. Here we describe the controlled vocabulary (CV) and metadata creation tool built by the METAFOR project for use in the context of describing the climate models, simulations and experiments of the fifth Coupled Model Intercomparison Project (CMIP5). The CV and resulting tool chain introduced here is designed for extensibility and reuse and should find applicability in many more projects.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Proceedings paper published by Society of American Archivists. Presented at conference in 2015 in Cleveland, OH (http://www2.archivists.org/proceedings/research-forum/2015/agenda#papers). Published by SAA in 2016.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper describes the process of creating a controlled vocabulary which can be used to systematically analyse the copyright transfer agreements (CTAs) of journal publishers with regard to self-archiving. The analysis formed the basis of the newly created Copyright Knowledge Bank of publishers’ self-archiving policies. Self-archiving terms appearing in publishers’ CTAs were identified and classified, with these then being simplified, merged, and discarded to form a definitive list. The controlled vocabulary consists of three categories that describe ‘what’ can be self-archived, the ‘conditions’ of self-archiving and the ‘restrictions’ of self-archiving. Condition terms include specifications such as ‘where’ an article can be self archived, restriction terms include specifications such as ‘when’ the article can be self archived. Additional information on any of these terms appears in ‘free-text’ fields. Although this controlled vocabulary provides an effective way of analysing CTAs, it will need to be continually reviewed and updated in light of any major new additions to the terms used in publishers’ copyright and self-archiving policies.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The World Wide Web (WWW) is useful for distributing scientific data. Most existing web data resources organize their information either in structured flat files or relational databases with basic retrieval capabilities. For databases with one or a few simple relations, these approaches are successful, but they can be cumbersome when there is a data model involving multiple relations between complex data. We believe that knowledge-based resources offer a solution in these cases. Knowledge bases have explicit declarations of the concepts in the domain, along with the relations between them. They are usually organized hierarchically, and provide a global data model with a controlled vocabulary, We have created the OWEB architecture for building online scientific data resources using knowledge bases. OWEB provides a shell for structuring data, providing secure and shared access, and creating computational modules for processing and displaying data. In this paper, we describe the translation of the online immunological database MHCPEP into an OWEB system called MHCWeb. This effort involved building a conceptual model for the data, creating a controlled terminology for the legal values for different types of data, and then translating the original data into the new structure. The 0 WEB environment allows for flexible access to the data by both users and computer programs.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Esse texto tem por objetivo apresentar os resultados parciais do projeto ???Mapeamento da Oferta de Capacita????o nas Escolas de Governo??? e debater as possibilidades que se abrem para a gest??o do conhecimento e para o gerenciamento das informa????es estrat??gicas visando ?? tomada de decis??es e a otimiza????o dos recursos dispon??veis para a capacita????o. O projeto mapeamento tem por objetivo coletar dados da oferta existente de capacita????o em escolas de governo brasileiras, em n??vel municipal, estadual e federal. Foi concebido em tr??s fases: coleta das informa????es cadastrais e sobre a disponibilidade de cursos; constru????o do vocabul??rio controlado e classifica????o dos cursos; e transposi????o dessas informa????es para um sistema de informa????es com interface web: o Sistema Mapa. O Sistema Mapa, se efetivamente incorporado e utilizado, poder?? constituir um sistema de informa????o que auxilie os gestores e dirigentes a planejar e realizar planos de capacita????o, utilizando de forma eficiente a oferta existente nas escolas de governo. Al??m disso, poder?? contribuir para melhor coordena????o das atividades de capacita????o, incentivo ?? complementaridade, redu????o de duplicidades e incentivo ?? amplia????o da comunica????o entre as institui????es da Rede.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Based in internet growth, through semantic web, together with communication speed improvement and fast development of storage device sizes, data and information volume rises considerably every day. Because of this, in the last few years there has been a growing interest in structures for formal representation with suitable characteristics, such as the possibility to organize data and information, as well as the reuse of its contents aimed for the generation of new knowledge. Controlled Vocabulary, specifically Ontologies, present themselves in the lead as one of such structures of representation with high potential. Not only allow for data representation, as well as the reuse of such data for knowledge extraction, coupled with its subsequent storage through not so complex formalisms. However, for the purpose of assuring that ontology knowledge is always up to date, they need maintenance. Ontology Learning is an area which studies the details of update and maintenance of ontologies. It is worth noting that relevant literature already presents first results on automatic maintenance of ontologies, but still in a very early stage. Human-based processes are still the current way to update and maintain an ontology, which turns this into a cumbersome task. The generation of new knowledge aimed for ontology growth can be done based in Data Mining techniques, which is an area that studies techniques for data processing, pattern discovery and knowledge extraction in IT systems. This work aims at proposing a novel semi-automatic method for knowledge extraction from unstructured data sources, using Data Mining techniques, namely through pattern discovery, focused in improving the precision of concept and its semantic relations present in an ontology. In order to verify the applicability of the proposed method, a proof of concept was developed, presenting its results, which were applied in building and construction sector.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The Smart Drug Search is publicly accessible at http://sing.ei.uvigo.es/sds/. The BIOMedical Search Engine Framework is freely available for non-commercial use at https://github.com/agjacome/biomsef

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The Andalusian Public Health System (Sistema Sanitario Público de Andalucía -SSPA) Repository is the open environment where all the scientific output generated by the SSPA professionals, resulting from their medical care, research and administrative activities, is comprehensively collected and managed. This repository possesses special features which determined its development: the SSPA organization and its purpose as a health institution, the specific sets of documents that it generates and the stakeholders involved in it. The repository uses DSpace 1.6.2, to which several changes were implemented in order to achieve the SSPA initial goals and requirements. The main changes were: the addition of specific qualifiers to the Metadata Dublin Core scheme, the modification of the submission form, the integration of the MeSH Thesaurus as controlled vocabulary and the optimization of the advanced search tool. Another key point during the setting up of the repository was the initial batch ingest of the documents.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

During the last few years, next-generation sequencing (NGS) technologies have accelerated the detection of genetic variants resulting in the rapid discovery of new disease-associated genes. However, the wealth of variation data made available by NGS alone is not sufficient to understand the mechanisms underlying disease pathogenesis and manifestation. Multidisciplinary approaches combining sequence and clinical data with prior biological knowledge are needed to unravel the role of genetic variants in human health and disease. In this context, it is crucial that these data are linked, organized, and made readily available through reliable online resources. The Swiss-Prot section of the Universal Protein Knowledgebase (UniProtKB/Swiss-Prot) provides the scientific community with a collection of information on protein functions, interactions, biological pathways, as well as human genetic diseases and variants, all manually reviewed by experts. In this article, we present an overview of the information content of UniProtKB/Swiss-Prot to show how this knowledgebase can support researchers in the elucidation of the mechanisms leading from a molecular defect to a disease phenotype.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

UniPathway (http://www.unipathway.org) is a fully manually curated resource for the representation and annotation of metabolic pathways. UniPathway provides explicit representations of enzyme-catalyzed and spontaneous chemical reactions, as well as a hierarchical representation of metabolic pathways. This hierarchy uses linear subpathways as the basic building block for the assembly of larger and more complex pathways, including species-specific pathway variants. All of the pathway data in UniPathway has been extensively cross-linked to existing pathway resources such as KEGG and MetaCyc, as well as sequence resources such as the UniProt KnowledgeBase (UniProtKB), for which UniPathway provides a controlled vocabulary for pathway annotation. We introduce here the basic concepts underlying the UniPathway resource, with the aim of allowing users to fully exploit the information provided by UniPathway.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Depuis quelques années, Internet est devenu un média incontournable pour la diffusion de ressources multilingues. Cependant, les différences linguistiques constituent souvent un obstacle majeur aux échanges de documents scientifiques, culturels, pédagogiques et commerciaux. En plus de cette diversité linguistique, on constate le développement croissant de bases de données et de collections composées de différents types de documents textuels ou multimédias, ce qui complexifie également le processus de repérage documentaire. En général, on considère l’image comme « libre » au point de vue linguistique. Toutefois, l’indexation en vocabulaire contrôlé ou libre (non contrôlé) confère à l’image un statut linguistique au même titre que tout document textuel, ce qui peut avoir une incidence sur le repérage. Le but de notre recherche est de vérifier l’existence de différences entre les caractéristiques de deux approches d’indexation pour les images ordinaires représentant des objets de la vie quotidienne, en vocabulaire contrôlé et en vocabulaire libre, et entre les résultats obtenus au moment de leur repérage. Cette étude suppose que les deux approches d’indexation présentent des caractéristiques communes, mais également des différences pouvant influencer le repérage de l’image. Cette recherche permet de vérifier si l’une ou l’autre de ces approches d’indexation surclasse l’autre, en termes d’efficacité, d’efficience et de satisfaction du chercheur d’images, en contexte de repérage multilingue. Afin d’atteindre le but fixé par cette recherche, deux objectifs spécifiques sont définis : identifier les caractéristiques de chacune des deux approches d’indexation de l’image ordinaire représentant des objets de la vie quotidienne pouvant influencer le repérage, en contexte multilingue et exposer les différences sur le plan de l’efficacité, de l’efficience et de la satisfaction du chercheur d’images à repérer des images ordinaires représentant des objets de la vie quotidienne indexées à l’aide d’approches offrant des caractéristiques variées, en contexte multilingue. Trois modes de collecte des données sont employés : l’analyse des termes utilisés pour l’indexation des images, la simulation du repérage d’un ensemble d’images indexées selon chacune des formes d’indexation à l’étude réalisée auprès de soixante répondants, et le questionnaire administré aux participants pendant et après la simulation du repérage. Quatre mesures sont définies pour cette recherche : l’efficacité du repérage d’images, mesurée par le taux de succès du repérage calculé à l’aide du nombre d’images repérées; l’efficience temporelle, mesurée par le temps, en secondes, utilisé par image repérée; l’efficience humaine, mesurée par l’effort humain, en nombre de requêtes formulées par image repérée et la satisfaction du chercheur d’images, mesurée par son autoévaluation suite à chaque tâche de repérage effectuée. Cette recherche montre que sur le plan de l’indexation de l’image ordinaire représentant des objets de la vie quotidienne, les approches d’indexation étudiées diffèrent fondamentalement l’une de l’autre, sur le plan terminologique, perceptuel et structurel. En outre, l’analyse des caractéristiques des deux approches d’indexation révèle que si la langue d’indexation est modifiée, les caractéristiques varient peu au sein d’une même approche d’indexation. Finalement, cette recherche souligne que les deux approches d’indexation à l’étude offrent une performance de repérage des images ordinaires représentant des objets de la vie quotidienne différente sur le plan de l’efficacité, de l’efficience et de la satisfaction du chercheur d’images, selon l’approche et la langue utilisées pour l’indexation.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Ce mémoire porte sur l’analyse documentaire en milieu universitaire. Deux approches générales sont d’abord étudiées : l’approche centrée sur le document (premier chapitre), prédominante dans la tradition bibliothéconomique, et l’approche centrée sur l’usager (deuxième chapitre), influencée par le développement d’outils le plus souvent associés au Web 2.0. L’opposition entre ces deux démarches reflète une dichotomie qui se trouve au cœur de la notion de sujet, c’est-à-dire les dimensions objective et subjective du sujet. Ce mémoire prend par conséquent la forme d’une dissertation dont l’avantage principal est de considérer à la fois d’importants acquis qui appartiennent à la tradition bibliothéconomique, à la fois des développements plus récents ayant un impact important sur l’évolution de l’analyse documentaire en milieu universitaire. Notre hypothèse est que ces deux tendances générales doivent être mises en relief afin d’approfondir la problématique de l’appariement, laquelle définit la difficulté d’accorder le vocabulaire qu’utilise l’usager dans ses recherches documentaires avec celui issu de l’analyse documentaire (métadonnées sujet). Dans le troisième chapitre, nous examinons certaines particularités liées à l’utilisation de la documentation en milieu universitaire dans le but de repérer certaines possibilités et certaines exigences de l’analyse documentaire dans un tel milieu. À partir d’éléments basés sur l’analyse des domaines d’études et sur la démarche analytico-synthétique, il s’agit d’accentuer l’interaction potentielle entre usagers et analystes documentaires sur le plan du vocabulaire utilisé de part et d’autre.