894 resultados para data integration
Resumo:
论文结合石化企业的数据组织现状和应用需求,提出了由数据转换层、数据集成层和数据接口层组成的石化企业数据集成平台的三层体系结构,三层体系结构的划分使得数据集成平台具有良好的开放性和可扩展性。设计了基于消息中间件、XML技术和适配器技术的数据集成平台的技术架构。论文提供了一个石化企业数据集成平台的整体解决方案,它对企业数据集成平台的建立具有很大的指导意义。
Resumo:
首先分析了企业对数据集成的需求,并指出数据集成所面临的主要问题.针对这些问题,提出了以企业核心数据模型为驱动的数据集成方法,而且给出了该集成方法的一种实现,称之为企业数据集成平台,并以炼化企业的具体应用为例,论述了该方法对于解决企业数据集成问题的有效性.
Resumo:
针对当今中国流程企业的现状,提出了一种对企业多种数据源进行整合和管理的解决方案--流程企业数据平台.该数据平台可以实现流程企业中多种数据的集成,并支持企业业务过程的持续改善和企业新应用的开发.流程企业数据平台由数据模型、数据发布接口、数据整合工具和平台管理工具组成,是平台级的软件产品.针对流程企业典型代表的石化企业的特点,给出了数据平台的实现.
Resumo:
针对目前企业普遍存在的信息孤岛林立和数据断层的局面,本文提出了一种进行企业数据集成的模型--企业统一数据模型,应用该模型可以较好地解决企业混乱的数据环境,建立一个统一高效的企业数据平台.同时本文结合在炼化企业建立统一数据模型的经验,详细论述了企业统一数据模型的建模方法.
Resumo:
流程企业中的信息主要包括管理数据和过程数据,随着企业信息化进程的推进,实现管理数据和过程数据的集成对提高流程企业生产管理与决策的质量、效益和水平起到越来越重要的作用,为实现流程企业的管控一体化创造了条件。本文在对流程企业的数据集成需求和现有的过程数据集成方法全面分析的基础上,针对现有的过程数据集成方法无法保证数据集成过程的主动性以及数据的时效性问题,提出了基于ECA规则的过程数据集成方法,该方法将具有主动机制的ECA规则引入到数据集成中,通过事件触发规则来完成数据集成任务,保证了数据的时效性和可靠性,具有一定的应用价值。论文主要工作如下: 1.分析了流程企业生产数据的特点和过程数据集成的重要性,总结了现有的解决过程数据集成问题方法的不足,针对已有方法无法保证过程数据集成的主动性和数据的时效性问题,提出了基于ECA规则的过程数据集成方法。 2.分析了流程企业数据集成的具体需求特点,给出了过程数据集成系统中的ECA规则的设计描述。基于实时数据库提供的发布/订阅机制,阐述了系统对规则中主要事件类型产生的捕获处理过程。 3.设计了基于ECA规则的过程数据集成框架,研究了框架中主要模块设计所用到的关键技术,包括复合事件检测机制、规则语法解析和规则调度策略,分析了该框架的特点,说明了系统运行流程。 4.基于上述研究,开发了基于ECA规则的企业过程数据集成系统。该系统已经投入石化和钢铁企业的实际应用中,取得了较好的效果。
Resumo:
数据集成是当今信息社会的必然需求,人们迫切需要从海量的数据获取宝贵的信息资源,以供科学管理及决策分析,而数据集成是其中最关键的步骤。数据集成主要用来解决数据的一致性,集成化问题,使我们能够从不同的业务系统和应用环境中获取数据,并进行高效的转换,清洗,成为数据分析系统期望数据,因而数据集成技术的研究具有重要意义。 本文对数据集成技术进行了相关的探索,提出来一种基于元数据的设计和实现方法。本文首先研究了数据集成本身存在的关键问题和难点,然后对当前国内外数据集成技术的研究状况进行了分析,进而提出了基于元数据技术,工作流技术以及插件机制三者结合的数据集成技术实现方案。该方案把整个数据集成过程需要的各个处理规则节点进行插件式封装,由它们组成灵活的的工作流,并把整个任务进行元数据描述,然后通过集成引擎解释元数据描述文件,执行相应处理规则,从而达到数据集成的目的。本文的核心内容是元数据机制、工作流机制和插件机制在数据集成工具中的应用设计与实现,这也是本文的创新之处,具体包括有系统元模型的设计、数据流控制模型的设计、系统体系结构设计以及系统的详细设计与实现等。其中,还构建了支持拖拽的流程建模工具,提高系统的易用性和可操作性。 本文的重点在弥补现有工具的一些不足,提高数据集成工具的扩展性、高效性、灵活性以及易操作性,并在保证工具具有良好的数据集成功能的同时,能提供良好的柔性支持,且有效降低成本。本文研究和设计的数据集成工具已经具备了的数据集成的基本功能,具有实际的应用价值,只要进行简单的扩展,便可应用到不同的领域。
Resumo:
针对MES与过程控制系统PCS系统的数据集成,提出一种基于实时历史数据库的数据集成平台模型,该平台可以满足生产过程中大量数据的存储要求,统一了MES对PCS的访问接口,提高了数据采集放率和数据通信的可靠性,是MES成功实施的基础。
Resumo:
For at least two millennia and probably much longer, the traditional vehicle for communicating geographical information to end-users has been the map. With the advent of computers, the means of both producing and consuming maps have radically been transformed, while the inherent nature of the information product has also expanded and diversified rapidly. This has given rise in recent years to the new concept of geovisualisation (GVIS), which draws on the skills of the traditional cartographer, but extends them into three spatial dimensions and may also add temporality, photorealistic representations and/or interactivity. Demand for GVIS technologies and their applications has increased significantly in recent years, driven by the need to study complex geographical events and in particular their associated consequences and to communicate the results of these studies to a diversity of audiences and stakeholder groups. GVIS has data integration, multi-dimensional spatial display advanced modelling techniques, dynamic design and development environments and field-specific application needs. To meet with these needs, GVIS tools should be both powerful and inherently usable, in order to facilitate their role in helping interpret and communicate geographic problems. However no framework currently exists for ensuring this usability. The research presented here seeks to fill this gap, by addressing the challenges of incorporating user requirements in GVIS tool design. It starts from the premise that usability in GVIS should be incorporated and implemented throughout the whole design and development process. To facilitate this, Subject Technology Matching (STM) is proposed as a new approach to assessing and interpreting user requirements. Based on STM, a new design framework called Usability Enhanced Coordination Design (UECD) is ten presented with the purpose of leveraging overall usability of the design outputs. UECD places GVIS experts in a new key role in the design process, to form a more coordinated and integrated workflow and a more focused and interactive usability testing. To prove the concept, these theoretical elements of the framework have been implemented in two test projects: one is the creation of a coastal inundation simulation for Whitegate, Cork, Ireland; the other is a flooding mapping tool for Zhushan Town, Jiangsu, China. The two case studies successfully demonstrated the potential merits of the UECD approach when GVIS techniques are applied to geographic problem solving and decision making. The thesis delivers a comprehensive understanding of the development and challenges of GVIS technology, its usability concerns, usability and associated UCD; it explores the possibility of putting UCD framework in GVIS design; it constructs a new theoretical design framework called UECD which aims to make the whole design process usability driven; it develops the key concept of STM into a template set to improve the performance of a GVIS design. These key conceptual and procedural foundations can be built on future research, aimed at further refining and developing UECD as a useful design methodology for GVIS scholars and practitioners.
Resumo:
BACKGROUND: A hierarchical taxonomy of organisms is a prerequisite for semantic integration of biodiversity data. Ideally, there would be a single, expansive, authoritative taxonomy that includes extinct and extant taxa, information on synonyms and common names, and monophyletic supraspecific taxa that reflect our current understanding of phylogenetic relationships. DESCRIPTION: As a step towards development of such a resource, and to enable large-scale integration of phenotypic data across vertebrates, we created the Vertebrate Taxonomy Ontology (VTO), a semantically defined taxonomic resource derived from the integration of existing taxonomic compilations, and freely distributed under a Creative Commons Zero (CC0) public domain waiver. The VTO includes both extant and extinct vertebrates and currently contains 106,947 taxonomic terms, 22 taxonomic ranks, 104,736 synonyms, and 162,400 cross-references to other taxonomic resources. Key challenges in constructing the VTO included (1) extracting and merging names, synonyms, and identifiers from heterogeneous sources; (2) structuring hierarchies of terms based on evolutionary relationships and the principle of monophyly; and (3) automating this process as much as possible to accommodate updates in source taxonomies. CONCLUSIONS: The VTO is the primary source of taxonomic information used by the Phenoscape Knowledgebase (http://phenoscape.org/), which integrates genetic and evolutionary phenotype data across both model and non-model vertebrates. The VTO is useful for inferring phenotypic changes on the vertebrate tree of life, which enables queries for candidate genes for various episodes in vertebrate evolution.
Resumo:
Query processing over the Internet involving autonomous data sources is a major task in data integration. It requires the estimated costs of possible queries in order to select the best one that has the minimum cost. In this context, the cost of a query is affected by three factors: network congestion, server contention state, and complexity of the query. In this paper, we study the effects of both the network congestion and server contention state on the cost of a query. We refer to these two factors together as system contention states. We present a new approach to determining the system contention states by clustering the costs of a sample query. For each system contention state, we construct two cost formulas for unary and join queries respectively using the multiple regression process. When a new query is submitted, its system contention state is estimated first using either the time slides method or the statistical method. The cost of the query is then calculated using the corresponding cost formulas. The estimated cost of the query is further adjusted to improve its accuracy. Our experiments show that our methods can produce quite accurate cost estimates of the submitted queries to remote data sources over the Internet.
Resumo:
Web databases are now pervasive. Such a database can be accessed via its query interface (usually HTML query form) only. Extracting Web query interfaces is a critical step in data integration across multiple Web databases, which creates a formal representation of a query form by extracting a set of query conditions in it. This paper presents a novel approach to extracting Web query interfaces. In this approach, a generic set of query condition rules are created to define query conditions that are semantically equivalent to SQL search conditions. Query condition rules represent the semantic roles that labels and form elements play in query conditions, and how they are hierarchically grouped into constructs of query conditions. To group labels and form elements in a query form, we explore both their structural proximity in the hierarchy of structures in the query form, which is captured by a tree of nested tags in the HTML codes of the form, and their semantic similarity, which is captured by various short texts used in labels, form elements and their properties. We have implemented the proposed approach and our experimental results show that the approach is highly effective.