969 resultados para Data Exchange
Resumo:
This paper presents an IEEE 802.11p full-stack prototype implementation to data exchange among vehicles and between vehicles and the roadway infrastructures. The prototype architecture is based on FPGAs for Intermediate Frequency (IF) and base band purposes, using 802.11a based transceivers for RF interfaces. Power amplifiers were also addressed, by using commercial and in-house solutions. This implementation aims to provide technical solutions for Intelligent Transportation Systems (ITS) field, namely for tolling and traffic management related services, in order to promote safety, mobility and driving comfort through the dynamic and real-time cooperation among vehicles and/or between vehicles and infrastructures. The performance of the proposed scheme is tested under realistic urban and suburban driving conditions. Preliminary results are promising, since they comply with most of the 802.11p standard requirements.
Resumo:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do Grau de Mestre em Engenharia Informática.
Resumo:
Belief revision is a critical issue in real world DAI applications. A Multi-Agent System not only has to cope with the intrinsic incompleteness and the constant change of the available knowledge (as in the case of its stand alone counterparts), but also has to deal with possible conflicts between the agents’ perspectives. Each semi-autonomous agent, designed as a combination of a problem solver – assumption based truth maintenance system (ATMS), was enriched with improved capabilities: a distributed context management facility allowing the user to dynamically focus on the more pertinent contexts, and a distributed belief revision algorithm with two levels of consistency. This work contributions include: (i) a concise representation of the shared external facts; (ii) a simple and innovative methodology to achieve distributed context management; and (iii) a reduced inter-agent data exchange format. The different levels of consistency adopted were based on the relevance of the data under consideration: higher relevance data (detected inconsistencies) was granted global consistency while less relevant data (system facts) was assigned local consistency. These abilities are fully supported by the ATMS standard functionalities.
Resumo:
Multi-agent architectures are well suited for complex inherently distributed problem solving domains. From the many challenging aspects that arise within this framework, a crucial one emerges: how to incorporate dynamic and conflicting agent beliefs? While the belief revision activity in a single agent scenario is concentrated on incorporating new information while preserving consistency, in a multi-agent system it also has to deal with possible conflicts between the agents perspectives. To provide an adequate framework, each agent, built as a combination of an assumption based belief revision system and a cooperation layer, was enriched with additional features: a distributed search control mechanism allowing dynamic context management, and a set of different distributed consistency methodologies. As a result, a Distributed Belief Revision Testbed (DiBeRT) was developed. This paper is a preliminary report presenting some of DiBeRT contributions: a concise representation of external beliefs; a simple and innovative methodology to achieve distributed context management; and a reduced inter-agent data exchange format.
Resumo:
Paper presented at the 5th European Conference Economics and Management of Energy in Industry, Vilamoura, Algarve. Apr. 14-17, 2009, 11p. URL: http:// www.cenertec.pt/ecemei/
Resumo:
OBJECTIVE: Quality assurance (QA) in clinical trials is essential to ensure treatment is safely and effectively delivered. As QA requirements have increased in complexity in parallel with evolution of radiation therapy (RT) delivery, a need to facilitate digital data exchange emerged. Our objective is to present the platform developed for the integration and standardization of QART activities across all EORTC trials involving RT. METHODS: The following essential requirements were identified: secure and easy access without on-site software installation; integration within the existing EORTC clinical remote data capture system; and the ability to both customize the platform to specific studies and adapt to future needs. After retrospective testing within several clinical trials, the platform was introduced in phases to participating sites and QART study reviewers. RESULTS: The resulting QA platform, integrating RT analysis software installed at EORTC Headquarters, permits timely, secure, and fully digital central DICOM-RT based data review. Participating sites submit data through a standard secure upload webpage. Supplemental information is submitted in parallel through web-based forms. An internal quality check by the QART office verifies data consistency, formatting, and anonymization. QART reviewers have remote access through a terminal server. Reviewers evaluate submissions for protocol compliance through an online evaluation matrix. Comments are collected by the coordinating centre and institutions are informed of the results. CONCLUSIONS: This web-based central review platform facilitates rapid, extensive, and prospective QART review. This reduces the risk that trial outcomes are compromised through inadequate radiotherapy and facilitates correlation of results with clinical outcomes.
Resumo:
Al llarg d'aquest treball es desenvolupa un sistema de micropagament basat en un model client - servidor en què el procés d'intercanvi de dades no necessita ser validat en temps real. El client té una cartera on guarda les monedes de què disposa per a fer les compres, i el servidor demana les monedes abans de fer la transferència del producte adquirit; aquestes monedes són guardades pel servidor i al final del dia es fa la conversió a diners.
Resumo:
A causa de l'increment del nombre de consumidors de dades geogràfiques, sobretot de dispositius mòbils, els sistemes d'informació geogràfica es troben davant el repte de reduir el coll d'ampolla que suposa l'arquitectura client-servidor clàssica. Una de les solucions a aquest problema és la translació de part de la intel·ligència al node client, així com la creació de xarxes descentralitzades (peer to peer),L'objectiu d'aquesta recerca és demostrar la viabilitat d'una infraestructura client mòbil-servidor, on el client no és únicament un consumidor més, si no que es torna un node intel·ligent.Per fer aquest estudi s'ha desenvolupat una aplicació pel sistema operatiu Android que consumeix dades d'OpenStreetMap. Aquesta aplicació utilitza tècniques de tessel·lat, catching i descàrrega de dades en background segons la posició de l'usuari, per facilitar el consum i reduir el flux de dades intercanviades entre el client i el servidor. També s'ha creat un servei web intern al dispositiu mòbil per a la creació de xarxes peer to peer, les quals permetin un intercanvi de dades entre els terminals mòbils. En aquest treball s'ha demostrat la viabilitat de la infraestructura a través del impacte de l'ús de les tècniques comentades anteriorment sobre el dispositiu client. Per mesurar l'impacte s'ha tingut en compte la càrrega de la CPU (la qual repercuteix al consum de la bateria) i el temps de resposta del sistema. Els resultats de les proves realitzades indiquen que aquestes tècniques redueixen el temps de cerca de punts d'interès d'una manera dràstica, però que també tenen un alt impacte a la CPU i en el temps de càrrega, sobretot en dispositius amb menys capacitats.
Resumo:
Roughly fifteen years ago, the Church of Jesus Christ of Latter-day Saints published a new proposed standard file format. They call it GEDCOM. It was designed to allow different genealogy programs to exchange data.Five years later, in may 2000, appeared the GENTECH Data Modeling Project, with the support of the Federation of Genealogical Societies (FGS) and other American genealogical societies. They attempted to define a genealogical logic data model to facilitate data exchange between different genealogical programs. Although genealogists deal with an enormous variety of data sources, one of the central concepts of this data model was that all genealogical data could be broken down into a series of short, formal genealogical statements. It was something more versatile than only export/import data records on a predefined fields. This project was finally absorbed in 2004 by the National Genealogical Society (NGS).Despite being a genealogical reference in many applications, these models have serious drawbacks to adapt to different cultural and social environments. At the present time we have no formal proposal for a recognized standard to represent the family domain.Here we propose an alternative conceptual model, largely inherited from aforementioned models. The design is intended to overcome their limitations. However, its major innovation lies in applying the ontological paradigm when modeling statements and entities.
Resumo:
A population register is an inventory of residents within a country, with their characteristics (date of birth, sex, marital status, etc.) and other socio-economic data, such as occupation or education. However, data on population are also stored in numerous other public registers such as tax, land, building and housing, military, foreigners, vehicles, etc. Altogether they contain vast amounts of personal and sensitive information. Access to public information is granted by law in many countries, but this transparency is generally subject to tensions with data protection laws. This paper proposes a framework to analyze data access (or protection) requirements, as well as a model of metadata for data exchange.
Resumo:
BACKGROUND: Molecular interaction Information is a key resource in modern biomedical research. Publicly available data have previously been provided in a broad array of diverse formats, making access to this very difficult. The publication and wide implementation of the Human Proteome Organisation Proteomics Standards Initiative Molecular Interactions (HUPO PSI-MI) format in 2004 was a major step towards the establishment of a single, unified format by which molecular interactions should be presented, but focused purely on protein-protein interactions. RESULTS: The HUPO-PSI has further developed the PSI-MI XML schema to enable the description of interactions between a wider range of molecular types, for example nucleic acids, chemical entities, and molecular complexes. Extensive details about each supported molecular interaction can now be captured, including the biological role of each molecule within that interaction, detailed description of interacting domains, and the kinetic parameters of the interaction. The format is supported by data management and analysis tools and has been adopted by major interaction data providers. Additionally, a simpler, tab-delimited format MITAB2.5 has been developed for the benefit of users who require only minimal information in an easy to access configuration. CONCLUSION: The PSI-MI XML2.5 and MITAB2.5 formats have been jointly developed by interaction data producers and providers from both the academic and commercial sector, and are already widely implemented and well supported by an active development community. PSI-MI XML2.5 enables the description of highly detailed molecular interaction data and facilitates data exchange between databases and users without loss of information. MITAB2.5 is a simpler format appropriate for fast Perl parsing or loading into Microsoft Excel.
Resumo:
In June 2006, the Swiss Parliament made two important decisions with regards to public registers' governance and individuals' identification. It adopted a new law on the harmonisation of population registers in order to simplify statistical data collection and data exchange from around 4'000 decentralized registers, and it also approved the introduction of a Unique Person Identifier (UPI). The law is rather vague about the implementation of this harmonisation and even though many projects are currently being undertaken in this domain, most of them are quite technical. We believe there is a need for analysis tools and therefore we propose a conceptual framework based on three pillars (Privacy, Identity and Governance) to analyse the requirements in terms of data management for population registers.
Resumo:
This document describes some of the technological aspects of a project devoted to the creation of a factory for language resources. The project’s objectives are explained, as well as the idea to create a distributed infrastructure of web services. This document focuses on two main topics of the factory: (1) the technological approaches chosen to develop the factory, i.e. software, protocols, servers, etc. (2) and Interoperability as the main challenge is to permit different NLP tools work together in the factory. This document explains why XCES and GrAF are chosen as the main formats used for the linguistic data exchange.
Resumo:
La Universitat de Vic disposa, entre altres equips, d’una cèl·lula flexible de fabricació, del fabricant Festo, que simula un procés de formació de palets amb els productes que es disposen en un magatzem intermedi. Aquesta cèl·lula està composta de quatre estacions de muntatge diferenciades (càrrega de palets, càrrega de plaques, magatzem intermedi i transport). Cada una disposa d'un PLC SIEMENS S7-300 per la seva automatització, i tots aquests es troben interconnectats amb una xarxa industrial Profibus. L'objectiu d'aquest projecte és implantar el sistema SCADA Vijeo Citect pel control i supervisió de l'estació magatzem d'aquesta cèl·lula flexible de fabricació, establint també un intercanvi de dades entre l'SCADA i el Microsoft Access, per poder ser utilitzat per la docència. Aquest projecte s'ha desenvolupat en cinc fases diferents: 1. La primera fase s'ha dedicat a l'automatització pròpiament de l'estació magatzem a partir de l'autòmat programable Siemens S7-300 i complint amb les necessitats plantejades. 2. En la segona fase s'ha programat i establert la comunicació per l'intercanvi de dades (lectura i escriptura) entre el sistema SCADA Vijeo Citect i la base de dades de Microsoft Access. 3. En la tercera fase s'ha elaborat i programat l'entorn gràfic de supervisió i control del procés a partir del sistema SCADA Vijeo Citect. 4. En la quarta fase s'ha instal·lat un OPC Server en el PC i s'ha establert la comunicació entre el PLC i el sistema SCADA. 5. Finalment s'ha anat revisant i depurant les diferents programacions i comunicacions per tal de que el sistema funcioni com a un conjunt.
Resumo:
Tämä diplomityö käsittelee kahden sotilaallisen koalition yhteentoimivuuden haasteita ja toteutusta kaupallisintyökaluin. Työ pohjautuu kahteen todelliseen, palvelusuuntautunutta arkkitehtuurisuunnittelua (SOA) hyödyntäneeseen integraatioprojektiin jotka on toteutettu Suomen IBM:n Palveluyksikössä vuosina 2006 - 2007. Työn tavoitteena on ollut tutkia sotilaallisten koalitioiden järjestelmä - ja tiedonvaihtoyhteentoimivuuden menetelmiä, näkökulmia ja teknistä toteutusta kaupallisin ohjelmistotuottein ja yhteistä tietomallia käyttäen. Lisäksi esitetään puolustustoimialan erityispiirteet tietojärjestelmätoimittajien ohjelmistokehitys - prosesseihin liittyen. Tätä varten tutkittiin koalitioiden käyttöön tarkoitettuja olemassa olevia ohjelmistoarkkitehtuureja ja yhteentoimivuusmalleja sekä sovitettiin niitä SOA - arkkitehtuuriajatteluun. Työn teoreettisena pohjana käytettiin organisatorisen ja teknisen yhteentoimivuuden kuvaavaa Layers of Coalition Interoperability (LCI) - mallia, minkä jälkeen mallin teknistä osiota käytettiin pohjana SOA - palveluihin perustuvan esimerkkijärjestelmän kehittämiseen kahden kuvitteellisen koalition tiedonvaihtoa varten. Työn keskeisinä tuloksina on syntynyt suunnitelma koalitioiden taustajärjestelmien yhdistämisestä dynaamisten SOA- palveluiden avulla yhteiseen JC3IEDM - tietomalliin. Tämä tietomalli vuorostaan antaa mahdollisuuden järjestelmän laajentamiseen esimerkiksi avustusjärjestöjen, poliisivoimien ja terveydenhuollon organisaatioiden tarpeisiin.