947 resultados para XML, Schema matching
Resumo:
Dissertação para obtenção do grau de Mestre em Engenharia Civil na Área de Especialização de Vias de Comunicação e Transportes
Resumo:
A manutenção do conteúdo web pode ser uma tarefa difícil, especialmente se considerarmos websites em que muitos utilizadores têm permissões para alterar o seu conteúdo. Um exemplo deste tipo de websites são os wikis. Se por um lado permitem rápida disseminação de conhecimento, por outro lado implicam um grande esforço para verificar a qualidade do seu conteúdo. Nesta tese analisamos diferentes abordagens à modelação de websites, especialmente para a verificação de conteúdo, onde contribuímos com uma extensão à ferramenta VeriFLog para a tornar mais adequada à verificação de conteúdos em websites colaborativos.
Resumo:
O contributo da área de investigação Interacção Humano-Computador (HCI) está patente não só na qualidade da interacção, mas também na diversificação das formas de interacção. A HCI define-se como sendo uma disciplina que se dedica ao desenho, desenvolvimento e implementação de sistemas de computação interactivos para uso humano e estudo dos fenómenos relevantes que os rodeiam. Pretende-se, no âmbito desta tese de mestrado, o desenvolvimento de um Editor Gráfico de Layout Fabril a integrar num SAD para suporte ao Planeamento e Controlo da Produção. O sistema deve ser capaz de gerar um layout fabril do qual constam, entre outros objectos, as representações gráficas e as respectivas características/atributos do conjunto de recursos (máquinas/processadores) existentes no sistema de produção a modelar. O módulo desenvolvido será integrado no projecto de I&D ADSyS (Adaptative Decision Support System for Interactive Scheduling with MetaCognition and User Modeling Experience), melhorando aspectos de interacção referentes ao sistema AutoDynAgents, um dedicado ao escalonamento, planeamento e controlo de produção. Foi realizada a análise de usabilidade a este módulo com a qual se pretendeu realizar a respectiva avaliação, através da realização de um teste de eficiência e do preenchimento de um inquérito, da qual se identificaram um conjunto de melhorias e sugestões a serem consideradas no refinamento deste módulo.
Resumo:
Dissertação apresentada ao Instituto Superior de Contabilidade e Administração do Porto para obtenção do Grau de Mestre em Assessoria de Administração Orientada pela professora Arminda Sá Sequeira
Resumo:
The discussion and analysis of the diverse outreach activities in this article provide guidance and suggestions for academic librarians who are interested in outreach and community engagement of any scale and nature. Cases are draw from a wide spectrum and are particularly strong in the setting of large academic libraries, special collections and programming for multicultural populations. The aim of this study is to present the results of research carried out regarding the needs, demand and consumption of European Union information by users in European Documentation Centres (EDC). A quantitative methodology was chosen based on a questionnaire with 24 items. This questionnaire was distributed within the EDC of Salamanca, Spain, and the EDC of Porto, Portugal, during specific time intervals between 2010 and 2011. We examined the level of EU information that EDC users possess, and identified the factors that facilitate or hinder access to EU information, the topics most demanded, and the types of documents consulted. Analysis was made of the use that the consumer of European information makes of databases and their behaviour during the consultation. Although the sample used was not very significant owing to its small size, it is a faithful reflection of the scarce visits made to EDCs. This study can be of use to managers of EDCs, providing them with better knowledge of the information needs and demands of their users. Ultimately this should lead to improvements in the services offered. The study lies within a frame of research scarcely addressed in specialized scholarly literature: European Union information.
Resumo:
OBJECTIVE: To analyze cause-specific mortality rates according to the relative income hypothesis. METHODS: All 96 administrative areas of the city of São Paulo, southeastern Brazil, were divided into two groups based on the Gini coefficient of income inequality: high (>0.25) and low (<0.25). The propensity score matching method was applied to control for confounders associated with socioeconomic differences among areas. RESULTS: The difference between high and low income inequality areas was statistically significant for homicide (8.57 per 10,000; 95%CI: 2.60;14.53); ischemic heart disease (5.47 per 10,000 [95%CI 0.76;10.17]); HIV/AIDS (3.58 per 10,000 [95%CI 0.58;6.57]); and respiratory diseases (3.56 per 10,000 [95%CI 0.18;6.94]). The ten most common causes of death accounted for 72.30% of the mortality difference. Infant mortality also had significantly higher age-adjusted rates in high inequality areas (2.80 per 10,000 [95%CI 0.86;4.74]), as well as among males (27.37 per 10,000 [95%CI 6.19;48.55]) and females (15.07 per 10,000 [95%CI 3.65;26.48]). CONCLUSIONS: The study results support the relative income hypothesis. After propensity score matching cause-specific mortality rates was higher in more unequal areas. Studies on income inequality in smaller areas should take proper accounting of heterogeneity of social and demographic characteristics.
Resumo:
Nowadays, due to the incredible grow of the mobile devices market, when we want to implement a client-server applications we must consider mobile devices limitations. In this paper we discuss which can be the more reliable and fast way to exchange information between a server and an Android mobile application. This is an important issue because with a responsive application the user experience is more enjoyable. In this paper we present a study that test and evaluate two data transfer protocols, socket and HTTP, and three data serialization formats (XML, JSON and Protocol Buffers) using different environments and mobile devices to realize which is the most practical and fast to use.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Informática e Computadores
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Redes de Comunicação e Multimédia
Resumo:
Gestational diabetes mellitus (GDM) and controversy are old friends. The impact of GDM on maternal and fetal health has been increasingly recognized. Nevertheless, universal consensus on the diagnostic methods and thresholds has long been lacking. Published guidelines from major societies differ significantly from one another, with recommendations ranging from aggressive screening to no routine screening at all. As a result, real-world practice is equally varied. This article recaps the latest evidence-based recommendations for the diagnosis and classification of GDM. It reviews the current evidence base for intensive multidisciplinary treatment of GDM and provides recommendations for postpartum management to delay and/or prevent progression to type 2 diabetes.
Resumo:
I (Prática Pedagógica) - O estágio foi realizado na Academia de Música de Almada (AMA) com três alunos de níveis diferentes: Aluno x, Iniciação Musical; Aluna y, 2º grau; e Aluna z, 5º grau. A planificação do trabalho foi feita com a colaboração do professor Vasco Broco e sob a supervisão do professor Pedro Saglimbeni Muñoz, tendo por base o programa da AMA e procurando corresponder às necessidades individuais de cada aluno, bem como respeitar os diferentes ritmos de aprendizagem. Dadas as diferenças de idade e de nível, os objectivos que procurei atingir com cada aluno durante o ano lectivo são naturalmente díspares, bem como a perspectiva de cada um em relação ao próprio instrumento e ao ensino da música. Enquanto que os mais novos encaram a aprendizagem musical como uma actividade meramente recreativa, pelo menos por enquanto, a Aluna z ambiciona fazer uma carreira como instrumentista. Mesmo entre os mais novos há diferenças, pois apesar de o Aluno x ainda estar a frequentar a Iniciação Musical, já iniciou a sua aprendizagem há mais tempo do que a Aluna y, encarando o ensino com mais seriedade. A AMA tem a vantagem de ter um pianista acompanhador disponível para cada dia da semana, o que tornou possível desenvolver um trabalho consistente com os alunos em termos de ensaios e audições. As audições e momentos de avaliação foram marcados em reuniões de Departamento Curricular, à excepção da calendarização das Provas Globais que foi feita em reunião de Conselho Pedagógico. O aproveitamento foi positivo para os três alunos no final do estágio, tendo todos transitado para o nível de ensino seguinte.
Resumo:
After a historical introduction, the bulk of the thesis concerns the study of a declarative semantics for logic programs. The main original contributions are: ² WFSX (Well–Founded Semantics with eXplicit negation), a new semantics for logic programs with explicit negation (i.e. extended logic programs), which compares favourably in its properties with other extant semantics. ² A generic characterization schema that facilitates comparisons among a diversity of semantics of extended logic programs, including WFSX. ² An autoepistemic and a default logic corresponding to WFSX, which solve existing problems of the classical approaches to autoepistemic and default logics, and clarify the meaning of explicit negation in logic programs. ² A framework for defining a spectrum of semantics of extended logic programs based on the abduction of negative hypotheses. This framework allows for the characterization of different levels of scepticism/credulity, consensuality, and argumentation. One of the semantics of abduction coincides with WFSX. ² O–semantics, a semantics that uniquely adds more CWA hypotheses to WFSX. The techniques used for doing so are applicable as well to the well–founded semantics of normal logic programs. ² By introducing explicit negation into logic programs contradiction may appear. I present two approaches for dealing with contradiction, and show their equivalence. One of the approaches consists in avoiding contradiction, and is based on restrictions in the adoption of abductive hypotheses. The other approach consists in removing contradiction, and is based in a transformation of contradictory programs into noncontradictory ones, guided by the reasons for contradiction.
Resumo:
The principal topic of this work is the application of data mining techniques, in particular of machine learning, to the discovery of knowledge in a protein database. In the first chapter a general background is presented. Namely, in section 1.1 we overview the methodology of a Data Mining project and its main algorithms. In section 1.2 an introduction to the proteins and its supporting file formats is outlined. This chapter is concluded with section 1.3 which defines that main problem we pretend to address with this work: determine if an amino acid is exposed or buried in a protein, in a discrete way (i.e.: not continuous), for five exposition levels: 2%, 10%, 20%, 25% and 30%. In the second chapter, following closely the CRISP-DM methodology, whole the process of construction the database that supported this work is presented. Namely, it is described the process of loading data from the Protein Data Bank, DSSP and SCOP. Then an initial data exploration is performed and a simple prediction model (baseline) of the relative solvent accessibility of an amino acid is introduced. It is also introduced the Data Mining Table Creator, a program developed to produce the data mining tables required for this problem. In the third chapter the results obtained are analyzed with statistical significance tests. Initially the several used classifiers (Neural Networks, C5.0, CART and Chaid) are compared and it is concluded that C5.0 is the most suitable for the problem at stake. It is also compared the influence of parameters like the amino acid information level, the amino acid window size and the SCOP class type in the accuracy of the predictive models. The fourth chapter starts with a brief revision of the literature about amino acid relative solvent accessibility. Then, we overview the main results achieved and finally discuss about possible future work. The fifth and last chapter consists of appendices. Appendix A has the schema of the database that supported this thesis. Appendix B has a set of tables with additional information. Appendix C describes the software provided in the DVD accompanying this thesis that allows the reconstruction of the present work.
Resumo:
Known algorithms capable of scheduling implicit-deadline sporadic tasks over identical processors at up to 100% utilisation invariably involve numerous preemptions and migrations. To the challenge of devising a scheduling scheme with as few preemptions and migrations as possible, for a given guaranteed utilisation bound, we respond with the algorithm NPS-F. It is configurable with a parameter, trading off guaranteed schedulable utilisation (up to 100%) vs preemptions. For any possible configuration, NPS-F introduces fewer preemptions than any other known algorithm matching its utilisation bound. A clustered variant of the algorithm, for systems made of multicore chips, eliminates (costly) off-chip task migrations, by dividing processors into disjoint clusters, formed by cores on the same chip (with the cluster size being a parameter). Clusters are independently scheduled (each, using non-clustered NPS-F). The utilisation bound is only moderately affected. We also formulate an important extension (applicable to both clustered and non-clustered NPS-F) which optimises the supply of processing time to executing tasks and makes it more granular. This reduces processing capacity requirements for schedulability without increasing preemptions.
Resumo:
Most current-generation Wireless Sensor Network (WSN) nodes are equipped with multiple sensors of various types, and therefore support for multi-tasking and multiple concurrent applications is becoming increasingly common. This trend has been fostering the design of WSNs allowing several concurrent users to deploy applications with dissimilar requirements. In this paper, we extend the advantages of a holistic programming scheme by designing a novel compiler-assisted scheduling approach (called REIS) able to identify and eliminate redundancies across applications. To achieve this useful high-level optimization, we model each user application as a linear sequence of executable instructions. We show how well-known string-matching algorithms such as the Longest Common Subsequence (LCS) and the Shortest Common Super-sequence (SCS) can be used to produce an optimal merged monolithic sequence of the deployed applications that takes into account embedded scheduling information. We show that our approach can help in achieving about 60% average energy savings in processor usage compared to the normal execution of concurrent applications.