940 resultados para Multi- Choice mixed integer goal programming


Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis has been realised through a scholarship offered by the Government of Canada to the Government of the Republic of Mauritius under the Programme Canadien de Bourses de la Francophonie

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Au cours des trois dernières décennies, le travail quotidien des cadres gestionnaires infirmiers de premier niveau (CGIPN) a subi de profondes transformations quant aux rôles qu’ils occupent et aux responsabilités qui leur incombent dans les divers établissements de santé. Ces CGIPN vivent et subissent de la pression au regard de leur charge de travail, faisant en sorte qu’ils perçoivent avoir peu de temps pour réaliser adéquatement leur travail en gestion, et que plusieurs infirmières de la relève qui souhaitent faire carrière en gestion se questionnent sur la possibilité d’œuvrer dans ce domaine de la pratique. Peu d’études portent sur l’état de la qualité de vie au travail (QVT) chez ces professionnels. Divisé en trois volets, le projet de recherche mixte séquentielle exploratoire a pour objectif général d’explorer la QVT chez les CGIPN et les facteurs qui l’influencent dans les divers établissements de santé québécois. Les objectifs spécifiques visent à : 1) décrire et comprendre ce que signifie la QVT auprès des 14 CGIPN œuvrant en Centre hospitalier affilié (CHA) à l’Université, 2) décrire et comprendre ce que signifie l’absence d’une QVT auprès des CGIPN en CHA, 3) développer et valider un nouvel instrument de mesure des facteurs favorables et défavorables pouvant influer sur la QVT à partir des données qualitatives auprès de CGIPN (n= 11) œuvrant en Centre de santé et des services sociaux et de CGIPN (n= 11) en Centre hospitalier universitaire, 4) identifier et mesurer les facteurs qui influencent la QVT des CGIPN (n= 291) travaillant dans divers établissements publics de santé et 5) déterminer si les caractéristiques sociodémographiques influencent le choix des facteurs de la QVT. Inspiré de la philosophie du Human caring de Watson comme perspective disciplinaire, le premier volet qualitatif fait appel à une méthode phénoménologique descriptive husserlienne. Pour ce faire, deux séries d’entretiens semi-dirigés à l’aide d’un guide d’entrevue visent à répondre aux deux premiers objectifs. Le second volet cible l’élaboration et la validation d’un instrument de mesure permettant de répondre au troisième objectif de la recherche et le dernier volet, de type quantitatif corrélationnel, s’applique à répondre aux deux derniers objectifs. Dans la première phase, l’analyse visuelle des données qualitatives (verbatim) permet l’émergence des résultats qualitatifs, soit de cinq eidos-thèmes favorables exprimés par les participants et décrits par ordre d’importance : 1) l’actualisation du leadership et des habiletés politiques pour l’amélioration de la qualité des soins infirmiers, 2) les éléments contextuels propices à l’humanisation organisationnelle, 3) le soutien organisationnel favorisant l’épanouissement socioprofessionnel et personnel, 4) l’organisation apprenante favorisant le développement des compétences en gestion des soins infirmiers et 5) l’accompagnement personnalisé répondant aux besoins spécifiques des novices en gestion des soins infirmiers. L’essence de la QVT pour les CGIPN se définit par l’émancipation socioprofessionnelle du cadre gestionnaire infirmier de premier niveau dans sa pratique clinico-administrative au sein d’une organisation humaniste. De plus, trois eidos-thèmes défavorables se dégagent des analyses qualitatives. Les résultats, décrits par ordre d’importance, sont : 1) la déshumanisation organisationnelle, 2) les conditions défavorables à la pratique en gestion des soins infirmiers et 3) l’accompagnement insuffisant des gestionnaires infirmiers novices. L’essence de l’absence d’une QVT pour les CGIPN se présente comme la dysharmonie au travail du cadre gestionnaire infirmier de premier niveau à l’intérieur d’une structure organisationnelle déshumanisante mettant en péril sa pratique clinico-administrative. Par la suite, une mise en relation des deux essences du phénomène a permis de faire émerger la signification universelle de l’expérience de la QVT et l’absence de celle-ci pour des CGIPN œuvrant en CHA comme étant : une dialectique en administration des services infirmiers dans laquelle se vit des pratiques humanisantes permettant l’atteinte d’un idéal de QVT, en coexistence avec des pratiques déshumanisantes conduisant à l’absence de QVT. Afin de respecter les postulats des méthodes quantitatives, les huit eidos-thèmes qui émergent de la première phase qualitative se transforment en facteurs favorables (FF) et défavorables (FD) à la QVT pour la réalisation des deuxième et troisième phases de la recherche. Dans la seconde phase, le construit théorique qui provient des huit eidos-thèmes et 23 thèmes permet de développer un nouvel instrument de mesure qui s’avère unique à cette population à l’étude. L’analyse psychométrique de l’instrument de mesure (questionnaire auto-administré en mode électronique) permet l’obtention d’un score global du coefficient alpha de Cronbach (α) de 0,93, ce qui est jugé excellent. Dans la troisième et dernière phase, les analyses statistiques des données (logiciel SPSS, version 22 pour Windows 7, 2013) issues de l’enquête quantitative provinciale en ligne (Web) mettent en évidence des résultats probants. Les résultats du questionnaire en ligne avec une échelle ordinale à cinq niveaux révèlent un score global du coefficient alpha de Cronbach à 0,95. De plus, les résultats quantitatifs suggèrent que les facteurs favorables (FF) à la QVT ont des scores moyens de 3,99 (FF1), 3,97 (FF2), 3,96 (FF3), 3,72 (FF4) et 3,53 (FF5) et les trois facteurs défavorables (FD) à la QVT ont des scores moyens de 3,91(FD1), 3,78 (FD2) et 3,56 (FD3). Ces facteurs correspondent aux eidos-thèmes favorables et défavorables de la QVT du volet qualitatif phénoménologique. Les résultats quantitatifs révèlent quelques relations significatives des caractéristiques sociodémographiques (âge et années d’expérience) sur le choix des facteurs qui influencent le plus fortement la QVT. En conclusion, le fait de pouvoir mieux décrire et comprendre la signification de la QVT ainsi que son absence et de pouvoir identifier les facteurs qui sont les plus susceptibles de l’influencer permet de formuler des recommandations pour les quatre domaines de la pratique infirmière. Des pistes d’avenues sont formulées pour maintenir la relève infirmière à des postes de CGIPN au sein des organisations de santé en lui assurant des moyens pour optimiser humainement sa QVT.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The theme of the thesis is centred around one important aspect of wireless sensor networks; the energy-efficiency.The limited energy source of the sensor nodes calls for design of energy-efficient routing protocols. The schemes for protocol design should try to minimize the number of communications among the nodes to save energy. Cluster based techniques were found energy-efficient. In this method clusters are formed and data from different nodes are collected under a cluster head belonging to each clusters and then forwarded it to the base station.Appropriate cluster head selection process and generation of desirable distribution of the clusters can reduce energy consumption of the network and prolong the network lifetime. In this work two such schemes were developed for static wireless sensor networks.In the first scheme, the energy wastage due to cluster rebuilding incorporating all the nodes were addressed. A tree based scheme is presented to alleviate this problem by rebuilding only sub clusters of the network. An analytical model of energy consumption of proposed scheme is developed and the scheme is compared with existing cluster based scheme. The simulation study proved the energy savings observed.The second scheme concentrated to build load-balanced energy efficient clusters to prolong the lifetime of the network. A voting based approach to utilise the neighbor node information in the cluster head selection process is proposed. The number of nodes joining a cluster is restricted to have equal sized optimum clusters. Multi-hop communication among the cluster heads is also introduced to reduce the energy consumption. The simulation study has shown that the scheme results in balanced clusters and the network achieves reduction in energy consumption.The main conclusion from the study was the routing scheme should pay attention on successful data delivery from node to base station in addition to the energy-efficiency. The cluster based protocols are extended from static scenario to mobile scenario by various authors. None of the proposals addresses cluster head election appropriately in view of mobility. An elegant scheme for electing cluster heads is presented to meet the challenge of handling cluster durability when all the nodes in the network are moving. The scheme has been simulated and compared with a similar approach.The proliferation of sensor networks enables users with large set of sensor information to utilise them in various applications. The sensor network programming is inherently difficult due to various reasons. There must be an elegant way to collect the data gathered by sensor networks with out worrying about the underlying structure of the network. The final work presented addresses a way to collect data from a sensor network and present it to the users in a flexible way.A service oriented architecture based application is built and data collection task is presented as a web service. This will enable composition of sensor data from different sensor networks to build interesting applications. The main objective of the thesis was to design energy-efficient routing schemes for both static as well as mobile sensor networks. A progressive approach was followed to achieve this goal.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Polycrystalline single phasic mixed ferrites belonging to the series Ni1−xZnxFe2O4 for various values of x have been prepared by conventional ceramic techniques. Pre-characterized nickel zinc ferrites were then incorporated into a natural rubber matrix according to a specific recipe for various loadings. The processability and cure parameters were then determined. The magnetic properties of the ceramic filler as well as the ferrite loaded rubber ferrite composites (RFC) were evaluated and compared. A general equation for predicting the magnetic properties was also formulated. The validity of these equations were then checked and correlated with the experimental data. The coercivity of the RFCs almost resemble that of the ceramic component in the RFC. Percolation threshold is not reached for a maximum loading of 120 phr (parts per hundred rubber by weight) of the filler. These studies indicate that flexible magnets can be made with appropriate magnetic properties namely saturation magnetisation (Ms) and magnetic field strength (Hc) by a judicious choice of x and a corresponding loading. These studies also suggest that there is no possible interaction between the filler and the matrix at least at the macroscopic level. The formulated equation will aid in synthesizing RFCs with predetermined magnetic

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A Multi-Objective Antenna Placement Genetic Algorithm (MO-APGA) has been proposed for the synthesis of matched antenna arrays on complex platforms. The total number of antennas required, their position on the platform, location of loads, loading circuit parameters, decoupling and matching network topology, matching network parameters and feed network parameters are optimized simultaneously. The optimization goal was to provide a given minimum gain, specific gain discrimination between the main and back lobes and broadband performance. This algorithm is developed based on the non-dominated sorting genetic algorithm (NSGA-II) and Minimum Spanning Tree (MST) technique for producing diverse solutions when the number of objectives is increased beyond two. The proposed method is validated through the design of a wideband airborne SAR

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Soil fertility constraints to crop production have been recognized widely as a major obstacle to food security and agro-ecosystem sustainability in sub-Saharan West Africa. As such, they have led to a multitude of research projects and policy debates on how best they should be overcome. Conclusions, based on long-term multi-site experiments, are lacking with respect to a regional assessment of phosphorus and nitrogen fertilizer effects, surface mulched crop residues, and legume rotations on total dry matter of cereals in this region. A mixed model time-trend analysis was used to investigate the effects of four nitrogen and phosphorus rates, annually applied crop residue dry matter at 500 and 2000 kg ha^-1, and cereal-legume rotation versus continuous cereal cropping on the total dry matter of cereals and legumes. The multi-factorial experiment was conducted over four years at eight locations, with annual rainfall ranging from 510 to 1300 mm, in Niger, Burkina Faso, and Togo. With the exception of phosphorus, treatment effects on legume growth were marginal. At most locations, except for typical Sudanian sites with very low base saturation and high rainfall, phosphorus effects on cereal total dry matter were much lower with rock phosphate than with soluble phosphorus, unless the rock phosphate was combined with an annual seed-placement of 4 kg ha^-1 phosphorus. Across all other treatments, nitrogen effects were negligible at 500 mm annual rainfall but at 900 mm, the highest nitrogen rate led to total dry matter increases of up to 77% and, at 1300 mm, to 183%. Mulch-induced increases in cereal total dry matter were larger with lower base saturation, reaching 45% on typical acid sandy Sahelian soils. Legume rotation effects tended to increase over time but were strongly species-dependent.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The process of developing software that takes advantage of multiple processors is commonly referred to as parallel programming. For various reasons, this process is much harder than the sequential case. For decades, parallel programming has been a problem for a small niche only: engineers working on parallelizing mostly numerical applications in High Performance Computing. This has changed with the advent of multi-core processors in mainstream computer architectures. Parallel programming in our days becomes a problem for a much larger group of developers. The main objective of this thesis was to find ways to make parallel programming easier for them. Different aims were identified in order to reach the objective: research the state of the art of parallel programming today, improve the education of software developers about the topic, and provide programmers with powerful abstractions to make their work easier. To reach these aims, several key steps were taken. To start with, a survey was conducted among parallel programmers to find out about the state of the art. More than 250 people participated, yielding results about the parallel programming systems and languages in use, as well as about common problems with these systems. Furthermore, a study was conducted in university classes on parallel programming. It resulted in a list of frequently made mistakes that were analyzed and used to create a programmers' checklist to avoid them in the future. For programmers' education, an online resource was setup to collect experiences and knowledge in the field of parallel programming - called the Parawiki. Another key step in this direction was the creation of the Thinking Parallel weblog, where more than 50.000 readers to date have read essays on the topic. For the third aim (powerful abstractions), it was decided to concentrate on one parallel programming system: OpenMP. Its ease of use and high level of abstraction were the most important reasons for this decision. Two different research directions were pursued. The first one resulted in a parallel library called AthenaMP. It contains so-called generic components, derived from design patterns for parallel programming. These include functionality to enhance the locks provided by OpenMP, to perform operations on large amounts of data (data-parallel programming), and to enable the implementation of irregular algorithms using task pools. AthenaMP itself serves a triple role: the components are well-documented and can be used directly in programs, it enables developers to study the source code and learn from it, and it is possible for compiler writers to use it as a testing ground for their OpenMP compilers. The second research direction was targeted at changing the OpenMP specification to make the system more powerful. The main contributions here were a proposal to enable thread-cancellation and a proposal to avoid busy waiting. Both were implemented in a research compiler, shown to be useful in example applications, and proposed to the OpenMP Language Committee.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Distributed systems are one of the most vital components of the economy. The most prominent example is probably the internet, a constituent element of our knowledge society. During the recent years, the number of novel network types has steadily increased. Amongst others, sensor networks, distributed systems composed of tiny computational devices with scarce resources, have emerged. The further development and heterogeneous connection of such systems imposes new requirements on the software development process. Mobile and wireless networks, for instance, have to organize themselves autonomously and must be able to react to changes in the environment and to failing nodes alike. Researching new approaches for the design of distributed algorithms may lead to methods with which these requirements can be met efficiently. In this thesis, one such method is developed, tested, and discussed in respect of its practical utility. Our new design approach for distributed algorithms is based on Genetic Programming, a member of the family of evolutionary algorithms. Evolutionary algorithms are metaheuristic optimization methods which copy principles from natural evolution. They use a population of solution candidates which they try to refine step by step in order to attain optimal values for predefined objective functions. The synthesis of an algorithm with our approach starts with an analysis step in which the wanted global behavior of the distributed system is specified. From this specification, objective functions are derived which steer a Genetic Programming process where the solution candidates are distributed programs. The objective functions rate how close these programs approximate the goal behavior in multiple randomized network simulations. The evolutionary process step by step selects the most promising solution candidates and modifies and combines them with mutation and crossover operators. This way, a description of the global behavior of a distributed system is translated automatically to programs which, if executed locally on the nodes of the system, exhibit this behavior. In our work, we test six different ways for representing distributed programs, comprising adaptations and extensions of well-known Genetic Programming methods (SGP, eSGP, and LGP), one bio-inspired approach (Fraglets), and two new program representations called Rule-based Genetic Programming (RBGP, eRBGP) designed by us. We breed programs in these representations for three well-known example problems in distributed systems: election algorithms, the distributed mutual exclusion at a critical section, and the distributed computation of the greatest common divisor of a set of numbers. Synthesizing distributed programs the evolutionary way does not necessarily lead to the envisaged results. In a detailed analysis, we discuss the problematic features which make this form of Genetic Programming particularly hard. The two Rule-based Genetic Programming approaches have been developed especially in order to mitigate these difficulties. In our experiments, at least one of them (eRBGP) turned out to be a very efficient approach and in most cases, was superior to the other representations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The traditional task of a central bank is to preserve price stability and, in doing so, not to impair the real economy more than necessary. To meet this challenge, it is of great relevance whether inflation is only driven by inflation expectations and the current output gap or whether it is, in addition, influenced by past inflation. In the former case, as described by the New Keynesian Phillips curve, the central bank can immediately and simultaneously achieve price stability and equilibrium output, the so-called ‘divine coincidence’ (Blanchard and Galí 2007). In the latter case, the achievement of price stability is costly in terms of output and will be pursued over several periods. Similarly, it is important to distinguish this latter case, which describes ‘intrinsic’ inflation persistence, from that of ‘extrinsic’ inflation persistence, where the sluggishness of inflation is not a ‘structural’ feature of the economy but merely ‘inherited’ from the sluggishness of the other driving forces, inflation expectations and output. ‘Extrinsic’ inflation persistence is usually considered to be the less challenging case, as policy-makers are supposed to fight against the persistence in the driving forces, especially to reduce the stickiness of inflation expectations by a credible monetary policy, in order to reestablish the ‘divine coincidence’. The scope of this dissertation is to contribute to the vast literature and ongoing discussion on inflation persistence: Chapter 1 describes the policy consequences of inflation persistence and summarizes the empirical and theoretical literature. Chapter 2 compares two models of staggered price setting, one with a fixed two-period duration and the other with a stochastic duration of prices. I show that in an economy with a timeless optimizing central bank the model with the two-period alternating price-setting (for most parameter values) leads to more persistent inflation than the model with stochastic price duration. This result amends earlier work by Kiley (2002) who found that the model with stochastic price duration generates more persistent inflation in response to an exogenous monetary shock. Chapter 3 extends the two-period alternating price-setting model to the case of 3- and 4-period price durations. This results in a more complex Phillips curve with a negative impact of past inflation on current inflation. As simulations show, this multi-period Phillips curve generates a too low degree of autocorrelation and too early turnings points of inflation and is outperformed by a simple Hybrid Phillips curve. Chapter 4 starts from the critique of Driscoll and Holden (2003) on the relative real-wage model of Fuhrer and Moore (1995). While taking the critique seriously that Fuhrer and Moore’s model will collapse to a much simpler one without intrinsic inflation persistence if one takes their arguments literally, I extend the model by a term for inequality aversion. This model extension is not only in line with experimental evidence but results in a Hybrid Phillips curve with inflation persistence that is observably equivalent to that presented by Fuhrer and Moore (1995). In chapter 5, I present a model that especially allows to study the relationship between fairness attitudes and time preference (impatience). In the model, two individuals take decisions in two subsequent periods. In period 1, both individuals are endowed with resources and are able to donate a share of their resources to the other individual. In period 2, the two individuals might join in a common production after having bargained on the split of its output. The size of the production output depends on the relative share of resources at the end of period 1 as the human capital of the individuals, which is built by means of their resources, cannot fully be substituted one against each other. Therefore, it might be rational for a well-endowed individual in period 1 to act in a seemingly ‘fair’ manner and to donate own resources to its poorer counterpart. This decision also depends on the individuals’ impatience which is induced by the small but positive probability that production is not possible in period 2. As a general result, the individuals in the model economy are more likely to behave in a ‘fair’ manner, i.e., to donate resources to the other individual, the lower their own impatience and the higher the productivity of the other individual. As the (seemingly) ‘fair’ behavior is modelled as an endogenous outcome and as it is related to the aspect of time preference, the presented framework might help to further integrate behavioral economics and macroeconomics.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Evaluation of major feed resources was conducted in four crop-livestock mixed farming systems of central southern Ethiopia, with 90 farmers, selected using multi-stage purposive and random sampling methods. Discussions were held with focused groups and key informants for vernacular name identification of feed, followed by feed sampling to analyse chemical composition (CP, ADF and NDF), in-vitro dry matter digestibility (IVDMD), and correlate with indigenous technical knowledge (ITK). Native pastures, crop residues (CR) and multi-purpose trees (MPT) are the major feed resources, demonstrated great variations in seasonality, chemical composition and IVDMD. The average CP, NDF and IVDMD values for grasses were 83.8 (ranged: 62.9–190), 619 (ranged: 357–877) and 572 (ranged: 317–743) g kg^(−1) DM, respectively. Likewise, the average CP, NDF and IVDMD for CR were 58 (ranged: 20–90), 760 (ranged: 340–931) and 461 (ranged: 285–637)g kg^(−1) DM, respectively. Generally, the MPT and non-conventional feeds (NCF, Ensete ventricosum and Ipomoea batatas) possessed higher CP (ranged: 155–164 g kg^(−1) DM) and IVDMD values (611–657 g kg^(−1) DM) while lower NDF (331–387 g kg^(−1) DM) and ADF (321–344 g kg^(−1) DM) values. The MPT and NCF were ranked as the best nutritious feeds by ITK while crop residues were the least. This study indicates that there are remarkable variations within and among forage resources in terms of chemical composition. There were also complementarities between ITK and feed laboratory results, and thus the ITK need to be taken into consideration in evaluation of local feed resources.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis defines Pi, a parallel architecture interface that separates model and machine issues, allowing them to be addressed independently. This provides greater flexibility for both the model and machine builder. Pi addresses a set of common parallel model requirements including low latency communication, fast task switching, low cost synchronization, efficient storage management, the ability to exploit locality, and efficient support for sequential code. Since Pi provides generic parallel operations, it can efficiently support many parallel programming models including hybrids of existing models. Pi also forms a basis of comparison for architectural components.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The goal of this research is to develop the prototype of a tactile sensing platform for anthropomorphic manipulation research. We investigate this problem through the fabrication and simple control of a planar 2-DOF robotic finger inspired by anatomic consistency, self-containment, and adaptability. The robot is equipped with a tactile sensor array based on optical transducer technology whereby localized changes in light intensity within an illuminated foam substrate correspond to the distribution and magnitude of forces applied to the sensor surface plane. The integration of tactile perception is a key component in realizing robotic systems which organically interact with the world. Such natural behavior is characterized by compliant performance that can initiate internal, and respond to external, force application in a dynamic environment. However, most of the current manipulators that support some form of haptic feedback either solely derive proprioceptive sensation or only limit tactile sensors to the mechanical fingertips. These constraints are due to the technological challenges involved in high resolution, multi-point tactile perception. In this work, however, we take the opposite approach, emphasizing the role of full-finger tactile feedback in the refinement of manual capabilities. To this end, we propose and implement a control framework for sensorimotor coordination analogous to infant-level grasping and fixturing reflexes. This thesis details the mechanisms used to achieve these sensory, actuation, and control objectives, along with the design philosophies and biological influences behind them. The results of behavioral experiments with a simple tactilely-modulated control scheme are also described. The hope is to integrate the modular finger into an %engineered analog of the human hand with a complete haptic system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Most Artificial Intelligence (AI) work can be characterized as either ``high-level'' (e.g., logical, symbolic) or ``low-level'' (e.g., connectionist networks, behavior-based robotics). Each approach suffers from particular drawbacks. High-level AI uses abstractions that often have no relation to the way real, biological brains work. Low-level AI, on the other hand, tends to lack the powerful abstractions that are needed to express complex structures and relationships. I have tried to combine the best features of both approaches, by building a set of programming abstractions defined in terms of simple, biologically plausible components. At the ``ground level'', I define a primitive, perceptron-like computational unit. I then show how more abstract computational units may be implemented in terms of the primitive units, and show the utility of the abstract units in sample networks. The new units make it possible to build networks using concepts such as long-term memories, short-term memories, and frames. As a demonstration of these abstractions, I have implemented a simulator for ``creatures'' controlled by a network of abstract units. The creatures exist in a simple 2D world, and exhibit behaviors such as catching mobile prey and sorting colored blocks into matching boxes. This program demonstrates that it is possible to build systems that can interact effectively with a dynamic physical environment, yet use symbolic representations to control aspects of their behavior.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The underlying assumptions for interpreting the meaning of data often change over time, which further complicates the problem of semantic heterogeneities among autonomous data sources. As an extension to the COntext INterchange (COIN) framework, this paper introduces the notion of temporal context as a formalization of the problem. We represent temporal context as a multi-valued method in F-Logic; however, only one value is valid at any point in time, the determination of which is constrained by temporal relations. This representation is then mapped to an abductive constraint logic programming framework with temporal relations being treated as constraints. A mediation engine that implements the framework automatically detects and reconciles semantic differences at different times. We articulate that this extended COIN framework is suitable for reasoning on the Semantic Web.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The underlying assumptions for interpreting the meaning of data often change over time, which further complicates the problem of semantic heterogeneities among autonomous data sources. As an extension to the COntext INterchange (COIN) framework, this paper introduces the notion of temporal context as a formalization of the problem. We represent temporal context as a multi-valued method in F-Logic; however, only one value is valid at any point in time, the determination of which is constrained by temporal relations. This representation is then mapped to an abductive constraint logic programming framework with temporal relations being treated as constraints. A mediation engine that implements the framework automatically detects and reconciles semantic differences at different times. We articulate that this extended COIN framework is suitable for reasoning on the Semantic Web.