850 resultados para Practice Domain Framework


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Within the information systems field, the task of conceptual modeling involves building a representation of selected phenomena in some domain. High-quality conceptual-modeling work is important because it facilitates early detection and correction of system development errors. It also plays an increasingly important role in activities like business process reengineering and documentation of best-practice data and process models in enterprise resource planning systems. Yet little research has been undertaken on many aspects of conceptual modeling. In this paper, we propose a framework to motivate research that addresses the following fundamental question: How can we model the world to better facilitate our developing, implementing, using, and maintaining more valuable information systems? The framework comprises four elements: conceptual-modeling grammars, conceptual-modeling methods, conceptual-modeling scripts, and conceptual-modeling contexts. We provide examples of the types of research that have already been undertaken on each element and illustrate research opportunities that exist.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper proposes a template for modelling complex datasets that integrates traditional statistical modelling approaches with more recent advances in statistics and modelling through an exploratory framework. Our approach builds on the well-known and long standing traditional idea of 'good practice in statistics' by establishing a comprehensive framework for modelling that focuses on exploration, prediction, interpretation and reliability assessment, a relatively new idea that allows individual assessment of predictions. The integrated framework we present comprises two stages. The first involves the use of exploratory methods to help visually understand the data and identify a parsimonious set of explanatory variables. The second encompasses a two step modelling process, where the use of non-parametric methods such as decision trees and generalized additive models are promoted to identify important variables and their modelling relationship with the response before a final predictive model is considered. We focus on fitting the predictive model using parametric, non-parametric and Bayesian approaches. This paper is motivated by a medical problem where interest focuses on developing a risk stratification system for morbidity of 1,710 cardiac patients given a suite of demographic, clinical and preoperative variables. Although the methods we use are applied specifically to this case study, these methods can be applied across any field, irrespective of the type of response.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Testamentary capacity (the capacity to make a will) is recognised in the literature as an important issue for speech-language pathologists' assessment of people with aphasia, but current guidelines for clinical practice lack an empirical base. Aims: The research aimed to suggest some guidelines for clinical practice based on information considered relevant for the court in determining testamentary capacity. Methods & Procedures: A recent legal case involving a challenge to the will of a woman with severe aphasia was critically examined with reference to current guidelines in the literature regarding assessment of testamentary capacity. Outcomes & Results: Examination of the information available on the case indicated that the judge gave priority to accounts of the everyday communication of the person with aphasia (including reported discourse samples) over the information provided by expert medical witnesses. The extent to which communication effectiveness could be maximised was found to be a matter of key significance to the determination of capacity. Conclusions: This study has implications for speech-language pathologists' assessment practices and reports, as well as for scope of practice with regard to legal decision making of people with aphasia. These issues are discussed in relation to the World Health Organisation's ICF framework of functioning for social participation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ontologies are becoming an important mechanism to build information systems. Nevertheless, there is still no systematic approach to support the design of such systems using tools that are common to information systems developers. In this paper, we propose an approach for deriving object frameworks from domain ontologies and then we show the application of this approach in the software process domain.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a taxonomy able to contribute to building a framework within the domain of Virtual Enterprises (VE). A VE taxonomy currently does not exist, and this lack is felt in the ambiguous way that some concepts are addressed, leading to a fragment understanding that hinders the development of the science of VE integration and management. The structure of the taxonomy developed is based on the view of the system as a 5-tuple consisting of Input, Control, Output, Mechanism, and Process, which is the underlying system-view in the well-know IDEF0 diagramming technique. In particular, this taxonomy addresses the VE extended lifecycle that implies the use of a meta-organization called Market of Resources, as an original contribution to the VE theory and practice. The taxonomy presented is constructed in a way to be easily complemented with other VE partial taxonomies that may be found in literature.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The purpose of this paper is to present a taxonomy able to contribute to building a framework within the domain of Virtual Enterprises (VE), to facilitate the sharing of knowledge and contributions to knowledge, as well as for trust building among VE stakeholders. A VE taxonomy currently does not exist, and this lack is felt in the ambiguous way that some concepts are addressed, leading to a fragment understanding that hinders the development of the science of VE integration and management. The structure of the taxonomy developed is based on the view of the system as a 5-tuple consisting of Input, Control, Output, Mechanism, and Process, which is the underlying system-view in the well-know IDEF0 diagramming technique. In particular, this taxonomy addresses the VE extended lifecycle that implies the use of a meta-organization called Market of Resources, as an original contribution to the VE theory and practice. The taxonomy presented does not repeat what the literature already includes, or the commonplaces, and it is constructed in a way to be easily complemented with other VE partial taxonomies that may be found in literature. Some suggestions for extensions to other interrelated domains (as evolution leaves taxonomies in an open or incompleteness state) are given in the text.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Purpose/objectives: This paper seeks to investigate whether performance management (PM) framework adopted in Portuguese local government (PLG) fit the Otley’s PM framework (1999). In particularly, the research questions are (1) whether PM framework adopted in PLG (SIADAP) fit the Otley´s framework, and (2) how local politicians (aldermen) see the operation of performance management systems (PMS) in PLG (focusing on the goal-setting process and incentive and reward structures). Theoretical positioning/contributions: With this paper we intend to contribute to literature on how the Otley’s PM framework can guide empirical research about the operation of PMS. In particular, the paper contributes to understand the fit between PMS implemented in PLG and the Otley´s PM framework. The analysis of this fit can be a good contribution to understand if PMS are used in PLG as a management tool or as a strategic response to external pressures (based on interviews conducted to aldermen). We believe that the Otley’s PM framework, as well as the extended PM framework presented by Ferreira and Otley (2009), can provide a useful research tool to understand the operation of PMS in PLG. Research method: The first research question is the central issue in this paper and is analyzed based on the main reforms introduced by Portuguese government on PM of public organizations (like municipalities). On the other hand, interviews conducted on three larger Portuguese municipalities (Oporto, Braga, and Matosinhos) show how aldermen see the operation of PMS in PLG, highlighting the goals setting process with targets associated and the existing of incentive and reward structures linked with performance. Findings: Generally we find that formal and regulated PM frameworks in PLG fit the main issues of the Otley’s PM framework. However, regarding the aldermen perceptions about PMS in practice we find a gap between theory and practice, especially regarding difficulties associated with the lack of a culture of goals and targets setting and the lack of incentive and reward structures linked with performance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recently, several distributed video coding (DVC) solutions based on the distributed source coding (DSC) paradigm have appeared in the literature. Wyner-Ziv (WZ) video coding, a particular case of DVC where side information is made available at the decoder, enable to achieve a flexible distribution of the computational complexity between the encoder and decoder, promising to fulfill novel requirements from applications such as video surveillance, sensor networks and mobile camera phones. The quality of the side information at the decoder has a critical role in determining the WZ video coding rate-distortion (RD) performance, notably to raise it to a level as close as possible to the RD performance of standard predictive video coding schemes. Towards this target, efficient motion search algorithms for powerful frame interpolation are much needed at the decoder. In this paper, the RD performance of a Wyner-Ziv video codec is improved by using novel, advanced motion compensated frame interpolation techniques to generate the side information. The development of these type of side information estimators is a difficult problem in WZ video coding, especially because the decoder only has available some reference, decoded frames. Based on the regularization of the motion field, novel side information creation techniques are proposed in this paper along with a new frame interpolation framework able to generate higher quality side information at the decoder. To illustrate the RD performance improvements, this novel side information creation framework has been integrated in a transform domain turbo coding based Wyner-Ziv video codec. Experimental results show that the novel side information creation solution leads to better RD performance than available state-of-the-art side information estimators, with improvements up to 2 dB: moreover, it allows outperforming H.264/AVC Intra by up to 3 dB with a lower encoding complexity.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

E-Learning frameworks are conceptual tools to organize networks of elearning services. Most frameworks cover areas that go beyond the scope of e-learning, from course to financial management, and neglects the typical activities in everyday life of teachers and students at schools such as the creation, delivery, resolution and evaluation of assignments. This paper presents the Ensemble framework - an e-learning framework exclusively focused on the teaching-learning process through the coordination of pedagogical services. The framework presents an abstract data, integration and evaluation model based on content and communications specifications. These specifications must base the implementation of networks in specialized domains with complex evaluations. In this paper we specialize the framework for two domains with complex evaluation: computer programming and computer-aided design (CAD). For each domain we highlight two Ensemble hotspots: data and evaluations procedures. In the former we formally describe the exercise and present possible extensions. In the latter, we describe the automatic evaluation procedures.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Embedded real-time applications increasingly present high computation requirements, which need to be completed within specific deadlines, but that present highly variable patterns, depending on the set of data available in a determined instant. The current trend to provide parallel processing in the embedded domain allows providing higher processing power; however, it does not address the variability in the processing pattern. Dimensioning each device for its worst-case scenario implies lower average utilization, and increased available, but unusable, processing in the overall system. A solution for this problem is to extend the parallel execution of the applications, allowing networked nodes to distribute the workload, on peak situations, to neighbour nodes. In this context, this report proposes a framework to develop parallel and distributed real-time embedded applications, transparently using OpenMP and Message Passing Interface (MPI), within a programming model based on OpenMP. The technical report also devises an integrated timing model, which enables the structured reasoning on the timing behaviour of these hybrid architectures.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Apresentação no âmbito da Dissertação de Mestrado Orientador: Doutora Alcina Dias

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work is a contribution to the e-Framework, arguably the most prominent e-learning framework today, and consists of the definition of a service for the automatic evaluation of programming exercises. This evaluation domain differs from trivial evaluations modelled by languages such as the IMS Question & Test Interoperability (QTI) specification. Complex evaluation domains justify the development of specialized evaluators that participate in several business processes. These business processes can combine other type of systems such as Programming Contest Management Systems, Learning Management Systems, Integrated Development Environments and Learning Object Repositories where programming exercises are stored as Learning Objects. This contribution describes the implementation approaches used, more precisely, behaviours & requests, use & interactions, applicable standards, interface definition and usage scenarios.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Mestrado em Computação e Instrumentação Médica

Relevância:

30.00% 30.00%

Publicador:

Resumo:

To meet the increasing demands of the complex inter-organizational processes and the demand for continuous innovation and internationalization, it is evident that new forms of organisation are being adopted, fostering more intensive collaboration processes and sharing of resources, in what can be called collaborative networks (Camarinha-Matos, 2006:03). Information and knowledge are crucial resources in collaborative networks, being their management fundamental processes to optimize. Knowledge organisation and collaboration systems are thus important instruments for the success of collaborative networks of organisations having been researched in the last decade in the areas of computer science, information science, management sciences, terminology and linguistics. Nevertheless, research in this area didn’t give much attention to multilingual contexts of collaboration, which pose specific and challenging problems. It is then clear that access to and representation of knowledge will happen more and more on a multilingual setting which implies the overcoming of difficulties inherent to the presence of multiple languages, through the use of processes like localization of ontologies. Although localization, like other processes that involve multilingualism, is a rather well-developed practice and its methodologies and tools fruitfully employed by the language industry in the development and adaptation of multilingual content, it has not yet been sufficiently explored as an element of support to the development of knowledge representations - in particular ontologies - expressed in more than one language. Multilingual knowledge representation is then an open research area calling for cross-contributions from knowledge engineering, terminology, ontology engineering, cognitive sciences, computational linguistics, natural language processing, and management sciences. This workshop joined researchers interested in multilingual knowledge representation, in a multidisciplinary environment to debate the possibilities of cross-fertilization between knowledge engineering, terminology, ontology engineering, cognitive sciences, computational linguistics, natural language processing, and management sciences applied to contexts where multilingualism continuously creates new and demanding challenges to current knowledge representation methods and techniques. In this workshop six papers dealing with different approaches to multilingual knowledge representation are presented, most of them describing tools, approaches and results obtained in the development of ongoing projects. In the first case, Andrés Domínguez Burgos, Koen Kerremansa and Rita Temmerman present a software module that is part of a workbench for terminological and ontological mining, Termontospider, a wiki crawler that aims at optimally traverse Wikipedia in search of domainspecific texts for extracting terminological and ontological information. The crawler is part of a tool suite for automatically developing multilingual termontological databases, i.e. ontologicallyunderpinned multilingual terminological databases. In this paper the authors describe the basic principles behind the crawler and summarized the research setting in which the tool is currently tested. In the second paper, Fumiko Kano presents a work comparing four feature-based similarity measures derived from cognitive sciences. The purpose of the comparative analysis presented by the author is to verify the potentially most effective model that can be applied for mapping independent ontologies in a culturally influenced domain. For that, datasets based on standardized pre-defined feature dimensions and values, which are obtainable from the UNESCO Institute for Statistics (UIS) have been used for the comparative analysis of the similarity measures. The purpose of the comparison is to verify the similarity measures based on the objectively developed datasets. According to the author the results demonstrate that the Bayesian Model of Generalization provides for the most effective cognitive model for identifying the most similar corresponding concepts existing for a targeted socio-cultural community. In another presentation, Thierry Declerck, Hans-Ulrich Krieger and Dagmar Gromann present an ongoing work and propose an approach to automatic extraction of information from multilingual financial Web resources, to provide candidate terms for building ontology elements or instances of ontology concepts. The authors present a complementary approach to the direct localization/translation of ontology labels, by acquiring terminologies through the access and harvesting of multilingual Web presences of structured information providers in the field of finance, leading to both the detection of candidate terms in various multilingual sources in the financial domain that can be used not only as labels of ontology classes and properties but also for the possible generation of (multilingual) domain ontologies themselves. In the next paper, Manuel Silva, António Lucas Soares and Rute Costa claim that despite the availability of tools, resources and techniques aimed at the construction of ontological artifacts, developing a shared conceptualization of a given reality still raises questions about the principles and methods that support the initial phases of conceptualization. These questions become, according to the authors, more complex when the conceptualization occurs in a multilingual setting. To tackle these issues the authors present a collaborative platform – conceptME - where terminological and knowledge representation processes support domain experts throughout a conceptualization framework, allowing the inclusion of multilingual data as a way to promote knowledge sharing and enhance conceptualization and support a multilingual ontology specification. In another presentation Frieda Steurs and Hendrik J. Kockaert present us TermWise, a large project dealing with legal terminology and phraseology for the Belgian public services, i.e. the translation office of the ministry of justice, a project which aims at developing an advanced tool including expert knowledge in the algorithms that extract specialized language from textual data (legal documents) and whose outcome is a knowledge database including Dutch/French equivalents for legal concepts, enriched with the phraseology related to the terms under discussion. Finally, Deborah Grbac, Luca Losito, Andrea Sada and Paolo Sirito report on the preliminary results of a pilot project currently ongoing at UCSC Central Library, where they propose to adapt to subject librarians, employed in large and multilingual Academic Institutions, the model used by translators working within European Union Institutions. The authors are using User Experience (UX) Analysis in order to provide subject librarians with a visual support, by means of “ontology tables” depicting conceptual linking and connections of words with concepts presented according to their semantic and linguistic meaning. The organizers hope that the selection of papers presented here will be of interest to a broad audience, and will be a starting point for further discussion and cooperation.