170 resultados para Multi-stage programming
Resumo:
Several standards appeared in recent years to formalize the metadata of learning objects, but they are still insufficient to fully describe a specialized domain. In particular, the programming exercise domain requires interdependent resources (e.g. test cases, solution programs, exercise description) usually processed by different services in the programming exercise life-cycle. Moreover, the manual creation of these resources is time-consuming and error-prone leading to what is an obstacle to the fast development of programming exercises of good quality. This paper focuses on the definition of an XML dialect called PExIL (Programming Exercises Interoperability Language). The aim of PExIL is to consolidate all the data required in the programming exercise life-cycle, from when it is created to when it is graded, covering also the resolution, the evaluation and the feedback. We introduce the XML Schema used to formalize the relevant data of the programming exercise life-cycle. The validation of this approach is made through the evaluation of the usefulness and expressiveness of the PExIL definition. In the former we present the tools that consume the PExIL definition to automatically generate the specialized resources. In the latter we use the PExIL definition to capture all the constraints of a set of programming exercises stored in a learning objects repository.
Resumo:
Learning computer programming requires solving programming exercises. In computer programming courses teachers need to assess and give feedback to a large number of exercises. These tasks are time consuming and error-prone since there are many aspects relating to good programming that should be considered. In this context automatic assessment tools can play an important role helping teachers in grading tasks as well to assist students with automatic feedback. In spite of its usefulness, these tools lack integration mechanisms with other eLearning systems such as Learning Management Systems, Learning Objects Repositories or Integrated Development Environments. In this paper we provide a survey on programming evaluation systems. The survey gathers information on interoperability features of these systems, categorizing and comparing them regarding content and communication standardization. This work may prove useful to instructors and computer science educators when they have to choose an assessment system to be integrated in their e-Learning environment.
Resumo:
This paper presents a tool called Petcha that acts as an automated Teaching Assistant in computer programming courses. The ultimate objective of Petcha is to increase the number of programming exercises effectively solved by students. Petcha meets this objective by helping both teachers to author programming exercises and students to solve them. It also coordinates a network of heterogeneous systems, integrating automatic program evaluators, learning management systems, learning object repositories and integrated programming environments. This paper presents the concept and the design of Petcha and sets this tool in a service oriented architecture for managing learning processes based on the automatic evaluation of programming exercises. The paper presents also a case study that validates the use of Petcha and of the proposed architecture.
Resumo:
Assessment plays a vital role in learning. This is certainly the case with assessment of computer programs, both in curricular and competitive learning. The lack of a standard – or at least a widely used format – creates a modern Ba- bel tower made of Learning Objects, of assessment items that cannot be shared among automatic assessment systems. These systems whose interoperability is hindered by the lack of a common format include contest management systems, evaluation engines, repositories of learning objects and authoring tools. A prag- matical approach to remedy this problem is to create a service to convert among existing formats. A kind of translation service specialized in programming prob- lems formats. To convert programming exercises on-the-fly among the most used formats is the purpose of the BabeLO – a service to cope with the existing Babel of Learning Object formats for programming exercises. BabeLO was designed as a service to act as a middleware in a network of systems typically used in auto- matic assessment of programs. It provides support for multiple exercise formats and can be used by: evaluation engines to assess exercises regardless of its format; repositories to import exercises from various sources; authoring systems to create exercises in multiple formats or based on exercises from other sources. This paper analyses several of existing formats to highlight both their differ- ences and their similar features. Based on this analysis it presents an approach to extensible format conversion. It presents also the features of PExIL, the pivotal format in which the conversion is based; and the function definitions of the proposed service – BabeLO. Details on the design and implementation of BabeLO, including the service API and the interfaces required to extend the conversion to a new format, are also provided. To evaluate the effectiveness and efficiency of this approach this paper reports on two actual uses of BabeLO: to relocate exercises to a different repository; and to use an evaluation engine in a network of heterogeneous systems.
Resumo:
Several standards have appeared in recent years to formalize the metadata of learning objects, but they are still insufficient to fully describe a specialized domain. In particular, the programming exercise domain requires interdependent resources (e.g. test cases, solution programs, exercise description) usually processed by different services in the programming exercise lifecycle. Moreover, the manual creation of these resources is time-consuming and error-prone, leading to an obstacle to the fast development of programming exercises of good quality. This chapter focuses on the definition of an XML dialect called PExIL (Programming Exercises Interoperability Language). The aim of PExIL is to consolidate all the data required in the programming exercise lifecycle from when it is created to when it is graded, covering also the resolution, the evaluation, and the feedback. The authors introduce the XML Schema used to formalize the relevant data of the programming exercise lifecycle. The validation of this approach is made through the evaluation of the usefulness and expressiveness of the PExIL definition. In the former, the authors present the tools that consume the PExIL definition to automatically generate the specialized resources. In the latter, they use the PExIL definition to capture all the constraints of a set of programming exercises stored in a learning objects repository.
Resumo:
The development and applications of thermoset polymeric composites, namely fibre reinforced plastics (FRP), have shifted in the last decades more and more into the mass market [1]. Despite of all advantages associated to FRP based products, the increasing production and consume also lead to an increasing amount of FRP wastes, either end-of-lifecycle products, or scrap and by-products generated by the manufacturing process itself. Whereas thermoplastic FRPs can be easily recycled, by remelting and remoulding, recyclability of thermosetting FRPs constitutes a more difficult task due to cross-linked nature of resin matrix. To date, most of the thermoset based FRP waste is being incinerated or landfilled, leading to negative environmental impacts and supplementary added costs to FRP producers and suppliers. This actual framework is putting increasing pressure on the industry to address the options available for FRP waste management, being an important driver for applied research undertaken cost efficient recycling methods. [1-2]. In spite of this, research on recycling solutions for thermoset composites is still at an elementary stage. Thermal and/or chemical recycling processes, with partial fibre recovering, have been investigated mostly for carbon fibre reinforced plastics (CFRP) due to inherent value of carbon fibre reinforcement; whereas for glass fibre reinforced plastics (GFRP), mechanical recycling, by means of milling and grinding processes, has been considered a more viable recycling method [1-2]. Though, at the moment, few solutions in the reuse of mechanically-recycled GFRP composites into valueadded products are being explored. Aiming filling this gap, in this study, a new waste management solution for thermoset GFRP based products was assessed. The mechanical recycling approach, with reduction of GFRP waste to powdered and fibrous materials was applied, and the potential added value of obtained recyclates was experimentally investigated as raw material for polyester based mortars. The use of a cementless concrete as host material for GFRP recyclates, instead of a conventional Portland cement based concrete, presents an important asset in avoiding the eventual incompatibility problems arisen from alkalis silica reaction between glass fibres and cementious binder matrix. Additionally, due to hermetic nature of resin binder, polymer based concretes present greater ability for incorporating recycled waste products [3]. Under this scope, different GFRP waste admixed polymer mortar (PM) formulations were analyzed varying the size grading and content of GFRP powder and fibre mix waste. Added value of potential recycling solution was assessed by means of flexural and compressive loading capacities of modified mortars with regard to waste-free polymer mortars.
Resumo:
No âmbito da unidade curricular Dissertação/Projeto/Estágio do 2º ano do Mestrado em Engenharia mecânica – Ramo Gestão Industrial do Instituto Superior de Engenharia do Porto, o presente trabalho de dissertação foi enquadrado num projeto de desenvolvimento de ferramentas de apoio à gestão de projetos. O projeto foi desenvolvido no Instituto de Engenharia Mecânica e Gestão Industrial (INEGI) na unidade de Desenvolvimento de Produto e Sistemas (DPS). A realização deste projeto teve como objetivo o desenvolvimento e adequação de ferramentas de apoio à gestão de multi-projeto no processo de desenvolvimento de produto na organização em estudo – o INEGI – DPS. A gestão de projetos tem hoje uma grande importância nos resultados das empresas essencialmente em virtude da necessidade de estas competirem num mundo em grande mudança com concorrentes ferozes, em que a capacidade de responder às mudanças a tempo e de uma forma integrada se torna cada vez mais importante. A atividade levada a cabo pela DPS impõe a necessidade de uma gestão de projetos mais eficaz e eficiente suportada numa gestão de informação centralizada. O presente projeto de investigação teve, numa primeira fase, uma adaptação à organização em estudo. De seguida, foi conduzida uma revisão da literatura com o objetivo de se obter a fundamentação teórica necessária ao desenvolvimento de ferramentas com base nas metodologias lean. Prosseguiu com o levantamento da situação inicial da organização e com a identificação dos problemas existentes na gestão de projetos. Incluiu também uma revisão e análise das ferramentas existentes na unidade em estudo. Este conhecimento permitiu delinear uma visão para guiar o desenvolvimento das ferramentas. Após a definição da visão foi, então, realizado o desenvolvimento das ferramentas de auxílio à gestão multi-projeto na organização. A concretização deste trabalho resultou no desenvolvimento de três ferramentas de auxílio à gestão multi-projeto na unidade. Estas ferramentas tornam o processo de gestão de projetos mais simples e fácil de assimilar, requerendo apenas alguns inputs por parte dos colaboradores. Estas ferramentas estão apoiadas nos pilares do lean, e deste modo estão vocacionadas para reduzir o desperdício, promover a melhoria contínua, aumentar o desempenho global dos vários atores nos projetos de modo a entregar mais valor e qualidade superior com menores custos. Acima de tudo, valorizar o trabalho dos colaboradores, tornando-os mais eficientes, eficazes, motivados e comprometidos com a organização.
Resumo:
The current ubiquitous network access and increase in network bandwidth are driving the sales of mobile location-aware user devices and, consequently, the development of context-aware applications, namely location-based services. The goal of this project is to provide consumers of location-based services with a richer end-user experience by means of service composition, personalization, device adaptation and continuity of service. Our approach relies on a multi-agent system composed of proxy agents that act as mediators and providers of personalization meta-services, device adaptation and continuity of service for consumers of pre-existing location-based services. These proxy agents, which have Web services interfaces to ensure a high level of interoperability, perform service composition and take in consideration the preferences of the users, the limitations of the user devices, making the usage of different types of devices seamless for the end-user. To validate and evaluate the performance of this approach, use cases were defined, tests were conducted and results gathered which demonstrated that the initial goals were successfully fulfilled.
Resumo:
Decentralised co-operative multi-agent systems are computational systems where conflicts are frequent due to the nature of the represented knowledge. Negotiation methodologies, in this case argumentation based negotiation methodologies, were developed and applied to solve unforeseeable and, therefore, unavoidable conflicts. The supporting computational model is a distributed belief revision system where argumentation plays the decisive role of revision. The distributed belief revision system detects, isolates and solves, whenever possible, the identified conflicts. The detection and isolation of the conflicts is automatically performed by the distributed consistency mechanism and the resolution of the conflict, or belief revision, is achieved via argumentation. We propose and describe two argumentation protocols intended to solve different types of identified information conflicts: context dependent and context independent conflicts. While the protocol for context dependent conflicts generates new consensual alternatives, the latter chooses to adopt the soundest, strongest argument presented. The paper shows the suitability of using argumentation as a distributed decentralised belief revision protocol to solve unavoidable conflicts.
Resumo:
The environmental management domain is vast and encompasses many identifiable activities: impact assessment, planning, project evaluation, etc. In particular, this paper focusses on the modelling of the project evaluation activity. The environmental decision support system under development aims to provide assistance to project developers in the selection of adequate locations, guaranteeing the compliance with the applicable regulations and the existing development plans as well as satisfying the specified project requirements. The inherent multidisciplinarity features of this activity lead to the adoption of the Multi-Agent paradigm, and, in particular, to the modelling of the involved agencies as a community of cooperative autonomous agents, where each agency contributes with its share of problem solving to the final system’s recommendation. To achieve this behaviour the many conclusions of the individual agencies have to be justifiably accommodated: not only they may differ, but can be interdependent, complementary, irreconcilable, or simply, independent. We propose different solutions (involving both local and global consistency) to support the adequate merge of the distinct perspectives that inevitably arise during this type of decision making.
Resumo:
The ability to respond sensibly to changing and conflicting beliefs is an integral part of intelligent agency. To this end, we outline the design and implementation of a Distributed Assumption-based Truth Maintenance System (DATMS) appropriate for controlling cooperative problem solving in a dynamic real world multi-agent community. Our DATMS works on the principle of local coherence which means that different agents can have different perspectives on the same fact provided that these stances are appropriately justified. The belief revision algorithm is presented, the meta-level code needed to ensure that all system-wide queries can be uniquely answered is described, and the DATMS’ implementation in a general purpose multi-agent shell is discussed.
Resumo:
Multi-agent architectures are well suited for complex inherently distributed problem solving domains. From the many challenging aspects that arise within this framework, a crucial one emerges: how to incorporate dynamic and conflicting agent beliefs? While the belief revision activity in a single agent scenario is concentrated on incorporating new information while preserving consistency, in a multi-agent system it also has to deal with possible conflicts between the agents perspectives. To provide an adequate framework, each agent, built as a combination of an assumption based belief revision system and a cooperation layer, was enriched with additional features: a distributed search control mechanism allowing dynamic context management, and a set of different distributed consistency methodologies. As a result, a Distributed Belief Revision Testbed (DiBeRT) was developed. This paper is a preliminary report presenting some of DiBeRT contributions: a concise representation of external beliefs; a simple and innovative methodology to achieve distributed context management; and a reduced inter-agent data exchange format.
Resumo:
In a real world multiagent system, where the agents are faced with partial, incomplete and intrinsically dynamic knowledge, conflicts are inevitable. Frequently, different agents have goals or beliefs that cannot hold simultaneously. Conflict resolution methodologies have to be adopted to overcome such undesirable occurrences. In this paper we investigate the application of distributed belief revision techniques as the support for conflict resolution in the analysis of the validity of the candidate beams to be produced in the CERN particle accelerators. This CERN multiagent system contains a higher hierarchy agent, the Specialist agent, which makes use of meta-knowledge (on how the con- flicting beliefs have been produced by the other agents) in order to detect which beliefs should be abandoned. Upon solving a conflict, the Specialist instructs the involved agents to revise their beliefs accordingly. Conflicts in the problem domain are mapped into conflicting beliefs of the distributed belief revision system, where they can be handled by proven formal methods. This technique builds on well established concepts and combines them in a new way to solve important problems. We find this approach generally applicable in several domains.
Resumo:
Generally, the evolution process of applications has impact on their underlining data models, thus becoming a time-consuming problem for programmers and database administrators. In this paper we address this problem within an aspect-oriented approach, which is based on a meta-model for orthogonal persistent programming systems. Applying reflection techniques, our meta-model aims to be simpler than its competitors. Furthermore, it enables database multi-version schemas. We also discuss two case studies in order to demonstrate the advantages of our approach.
Resumo:
Smart grids with an intensive penetration of distributed energy resources will play an important role in future power system scenarios. The intermittent nature of renewable energy sources brings new challenges, requiring an efficient management of those sources. Additional storage resources can be beneficially used to address this problem; the massive use of electric vehicles, particularly of vehicle-to-grid (usually referred as gridable vehicles or V2G), becomes a very relevant issue. This paper addresses the impact of Electric Vehicles (EVs) in system operation costs and in power demand curve for a distribution network with large penetration of Distributed Generation (DG) units. An efficient management methodology for EVs charging and discharging is proposed, considering a multi-objective optimization problem. The main goals of the proposed methodology are: to minimize the system operation costs and to minimize the difference between the minimum and maximum system demand (leveling the power demand curve). The proposed methodology perform the day-ahead scheduling of distributed energy resources in a distribution network with high penetration of DG and a large number of electric vehicles. It is used a 32-bus distribution network in the case study section considering different scenarios of EVs penetration to analyze their impact in the network and in the other energy resources management.