10 resultados para Common Assessment Framework (CAF)
em Instituto Politécnico do Porto, Portugal
Resumo:
Apresentação no âmbito da Dissertação de Mestrado Orientador: Doutora Alcina Dias
Resumo:
One of the most difficult issues of e-Learning is the students’ assessment. Being this an outstanding task regarding theoretical topics, it becomes even more challenging when the topics under evaluation are practical. ISCAP’s Information Systems Department is composed of about twenty teachers who have been for several years using an e-learning environment (at the moment Moodle 2.3) combined with traditional assessment. They are now planning and implementing a new e-learning assessment strategy. This effort was undertaken in order to evaluate a practical topic (the use of spreadsheets to solve management problems) common to shared courses of several undergraduate degree programs. The same team group is already experienced in the assessment of theoretical information systems topics using the b-learning platform. Therefore, this project works as an extension to previous experiences being the team aware of the additional difficulties due to the practical nature of the topics. This paper describes this project and presents two cycles of the action research methodology, used to conduct the research. The first cycle goal was to produce a database of questions. When it was implemented in order to be used with a pilot group of students, several problems were identified. Subsequently, the second cycle consisted in solving the identified problems preparing the database and all the players to a broader scope implementation. For each cycle, all the phases, its drawbacks and achievements are described. This paper suits all those who are or are planning to be in the process of shifting their assessment strategy from a traditional to one supported by an e-learning platform.
RadiaLE: A framework for designing and assessing link quality estimators in wireless sensor networks
Resumo:
Stringent cost and energy constraints impose the use of low-cost and low-power radio transceivers in large-scale wireless sensor networks (WSNs). This fact, together with the harsh characteristics of the physical environment, requires a rigorous WSN design. Mechanisms for WSN deployment and topology control, MAC and routing, resource and mobility management, greatly depend on reliable link quality estimators (LQEs). This paper describes the RadiaLE framework, which enables the experimental assessment, design and optimization of LQEs. RadiaLE comprises (i) the hardware components of the WSN testbed and (ii) a software tool for setting-up and controlling the experiments, automating link measurements gathering through packets-statistics collection, and analyzing the collected data, allowing for LQEs evaluation. We also propose a methodology that allows (i) to properly set different types of links and different types of traffic, (ii) to collect rich link measurements, and (iii) to validate LQEs using a holistic and unified approach. To demonstrate the validity and usefulness of RadiaLE, we present two case studies: the characterization of low-power links and a comparison between six representative LQEs. We also extend the second study for evaluating the accuracy of the TOSSIM 2 channel model.
Resumo:
Wireless Sensor Networks (WSN) are being used for a number of applications involving infrastructure monitoring, building energy monitoring and industrial sensing. The difficulty of programming individual sensor nodes and the associated overhead have encouraged researchers to design macro-programming systems which can help program the network as a whole or as a combination of subnets. Most of the current macro-programming schemes do not support multiple users seamlessly deploying diverse applications on the same shared sensor network. As WSNs are becoming more common, it is important to provide such support, since it enables higher-level optimizations such as code reuse, energy savings, and traffic reduction. In this paper, we propose a macro-programming framework called Nano-CF, which, in addition to supporting in-network programming, allows multiple applications written by different programmers to be executed simultaneously on a sensor networking infrastructure. This framework enables the use of a common sensing infrastructure for a number of applications without the users having to worrying about the applications already deployed on the network. The framework also supports timing constraints and resource reservations using the Nano-RK operating system. Nano- CF is efficient at improving WSN performance by (a) combining multiple user programs, (b) aggregating packets for data delivery, and (c) satisfying timing and energy specifications using Rate- Harmonized Scheduling. Using representative applications, we demonstrate that Nano-CF achieves 90% reduction in Source Lines-of-Code (SLoC) and 50% energy savings from aggregated data delivery.
Resumo:
The characteristics of carbon fibre reinforced laminates had widened their use, from aerospace to domestic appliances. A common characteristic is the need of drilling for assembly purposes. It is known that a drilling process that reduces the drill thrust force can decrease the risk of delamination. In this work, delamination assessment methods based on radiographic data are compared and correlated with mechanical test results (bearing test).
Resumo:
ISCAP’s Information Systems Department is composed of about twenty teachers who have, for several years, been using an e-learning environment (Moodle) combined with traditional assessment. A new e-assessment strategy was implemented recently in order to evaluate a practical topic, the use of spreadsheets to solve management problems. This topic is common to several courses of different undergraduate degree programs. Being e-assessment an outstanding task regarding theoretical topics, it becomes even more challenging when the topics under evaluation are practical. In order to understand the implications of this new type of assessment from the viewpoint of the students, questionnaires and interviews were undertaken. In this paper the analysis of the questionnaires are presented and discussed.
Resumo:
The excessive use of pesticides and fertilisers in agriculture has generated a decrease in groundwater and surface water quality in many regions of the EU, constituting a hazard for human health and the environment. Besides, on-site sewage disposal is an important source of groundwater contamination in urban and peri-urban areas. The assessment of groundwater vulnerability to contamination is an important tool to fulfil the demands of EU Directives. The purpose of this study is to assess the groundwater vulnerability to contamination related mainly to agricultural activities in a peri-urban area (Vila do Conde, NW Portugal). The hydrogeological framework is characterised mainly by fissured granitic basement and sedimentary cover. Water samples were collected and analysed for temperature, pH, electrical conductivity, chloride, phosphate, nitrate and nitrite. An evaluation of groundwater vulnerability to contamination was applied (GOD-S, Pesticide DRASTIC-Fm, SINTACS and SI) and the potential nitrate contamination risk was assessed, both on a hydrogeological GIS-based mapping. A principal component analysis was performed to characterised patterns of relationship among groundwater contamination, vulnerability, and the hydrogeological setting assessed. Levels of nitrate above legislation limits were detected in 75 % of the samples analysed. Alluvia units showed the highest nitrate concentrations and also the highest vulnerability and risk. Nitrate contamination is a serious problem affecting groundwater, particularly shallow aquifers, especially due to agriculture activities, livestock and cesspools. GIS-based cartography provided an accurate way to improve knowledge on water circulation models and global functioning of local aquifer systems. Finally, this study highlights the adequacy of an integrated approach, combining hydrogeochemical data, vulnerability assessments and multivariate analysis, to understand groundwater processes in peri-urban areas.
Resumo:
The recent technological advancements and market trends are causing an interesting phenomenon towards the convergence of High-Performance Computing (HPC) and Embedded Computing (EC) domains. On one side, new kinds of HPC applications are being required by markets needing huge amounts of information to be processed within a bounded amount of time. On the other side, EC systems are increasingly concerned with providing higher performance in real-time, challenging the performance capabilities of current architectures. The advent of next-generation many-core embedded platforms has the chance of intercepting this converging need for predictable high-performance, allowing HPC and EC applications to be executed on efficient and powerful heterogeneous architectures integrating general-purpose processors with many-core computing fabrics. To this end, it is of paramount importance to develop new techniques for exploiting the massively parallel computation capabilities of such platforms in a predictable way. P-SOCRATES will tackle this important challenge by merging leading research groups from the HPC and EC communities. The time-criticality and parallelisation challenges common to both areas will be addressed by proposing an integrated framework for executing workload-intensive applications with real-time requirements on top of next-generation commercial-off-the-shelf (COTS) platforms based on many-core accelerated architectures. The project will investigate new HPC techniques that fulfil real-time requirements. The main sources of indeterminism will be identified, proposing efficient mapping and scheduling algorithms, along with the associated timing and schedulability analysis, to guarantee the real-time and performance requirements of the applications.
Resumo:
Os Sistemas de Apoio à Tomada de Decisão em Grupo (SADG) surgiram com o objetivo de apoiar um conjunto de decisores no processo de tomada de decisão. Uma das abordagens mais comuns na literatura para a implementação dos SADG é a utilização de Sistemas Multi-Agente (SMA). Os SMA permitem refletir com maior transparência o contexto real, tanto na representação que cada agente faz do decisor que representa como no formato de comunicação utilizado. Com o crescimento das organizações, atualmente vive-se uma viragem no conceito de tomada de decisão. Cada vez mais, devido a questões como: o estilo de vida, os mercados globais e o tipo de tecnologias disponíveis, faz sentido falar de decisão ubíqua. Isto significa que o decisor deverá poder utilizar o sistema a partir de qualquer local, a qualquer altura e através dos mais variados tipos de dispositivos eletrónicos tais como tablets, smartphones, etc. Neste trabalho é proposto um novo modelo de argumentação, adaptado ao contexto da tomada de decisão ubíqua para ser utilizado por um SMA na resolução de problemas multi-critério. É assumido que cada agente poderá utilizar um estilo de comportamento que afeta o modo como esse agente interage com outros agentes em situações de conflito. Sendo assim, pretende-se estudar o impacto da utilização de estilos de comportamento ao longo do processo da tomada de decisão e perceber se os agentes modelados com estilos de comportamento conseguem atingir o consenso mais facilmente quando comparados com agentes que não apresentam nenhum estilo de comportamento. Pretende-se ainda estudar se o número de argumentos trocados entre os agentes é proporcional ao nível de consenso final após o processo de tomada de decisão. De forma a poder estudar as hipóteses de investigação desenvolveu-se um protótipo de um SADG, utilizando um SMA. Desenvolveu-se ainda uma framework de argumentação que foi adaptada ao protótipo desenvolvido. Os resultados obtidos permitiram validar as hipóteses definidas neste trabalho tendo-se concluído que os agentes modelados com estilos de comportamento conseguem na maioria das vezes atingir um consenso mais facilmente comparado com agentes que não apresentam nenhum estilo de comportamento e que o número de argumentos trocados entre os agentes durante o processo de tomada de decisão não é proporcional ao nível de consenso final.
Resumo:
Com o aumento de plataformas móveis disponíveis no mercado e com o constante incremento na sua capacidade computacional, a possibilidade de executar aplicações e em especial jogos com elevados requisitos de desempenho aumentou consideravelmente. O mercado dos videojogos tem assim um cada vez maior número de potenciais clientes. Em especial, o mercado de jogos massive multiplayer online (MMO) tem-se tornado muito atractivo para as empresas de desenvolvimento de jogos. Estes jogos suportam uma elevada quantidade de jogadores em simultâneo que podem estar a executar o jogo em diferentes plataformas e distribuídos por um "mundo" de jogo extenso. Para incentivar a exploração desse "mundo", distribuem-se de forma inteligente pontos de interesse que podem ser explorados pelo jogador. Esta abordagem leva a um esforço substancial no planeamento e construção desses mundos, gastando tempo e recursos durante a fase de desenvolvimento. Isto representa um problema para as empresas de desenvolvimento de jogos, e em alguns casos, e impraticável suportar tais custos para equipas indie. Nesta tese e apresentada uma abordagem para a criação de mundos para jogos MMO. Estudam-se vários jogos MMO que são casos de sucesso de modo a identificar propriedades comuns nos seus mundos. O objectivo e criar uma framework flexível capaz de gerar mundos com estruturas que respeitam conjuntos de regras definidas por game designers. Para que seja possível usar a abordagem aqui apresentada em v arias aplicações diferentes, foram desenvolvidos dois módulos principais. O primeiro, chamado rule-based-map-generator, contem a lógica e operações necessárias para a criação de mundos. O segundo, chamado blocker, e um wrapper à volta do módulo rule-based-map-generator que gere as comunicações entre servidor e clientes. De uma forma resumida, o objectivo geral e disponibilizar uma framework para facilitar a geração de mundos para jogos MMO, o que normalmente e um processo bastante demorado e aumenta significativamente o custo de produção, através de uma abordagem semi-automática combinando os benefícios de procedural content generation (PCG) com conteúdo gráfico gerado manualmente.