46 resultados para Minimalist Running


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Modeling the fundamental performance limits of Wireless Sensor Networks (WSNs) is of paramount importance to understand their behavior under the worst-case conditions and to make the appropriate design choices. This is particular relevant for time-sensitive WSN applications, where the timing behavior of the network protocols (message transmission must respect deadlines) impacts on the correct operation of these applications. In that direction this paper contributes with a methodology based on Network Calculus, which enables quick and efficient worst-case dimensioning of static or even dynamically changing cluster-tree WSNs where the data sink can either be static or mobile. We propose closed-form recurrent expressions for computing the worst-case end-to-end delays, buffering and bandwidth requirements across any source-destination path in a cluster-tree WSN. We show how to apply our methodology to the case of IEEE 802.15.4/ZigBee cluster-tree WSNs. Finally, we demonstrate the validity and analyze the accuracy of our methodology through a comprehensive experimental study using commercially available technology, namely TelosB motes running TinyOS.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Modeling the fundamental performance limits of Wireless Sensor Networks (WSNs) is of paramount importance to understand their behavior under worst-case conditions and to make the appropriate design choices. In that direction this paper contributes with an analytical methodology for modeling cluster-tree WSNs where the data sink can either be static or mobile. We assess the validity and pessimism of analytical model by comparing the worst-case results with the values measured through an experimental test-bed based on Commercial-Off- The-Shelf (COTS) technologies, namely TelosB motes running TinyOS.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Constrained and unconstrained Nonlinear Optimization Problems often appear in many engineering areas. In some of these cases it is not possible to use derivative based optimization methods because the objective function is not known or it is too complex or the objective function is non-smooth. In these cases derivative based methods cannot be used and Direct Search Methods might be the most suitable optimization methods. An Application Programming Interface (API) including some of these methods was implemented using Java Technology. This API can be accessed either by applications running in the same computer where it is installed or, it can be remotely accessed through a LAN or the Internet, using webservices. From the engineering point of view, the information needed from the API is the solution for the provided problem. On the other hand, from the optimization methods researchers’ point of view, not only the solution for the problem is needed. Also additional information about the iterative process is useful, such as: the number of iterations; the value of the solution at each iteration; the stopping criteria, etc. In this paper are presented the features added to the API to allow users to access to the iterative process data.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In practice the robotic manipulators present some degree of unwanted vibrations. The advent of lightweight arm manipulators, mainly in the aerospace industry, where weight is an important issue, leads to the problem of intense vibrations. On the other hand, robots interacting with the environment often generate impacts that propagate through the mechanical structure and produce also vibrations. In order to analyze these phenomena a robot signal acquisition system was developed. The manipulator motion produces vibrations, either from the structural modes or from endeffector impacts. The instrumentation system acquires signals from several sensors that capture the joint positions, mass accelerations, forces and moments, and electrical currents in the motors. Afterwards, an analysis package, running off-line, reads the data recorded by the acquisition system and extracts the signal characteristics. Due to the multiplicity of sensors, the data obtained can be redundant because the same type of information may be seen by two or more sensors. Because of the price of the sensors, this aspect can be considered in order to reduce the cost of the system. On the other hand, the placement of the sensors is an important issue in order to obtain the suitable signals of the vibration phenomenon. Moreover, the study of these issues can help in the design optimization of the acquisition system. In this line of thought a sensor classification scheme is presented. Several authors have addressed the subject of the sensor classification scheme. White (White, 1987) presents a flexible and comprehensive categorizing scheme that is useful for describing and comparing sensors. The author organizes the sensors according to several aspects: measurands, technological aspects, detection means, conversion phenomena, sensor materials and fields of application. Michahelles and Schiele (Michahelles & Schiele, 2003) systematize the use of sensor technology. They identified several dimensions of sensing that represent the sensing goals for physical interaction. A conceptual framework is introduced that allows categorizing existing sensors and evaluates their utility in various applications. This framework not only guides application designers for choosing meaningful sensor subsets, but also can inspire new systems and leads to the evaluation of existing applications. Today’s technology offers a wide variety of sensors. In order to use all the data from the diversity of sensors a framework of integration is needed. Sensor fusion, fuzzy logic, and neural networks are often mentioned when dealing with problem of combing information from several sensors to get a more general picture of a given situation. The study of data fusion has been receiving considerable attention (Esteban et al., 2005; Luo & Kay, 1990). A survey of the state of the art in sensor fusion for robotics can be found in (Hackett & Shah, 1990). Henderson and Shilcrat (Henderson & Shilcrat, 1984) introduced the concept of logic sensor that defines an abstract specification of the sensors to integrate in a multisensor system. The recent developments of micro electro mechanical sensors (MEMS) with unwired communication capabilities allow a sensor network with interesting capacity. This technology was applied in several applications (Arampatzis & Manesis, 2005), including robotics. Cheekiralla and Engels (Cheekiralla & Engels, 2005) propose a classification of the unwired sensor networks according to its functionalities and properties. This paper presents a development of a sensor classification scheme based on the frequency spectrum of the signals and on a statistical metrics. Bearing these ideas in mind, this paper is organized as follows. Section 2 describes briefly the robotic system enhanced with the instrumentation setup. Section 3 presents the experimental results. Finally, section 4 draws the main conclusions and points out future work.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper proposes an online mechanism that can evaluate the sensitivity of single event upsets (SEUs) of field programmable gate arrays (FPGAs). The online detection mechanism cyclically reads and compares the values form the external and internal configuration memories, taking into account the mask information. This remote detection method also signals any mismatch as a result of a SEU that affects both used and not-used FPGA parts, which maximizes the monitored area. By utilizing an external, Web-accessible controller that is connected to the test infrastructure, the possibility of running the same operation in a remote manner is enabled. Moreover, the need for a local memory to store the mask values is also eliminated.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work extends a recent comparative study covering four different courses lectured at the Polytechnic of Porto - School of Engineering, in respect to the usage of a particular Learning Management System, i.e. Moodle, and its impact on students' results. A fifth course, which includes a number of resources especially supporting laboratory classes, is now added to the analysis. This particular course includes a number of remote experiments, made available through VISIR (Virtual Instrument Systems in Reality) and directly accessible through links included in the Moodle course page. We have analyzed the students' behavior in following these links and in effectively running experiments in VISIR (and also using other lab related resources, in Moodle). This data have been correlated with students' classifications in the lab component and in the exam, each one weighting 50% of their final marks. We aimed to compare students' performance in a richly Moodle-supported environment (with lab component) and in a poorly Moodle-supported environment (with only theoretical component). This question followed from conclusions drawn in the above referred comparative study, where it was shown that even though a positive correlation factor existed between the number of Moodle accesses and the final exam grade obtained by each student, its explanation behind was not straightforward, as the quality of the resources was preponderant over its quantity.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Até 2020, a Europa terá de reduzir 20% das suas emissões de gases com efeito de estufa, 20% da produção de energia terá de ser proveniente de fontes renováveis e a eficiência energética deverá aumentar 20%. Estas são as metas apresentadas pela União Europeia, que ficaram conhecidas por 20/20/20 [1]. A Refinaria de Matosinhosé um complexo industrial que opera no sector da refinação e que apresenta preocupações ao nível da eficiência energética e dos aspectos ambientais subjacentes. No âmbito da racionalização energética das refinarias, a Galp Energia tem vindo a implementar um conjunto de medidas, adoptando as melhores tecnologias disponíveis com o objectivo de diminuir os consumos de energia, promover a eficiência energética e reduzir as emissões de dióxido de carbono. Para ir de encontro a estas medidas foi elaborado um estudo comparativo que permitiu à empresa definir as medidas consideradas prioritárias. Uma solução encontrada visa a execução de projectos que não requerem investimento e que têm acções imediatas, tais como o aumento da eficiência energética das fornalhas [1]. Este trabalho realizado na Galp Energia S.A. teve como objectivo principal a optimização energética da Unidade de Desalfatação do Propano da Fábrica de Óleos Base. Esta optimização baseou-se no aproveitamento energético da corrente de fundo da coluna de rectificação T2003C com uma potência calorífica de 2,79 Gcal/h. Após levantamento de todas as variáveis do processo relativas a esta unidade, especialmente a potência calorífica das correntes envolvidas chegou-se á conclusão que a fornalha H2101 poderá ser substituída por dois permutadores, reduzindo desta forma os consumos energéticos. Pois a corrente de fundo da coluna T2003 com uma potência calorífica 2,79 Gcal/h poderá permutar calor com a corrente da mistura asfalto com propano, fazendo com que esta atinja temperatura superior à obtida com a fornalha em funcionamento. A análise económica ao consumo e respectivo custo do fuelóleo na fornalha para o período de um ano foi realizada, sendo o seu custo de combustível de 611.396,00 €. O valor da aquisição dos permutadores é 86.355,97€, sendo rentável a alteração proposta neste projecto.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Concepts like E-learning and M-learning are changing the traditional learning place. No longer restricted to well-defined physical places, education on Automation and other Engineering areas is entering the so-called ubiquitous learning place, where even the more practical knowledge (acquired at lab classes) is now moving into, due to emergent concepts such as Remote Experimentation or Mobile Experimentation. While Remote Experimentation is traditionally regarded as the remote access to real-world experiments through a simple web browser running on a PC connected to the Internet, Mobile Experimentation may be seen as the access to those same (or others) experiments, through mobile devices, used in M-learning contexts. These two distinct client types (PCs versus mobile devices) pose specific requirements for the remote lab infrastructure, namely the ability to tune the experiment interface according to the characteristics (e.g. display size) of the accessing device. This paper addresses those requirements, namely by proposing a new architecture for the remote lab infrastructure able to accommodate both Remote and Mobile Experimentation scenarios.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper presents a low-cost scaled model of a silo for drying and airing cereal grains. It allows the control and monitor of several parameters associated to the silo's operation, through a remote accessible infrastructure. The scaled model consists of a 2.50 m wide × 2.10 m long plant with all control and monitor capacities provided by micro-Web servers. An application running on the micro-Web servers enables storing all parameters in a data basis for later analysis. The implemented model aims to support a remote experimentation facility for technological education, research-oriented tutorials, and industrial applications. Given the low-cost requirement, this remote facility can be easily replicated in other institutions to support a network of remote labs, which encompasses the concurrent access of several users (e.g. students).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Dynamically reconfigurable systems have benefited from a new class of FPGAs recently introduced into the market, which allow partial and dynamic reconfiguration at run-time, enabling multiple independent functions from different applications to share the same device, swapping resources as needed. When the sequence of tasks to be performed is not predictable, resource allocation decisions have to be made on-line, fragmenting the FPGA logic space. A rearrangement may be necessary to get enough contiguous space to efficiently implement incoming functions, to avoid spreading their components and, as a result, degrading their performance. This paper presents a novel active replication mechanism for configurable logic blocks (CLBs), able to implement on-line rearrangements, defragmenting the available FPGA resources without disturbing those functions that are currently running.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Reconfigurable computing experienced a considerable expansion in the last few years, due in part to the fast run-time partial reconfiguration features offered by recent SRAM-based Field Programmable Gate Arrays (FPGAs), which allowed the implementation in real-time of dynamic resource allocation strategies, with multiple independent functions from different applications sharing the same logic resources in the space and temporal domains. However, when the sequence of reconfigurations to be performed is not predictable, the efficient management of the logic space available becomes the greatest challenge posed to these systems. Resource allocation decisions have to be made concurrently with system operation, taking into account function priorities and optimizing the space currently available. As a consequence of the unpredictability of this allocation procedure, the logic space becomes fragmented, with many small areas of free resources failing to satisfy most requests and so remaining unused. A rearrangement of the currently running functions is therefore necessary, so as to obtain enough contiguous space to implement incoming functions, avoiding the spreading of their components and the resulting degradation of system performance. A novel active relocation procedure for Configurable Logic Blocks (CLBs) is herein presented, able to carry out online rearrangements, defragmenting the available FPGA resources without disturbing functions currently running.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Consider the problem of assigning implicit-deadline sporadic tasks on a heterogeneous multiprocessor platform comprising two different types of processors—such a platform is referred to as two-type platform. We present two low degree polynomial time-complexity algorithms, SA and SA-P, each providing the following guarantee. For a given two-type platform and a task set, if there exists a task assignment such that tasks can be scheduled to meet deadlines by allowing them to migrate only between processors of the same type (intra-migrative), then (i) using SA, it is guaranteed to find such an assignment where the same restriction on task migration applies but given a platform in which processors are 1+α/2 times faster and (ii) SA-P succeeds in finding a task assignment where tasks are not allowed to migrate between processors (non-migrative) but given a platform in which processors are 1+α times faster. The parameter 0<α≤1 is a property of the task set; it is the maximum of all the task utilizations that are no greater than 1. We evaluate average-case performance of both the algorithms by generating task sets randomly and measuring how much faster processors the algorithms need (which is upper bounded by 1+α/2 for SA and 1+α for SA-P) in order to output a feasible task assignment (intra-migrative for SA and non-migrative for SA-P). In our evaluations, for the vast majority of task sets, these algorithms require significantly smaller processor speedup than indicated by their theoretical bounds. Finally, we consider a special case where no task utilization in the given task set can exceed one and for this case, we (re-)prove the performance guarantees of SA and SA-P. We show, for both of the algorithms, that changing the adversary from intra-migrative to a more powerful one, namely fully-migrative, in which tasks can migrate between processors of any type, does not deteriorate the performance guarantees. For this special case, we compare the average-case performance of SA-P and a state-of-the-art algorithm by generating task sets randomly. In our evaluations, SA-P outperforms the state-of-the-art by requiring much smaller processor speedup and by running orders of magnitude faster.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Task scheduling is one of the key mechanisms to ensure timeliness in embedded real-time systems. Such systems have often the need to execute not only application tasks but also some urgent routines (e.g. error-detection actions, consistency checkers, interrupt handlers) with minimum latency. Although fixed-priority schedulers such as Rate-Monotonic (RM) are in line with this need, they usually make a low processor utilization available to the system. Moreover, this availability usually decreases with the number of considered tasks. If dynamic-priority schedulers such as Earliest Deadline First (EDF) are applied instead, high system utilization can be guaranteed but the minimum latency for executing urgent routines may not be ensured. In this paper we describe a scheduling model according to which urgent routines are executed at the highest priority level and all other system tasks are scheduled by EDF. We show that the guaranteed processor utilization for the assumed scheduling model is at least as high as the one provided by RM for two tasks, namely 2(2√−1). Seven polynomial time tests for checking the system timeliness are derived and proved correct. The proposed tests are compared against each other and to an exact but exponential running time test.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Os aproveitamentos geotérmicos têm vindo a aumentar significativamente em todo o mundo, sendo os Estados Unidos da América, o maior produtor desta energia proveniente do interior da Terra, com cerca de 3.187 MW de capacidade instalada. Portugal tem capacidade instalada total de 29 MW, no entanto no que se refere ao aproveitamento de “alta entalpia”, isto é, o aproveitamento geotérmico para produção elétrica, apenas se encontra no arquipélago dos Açores, na ilha de S. Miguel, onde estão instaladas e em funcionamento duas centrais geotérmicas com a potência total de 23 MW, com produção de energia de 185 GWh. Em Portugal Continental, não se consegue produzir energia elétrica devido às temperaturas existentes, restringindo esta utilização apenas ao aproveitamento de baixa entalpia (máximo de 76 ºC). Este aproveitamento normalmente é feito em cascata, segundo, predominando o aquecimento de águas sanitárias, climatização, e para termas, usando águas termominerais. Para a exploração deste recurso renovável, é necessário conhecer a hidrogeologia do país, e relacioná-la com a fracturação, e acidentes tectónicos. Portugal Continental, está divido em quatros partes distintas a nível hidrogeológico, o Maciço Antigo, a Orla Ocidental, a Bacia Tejo-Sado e a Orla Meridional. Qualquer aproveitamento geotérmico em Portugal terá de atender a estas características, potenciando também, novas explorações geotérmicas orientadas para as pessoas, respeitando os valores sociais, culturais e ambientais. Neste contexto, existem alguns complexos geotérmicos em funcionamento, outros abandonados, e muitos outros em estudo para uma breve aplicação. Um exemplo de sucesso no aproveitamento do calor geotérmico, é o complexo de Chaves, que foi evoluindo desde 1985, até aos dias de hoje, continuando em exploração e em expansão para um melhor servir da população local. A existência de dois furos, e brevemente dum terceiro, servem para o abastecimento duma piscina, dum hotel, das termas, e da balneoterapia. Devido à riqueza a nível das temperaturas, dos caudais, e ao nível das necessidades energéticas existentes, este complexo apresenta um tempo de retorno de investimento de cerca de 7 anos, o que é geralmente considerado para investimentos para fins públicos, como é o caso. No âmbito das investigações agora realizadas, foi constatado que estes projetos suportam a cobertura de alguma incerteza hidrogeológica, dada a importante procura existente.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Com um mercado automóvel cada vez mais competitivo e com os construtores automóveis à procura de atingir os zero defeitos nos seus produtos, a Bosch Car Multimédia Portugal S.A, fabricante de sistemas multimédia para o mercado automóvel, tem como objetivo a qualidade perfeita dos seus produtos. Tal perfeição exige processos de fabrico cada vez mais evoluídos e com melhores sistemas de auxílio à montagem. Nesse sentido, a incorporação de sistemas de visão artificial para verificação da montagem correta dos componentes em sistemas multimédia tem vindo a crescer largamente. Os sistemas de inspeção visual da Cognex tornaram-se o standard da Bosch para a verifi-cação da montagem de componentes por serem sistemas bastante completos, fáceis de con-figurar e com um suporte técnico bastante completo. Estes sistemas têm vindo a ser inte-grados em diversas máquinas (postos) de montagem e nunca foi desenvolvida uma ferra-menta normalizada para integração destes sistemas com as máquinas. A ideia principal deste projeto passou por desenvolver um sistema (uma aplicação informá-tica) que permita controlar os indicadores de qualidade destes sistemas de visão, garantir o seguimento dos produtos montados e, ao mesmo tempo, efetuar cópias de segurança de todo o sistema para utilização em caso de avaria ou de troca de equipamento. Tal sistema foi desenvolvido recorrendo à programação de uma Dynamic Link Library (DLL), através da linguagem VisualBasic.NET, que permite às aplicações dos equipamen-tos (máquinas) da Bosch Car Multimédia comunicarem de uma forma universal e transpa-rente com os sistemas de inspeção visual da marca Cognex. Os objetivos a que o autor se propôs no desenvolvimento deste sistema foram na sua maioria alcançados e o projeto encontra-se atualmente implementado e em execução nas linhas de produção da Bosch Car Multimédia.