50 resultados para self revelation mechanism
Resumo:
The IEEE 802.15.4 Medium Access Control (MAC) protocol is an enabling technology for time sensitive wireless sensor networks thanks to its Guaranteed-Time Slot (GTS) mechanism in the beacon-enabled mode. However, the protocol only supports explicit GTS allocation, i.e. a node allocates a number of time slots in each superframe for exclusive use. The limitation of this explicit GTS allocation is that GTS resources may quickly disappear, since a maximum of seven GTSs can be allocated in each superframe, preventing other nodes to benefit from guaranteed service. Moreover, the GTSs may be only partially used, resulting in wasted bandwidth. To overcome these limitations, this paper proposes i-GAME, an implicit GTS Allocation Mechanism in beacon-enabled IEEE 802.15.4 networks. The allocation is based on implicit GTS allocation requests, taking into account the traffic specifications and the delay requirements of the flows. The i-GAME approach enables the use of a GTS by multiple nodes, while all their (delay, bandwidth) requirements are still satisfied. For that purpose, we propose an admission control algorithm that enables to decide whether to accept a new GTS allocation request or not, based not only on the remaining time slots, but also on the traffic specifications of the flows, their delay requirements and the available bandwidth resources. We show that our proposal improves the bandwidth utilization compared to the explicit allocation used in the IEEE 802.15.4 protocol standard. We also present some practical considerations for the implementation of i-GAME, ensuring backward compatibility with the IEEE 801.5.4 standard with only minor add-ons.
Resumo:
The IEEE 802.15.4 protocol proposes a flexible communication solution for Low-Rate Wireless Personal Area Networks (LR-WPAN) including wireless sensor networks (WSNs). It presents the advantage to fit different requirements of potential applications by adequately setting its parameters. When in beaconenabled mode, the protocol can provide timeliness guarantees by using its Guaranteed Time Slot (GTS) mechanism. However, power-efficiency and timeliness guarantees are often two antagonistic requirements in wireless sensor networks. The purpose of this paper is to analyze and propose a methodology for setting the relevant parameters of IEEE 802.15.4-compliant WSNs that takes into account a proper trade-off between power-efficiency and delay bound guarantees. First, we propose two accurate models of service curves for a GTS allocation as a function of the IEEE 802.15.4 parameters, using Network Calculus formalism. We then evaluate the delay bound guaranteed by a GTS allocation and express it as a function of the duty cycle. Based on the relation between the delay requirement and the duty cycle, we propose a power-efficient superframe selection method that simultaneously reduces power consumption and enables meeting the delay requirements of real-time flows allocating GTSs. The results of this work may pave the way for a powerefficient management of the GTS mechanism in an IEEE 802.15.4 cluster.
Resumo:
This technical report describes the implementation details of the Implicit GTS Allocation Mechanism (i-GAME), for the IEEE 802.15.4 protocol. The i-GAME was implemented in nesC/TinyOS for the CrossBow MICAz mote, over our own implementation of the IEEE 802.15.4 protocol stack. This document provides the implementation details, including a description of the i-GAME software interfaces.
Resumo:
Solving systems of nonlinear equations is a very important task since the problems emerge mostly through the mathematical modelling of real problems that arise naturally in many branches of engineering and in the physical sciences. The problem can be naturally reformulated as a global optimization problem. In this paper, we show that a self-adaptive combination of a metaheuristic with a classical local search method is able to converge to some difficult problems that are not solved by Newton-type methods.
Resumo:
The interest in the development of climbing robots has grown rapidly in the last years. Climbing robots are useful devices that can be adopted in a variety of applications, such as maintenance and inspection in the process and construction industries. These systems are mainly adopted in places where direct access by a human operator is very expensive, because of the need for scaffolding, or very dangerous, due to the presence of an hostile environment. The main motivations are to increase the operation efficiency, by eliminating the costly assembly of scaffolding, or to protect human health and safety in hazardous tasks. Several climbing robots have already been developed, and other are under development, for applications ranging from cleaning to inspection of difficult to reach constructions. A wall climbing robot should not only be light, but also have large payload, so that it may reduce excessive adhesion forces and carry instrumentations during navigation. These machines should be capable of travelling over different types of surfaces, with different inclinations, such as floors, walls, or ceilings, and to walk between such surfaces (Elliot et al. (2006); Sattar et al. (2002)). Furthermore, they should be able of adapting and reconfiguring for various environment conditions and to be self-contained. Up to now, considerable research was devoted to these machines and various types of experimental models were already proposed (according to Chen et al. (2006), over 200 prototypes aimed at such applications had been developed in the world by the year 2006). However, we have to notice that the application of climbing robots is still limited. Apart from a couple successful industrialized products, most are only prototypes and few of them can be found in common use due to unsatisfactory performance in on-site tests (regarding aspects such as their speed, cost and reliability). Chen et al. (2006) present the main design problems affecting the system performance of climbing robots and also suggest solutions to these problems. The major two issues in the design of wall climbing robots are their locomotion and adhesion methods. With respect to the locomotion type, four types are often considered: the crawler, the wheeled, the legged and the propulsion robots. Although the crawler type is able to move relatively faster, it is not adequate to be applied in rough environments. On the other hand, the legged type easily copes with obstacles found in the environment, whereas generally its speed is lower and requires complex control systems. Regarding the adhesion to the surface, the robots should be able to produce a secure gripping force using a light-weight mechanism. The adhesion method is generally classified into four groups: suction force, magnetic, gripping to the surface and thrust force type. Nevertheless, recently new methods for assuring the adhesion, based in biological findings, were proposed. The vacuum type principle is light and easy to control though it presents the problem of supplying compressed air. An alternative, with costs in terms of weight, is the adoption of a vacuum pump. The magnetic type principle implies heavy actuators and is used only for ferromagnetic surfaces. The thrust force type robots make use of the forces developed by thrusters to adhere to the surfaces, but are used in very restricted and specific applications. Bearing these facts in mind, this chapter presents a survey of different applications and technologies adopted for the implementation of climbing robots locomotion and adhesion to surfaces, focusing on the new technologies that are recently being developed to fulfill these objectives. The chapter is organized as follows. Section two presents several applications of climbing robots. Sections three and four present the main locomotion principles, and the main "conventional" technologies for adhering to surfaces, respectively. Section five describes recent biological inspired technologies for robot adhesion to surfaces. Section six introduces several new architectures for climbing robots. Finally, section seven outlines the main conclusions.
Resumo:
With advancement in computer science and information technology, computing systems are becoming increasingly more complex with an increasing number of heterogeneous components. They are thus becoming more difficult to monitor, manage, and maintain. This process has been well known as labor intensive and error prone. In addition, traditional approaches for system management are difficult to keep up with the rapidly changing environments. There is a need for automatic and efficient approaches to monitor and manage complex computing systems. In this paper, we propose an innovative framework for scheduling system management by combining Autonomic Computing (AC) paradigm, Multi-Agent Systems (MAS) and Nature Inspired Optimization Techniques (NIT). Additionally, we consider the resolution of realistic problems. The scheduling of a Cutting and Treatment Stainless Steel Sheet Line will be evaluated. Results show that proposed approach has advantages when compared with other scheduling systems
Resumo:
This paper proposes a novel agent-based approach to Meta-Heuristics self-configuration. Meta-heuristics are algorithms with parameters which need to be set up as efficient as possible in order to unsure its performance. A learning module for self-parameterization of Meta-heuristics (MH) in a Multi-Agent System (MAS) for resolution of scheduling problems is proposed in this work. The learning module is based on Case-based Reasoning (CBR) and two different integration approaches are proposed. A computational study is made for comparing the two CBR integration perspectives. Finally, some conclusions are reached and future work outlined.
Resumo:
A vasta aplicação de antibióticos em sistemas de aquacultura tem conduzido ao aparecimento de espécies microbianas resistentes, devendo ser evitada ou, se possível, minimizada. A minimização implica um controlo muito rigoroso das doses aplicadas, algo que só pode ser realizado se todo o processo analítico for simples, barato e puder ser implementado no local. O presente trabalho descreve para este efeito um sistema sensor constituído por papel quimicamente modificado, cujo procedimento analítico foi semelhante à monitorização de pH através das conhecidas tiras indicadoras de pH universal. O papel foi modificado com base em reacções de monocamadas e assumiu uma coloração típica após contacto com o antibiótico. A intensidade da coloração estava também relacionada com a concentração desse antibiótico. Como prova deste conceito, o sensor proposto foi desenhado tendo em vista a monitorização de oxitetraciclina, um dos antibióticos utilizados neste contexto com elevada frequência. A modificação do papel foi baseada na alteração química das unidades de glucose na matriz sólida por reacção covalente com reagentes apropriados. Foram utilizadas duas estratégias diferentes para este efeito: uma foi baseada em reacções de tipo sol-gel e a outra em reacção de adição nucleófila. Posteriormente, o papel foi modificado com espécies metálicas capazes de mudar de cor na presença da oxitetraciclina. Estas modificações químicas foram avaliadas e optimizadas relativamente a vários parâmetros, no sentido de promover uma variação de cor intensa face à concentração de antibiótico. Procedeu-se ainda ao controlo desta modificação por técnicas de espectroscopia de infravermelho. A variação de cor foi avaliada por comparação visual, mas registada por imagem digital. Os materiais sensores preparados foram ainda caracterizados do ponto de vista do seu desempenho analítico. Incluíram-se aqui a identificação de uma relação linear entre coordenadas de cor e concentração, a atribuição de uma gama de concentrações capaz de fornecer uma resposta previsível e resposta cruzada face a outros antibióticos. Procedeu-se ainda à aplicação dos sensores numa amostra de água ambiental dopada com antibiótico. De uma forma geral, foi possível estabelecer um processo simples de modificação de papel capaz de aferir a presença e a quantidade de tetraciclinas, mais concretamente a oxitetraciclina. O processo aqui estabelecido é promissor, antevendo a concretização de uma metodologia simples, barata e local para a monitorização de anitbióticos em águas.
Resumo:
The interest in the development of climbing robots is growing rapidly. Motivations are typically to increase the operation efficiency by obviating the costly assembly of scaffolding or to protect human health and safety in hazardous tasks. Climbing robots are starting to be developed for applications ranging from cleaning to inspection of difficult to reach constructions. These robots should be capable of travelling on different types of surfaces, with varying inclinations, such as floors, walls, ceilings, and to walk between such surfaces. Furthermore, these machines should be capable of adapting and reconfiguring for various environment conditions and to be self-contained. Regarding the adhesion to the surface, they should be able to produce a secure gripping force using a light-weight mechanism. This paper presents a survey of different applications and technologies proposed for the implementation of climbing robots.
Resumo:
To boost logic density and reduce per unit power consumption SRAM-based FPGAs manufacturers adopted nanometric technologies. However, this technology is highly vulnerable to radiation-induced faults, which affect values stored in memory cells, and to manufacturing imperfections. Fault tolerant implementations, based on Triple Modular Redundancy (TMR) infrastructures, help to keep the correct operation of the circuit. However, TMR is not sufficient to guarantee the safe operation of a circuit. Other issues like module placement, the effects of multi- bit upsets (MBU) or fault accumulation, have also to be addressed. In case of a fault occurrence the correct operation of the affected module must be restored and/or the current state of the circuit coherently re-established. A solution that enables the autonomous restoration of the functional definition of the affected module, avoiding fault accumulation, re-establishing the correct circuit state in real-time, while keeping the normal operation of the circuit, is presented in this paper.
Resumo:
To increase the amount of logic available in SRAM-based FPGAs manufacturers are using nanometric technologies to boost logic density and reduce prices. However, nanometric scales are highly vulnerable to radiation-induced faults that affect values stored in memory cells. Since the functional definition of FPGAs relies on memory cells, they become highly prone to this type of faults. Fault tolerant implementations, based on triple modular redundancy (TMR) infrastructures, help to keep the correct operation of the circuit. However, TMR is not sufficient to guarantee the safe operation of a circuit. Other issues like the effects of multi-bit upsets (MBU) or fault accumulation, have also to be addressed. Furthermore, in case of a fault occurrence the correct operation of the affected module must be restored and the current state of the circuit coherently re-established. A solution that enables the autonomous correct restoration of the functional definition of the affected module, avoiding fault accumulation, re-establishing the correct circuit state in realtime, while keeping the normal operation of the circuit, is presented in this paper.
Resumo:
The new generations of SRAM-based FPGA (field programmable gate array) devices are the preferred choice for the implementation of reconfigurable computing platforms intended to accelerate processing in real-time systems. However, FPGA's vulnerability to hard and soft errors is a major weakness to robust configurable system design. In this paper, a novel built-in self-healing (BISH) methodology, based on run-time self-reconfiguration, is proposed. A soft microprocessor core implemented in the FPGA is responsible for the management and execution of all the BISH procedures. Fault detection and diagnosis is followed by repairing actions, taking advantage of the dynamic reconfiguration features offered by new FPGA families. Meanwhile, modular redundancy assures that the system still works correctly
Resumo:
No âmbito da Unidade Curricular Dissertação, inserida no segundo ano do Mestrado em Engenharia Mecânica – Gestão Industrial do Instituto Superior de Engenharia do Porto, foi proposto um projecto que está a ser desenvolvido pela equipa de Engenharia e Gestão Industrial, da unidade de negócios AMT, intitulado por: “Projecto Índia – Desenvolvimento da nova unidade fabril” Este projecto tem como principal objectivo o desenvolvimento de uma fábrica de excelência na Índia de fabricação de componentes de média tensão, isto é, com processos logísticos bem definidos e com linhas de produção o mais automatizadas possivel. Esta nova fábrica de raiz, vai ser gerida e decalcada do modelo atualmente existe na fábrica da EFACEC de componentes de média tensão existente em Portugal. Numa primeira fase do desenvolvimento do projecto, foi seleccionado um edifício com cerca de 1600m2 em Nashik, uma localidade a cerca de 171 Km de Mumbai, onde se encontram 80% dos fornecedores da EFACEC. Foram identificados os produtos a serem fabricados e quantificada a respectiva procura anual. Foi efectuado o balanceamento de cada uma das linhas e desenhado o layout. Neste layout contemplou-se as áreas de produção, laboratório, gabinetes de chefes de equipa, expedição, recepção e armazém. Após a definição das áreas de montagem de cada produto, iniciou-se a concepção das linhas de produção, sobretudo automáticas, com a definição da cadência de produção. A linha de fabricação que é especialmente detalhada neste documento é a linha de montagem dos comandos CI. Este é o produto com mais procura. Foi também definido o processo logístico do fluxo interno da fábrica. Nas linhas de produção foi implementado o sistema de controlo de fluxo baseado em cartões Kanban e no armazém criou-se um novo conceito de controlo e localização de produtos, o “Aquiles”. O Aquiles permite automaticamente e através da leitura de código de barras, indexar os artigos nas estantes. Cada artigo e cada estante e/ou localização estão codificados e no momento de recepção de material o código do artigo é associado ao código da estante. No âmbito de explorar todas as soluções possíveis para a um melhor desenvolvimento desta nova fábrica foram abordados temas como “JIT”, “Pull Flow”, “Kanban”, “Takttime”.
Resumo:
Power law (PL) distributions have been largely reported in the modeling of distinct real phenomena and have been associated with fractal structures and self-similar systems. In this paper, we analyze real data that follows a PL and a double PL behavior and verify the relation between the PL coefficient and the capacity dimension of known fractals. It is to be proved a method that translates PLs coefficients into capacity dimension of fractals of any real data.
Resumo:
This paper presents a modified Particle Swarm Optimization (PSO) methodology to solve the problem of energy resources management with high penetration of distributed generation and Electric Vehicles (EVs) with gridable capability (V2G). The objective of the day-ahead scheduling problem in this work is to minimize operation costs, namely energy costs, regarding the management of these resources in the smart grid context. The modifications applied to the PSO aimed to improve its adequacy to solve the mentioned problem. The proposed Application Specific Modified Particle Swarm Optimization (ASMPSO) includes an intelligent mechanism to adjust velocity limits during the search process, as well as self-parameterization of PSO parameters making it more user-independent. It presents better robustness and convergence characteristics compared with the tested PSO variants as well as better constraint handling. This enables its use for addressing real world large-scale problems in much shorter times than the deterministic methods, providing system operators with adequate decision support and achieving efficient resource scheduling, even when a significant number of alternative scenarios should be considered. The paper includes two realistic case studies with different penetration of gridable vehicles (1000 and 2000). The proposed methodology is about 2600 times faster than Mixed-Integer Non-Linear Programming (MINLP) reference technique, reducing the time required from 25 h to 36 s for the scenario with 2000 vehicles, with about one percent of difference in the objective function cost value.