881 resultados para distributed computing
Resumo:
Free independent travelers require flexible, reactive service delivery due to their regularly changing location and activities and the lack of a wired Internet connection. A ubiquitous travel service delivery system that is able to dynamically deliver services in response to relevant events, such as changing location, availability of new last-minute specials, work opportunities, and safety issues can provide added value while retaining the flexibility that is so important to independent travelers. This article describes such a system. An engineering design research approach has been adopted to design the system. Issues addressed include traveler and service states and events, contexts, situations, and situation-action rules. An architecture is proposed that is based on distributed, cooperating software agents and mobile data technologies. The role of these agents is to continuously monitor situations that are occurring in the physical and virtual service spaces and to take the required action for any situations that are relevant to the traveler.
Resumo:
Graphical user interfaces (GUIs) are critical components of today's open source software. Given their increased relevance, the correctness and usability of GUIs are becoming essential. This paper describes the latest results in the development of our tool to reverse engineer the GUI layer of interactive computing open source systems. We use static analysis techniques to generate models of the user interface behavior from source code. Models help in graphical user interface inspection by allowing designers to concentrate on its more important aspects. One particular type of model that the tool is able to generate is state machines. The paper shows how graph theory can be useful when applied to these models. A number of metrics and algorithms are used in the analysis of aspects of the user interface's quality. The ultimate goal of the tool is to enable analysis of interactive system through GUIs source code inspection.
Resumo:
A sociedade atual é marcada por uma crescente competição entre as organizações, as quais para garantirem a sua sobrevivência têm que evoluir e utilizar a tecnologia de informação mais adequada. Num período em que as PMES portuguesas necessitam de reduzir custos e concentrarem-se mais no seu processo de negócio surge um novo paradigma tecnológico, a Computação na Nuvem, esta ainda é um paradigma bastante recente, onde os recursos de TI que apresentam características de elasticidade e dinâmicas são oferecidos ao cliente como um serviço, no qual este paga apenas o que consome (pagamento por utilização). A Cloud Computing com o seu novo paradigma permite às organizações, principalmente às PMES reduzirem custos, concentrarem-se no seu processo de negócio sem se preocuparem com a parte informática (cópias de segurança, assistência e perda de dados) e obter um retorno mais rápido do seu investimento. Este trabalho consiste no Estudo da Adoção Individual da Cloud Computing no contexto das PMEs Portuguesas, o qual apresenta como principais objetivos obter uma melhor compreensão do conceito Cloud Computing, entender a sua importância para os indivíduos das PMES, identificar os seus fatores de sucesso e insucesso e identificar as limitações sentidas de uma forma generalizada e de uma forma específica para as PMES Portuguesas. Para cumprir os objetivos mencionados anteriormente, efetuou-se uma extensa revisão de literatura, propôs-se um modelo conceptual de investigação com adaptações à Teoria Unificada de Aceitação e Uso de Tecnologia (UTAUT) e utilizou-se como instrumento de recolha de dados um questionário que foi enviado às PMES Portuguesas.
Resumo:
One of the most efficient approaches to generate the side information (SI) in distributed video codecs is through motion compensated frame interpolation where the current frame is estimated based on past and future reference frames. However, this approach leads to significant spatial and temporal variations in the correlation noise between the source at the encoder and the SI at the decoder. In such scenario, it would be useful to design an architecture where the SI can be more robustly generated at the block level, avoiding the creation of SI frame regions with lower correlation, largely responsible for some coding efficiency losses. In this paper, a flexible framework to generate SI at the block level in two modes is presented: while the first mode corresponds to a motion compensated interpolation (MCI) technique, the second mode corresponds to a motion compensated quality enhancement (MCQE) technique where a low quality Intra block sent by the encoder is used to generate the SI by doing motion estimation with the help of the reference frames. The novel MCQE mode can be overall advantageous from the rate-distortion point of view, even if some rate has to be invested in the low quality Intra coding blocks, for blocks where the MCI produces SI with lower correlation. The overall solution is evaluated in terms of RD performance with improvements up to 2 dB, especially for high motion video sequences and long Group of Pictures (GOP) sizes.
Resumo:
Motion compensated frame interpolation (MCFI) is one of the most efficient solutions to generate side information (SI) in the context of distributed video coding. However, it creates SI with rather significant motion compensated errors for some frame regions while rather small for some other regions depending on the video content. In this paper, a low complexity Infra mode selection algorithm is proposed to select the most 'critical' blocks in the WZ frame and help the decoder with some reliable data for those blocks. For each block, the novel coding mode selection algorithm estimates the encoding rate for the Intra based and WZ coding modes and determines the best coding mode while maintaining a low encoder complexity. The proposed solution is evaluated in terms of rate-distortion performance with improvements up to 1.2 dB regarding a WZ coding mode only solution.
Resumo:
A necessidade de poder computacional é crescente nas diversas áreas de actuação humana, tanto na indústria, como em ambientes académicos. Grid Computing permite a ligação de recursos computacionais dispersos de maneira a permitir a sua utilização mais eficaz, fornecendo aos utilizadores um acesso simplificado ao poder computacional de diversos sistemas. Os primeiros projectos de Grid Computing implicavam a ligação de máquinas paralelas ou aglomerados de alto desempenho e alto custo, disponíveis apenas em algumas instituições. Contrastando com o elevado custo dos super-computadores, os computadores pessoais e a Internet sofreram uma evolução significativa nos últimos anos. O uso de computadores dispersos em uma WAN pode representar um ambiente muito interessante para processamento de alto desempenho. Os sistemas em Grid fornecem a possibilidade de se utilizar um conjunto de computadores pessoais de modo a fornecer uma computação que utiliza recursos que de outra maneira estariam omissos. Este trabalho consiste no estudo de Grid Computing a nível de conceito e de arquitectura e numa análise ao seu estado actual hoje em dia. Como complemento foi desenvolvido um componente que permite o desenvolvimento de serviços para Grids (Grid Services) mais eficaz do que o modelo de suporte a serviços actualmente utilizado. Este componente é disponibilizado sob a forma um plug-in para a plataforma Eclipse IDE.
Resumo:
Integrated manufacturing constitutes a complex system made of heterogeneous information and control subsystems. Those subsystems are not designed to the cooperation. Typically each subsystem automates specific processes, and establishes closed application domains, therefore it is very difficult to integrate it with other subsystems in order to respond to the needed process dynamics. Furthermore, to cope with ever growing marketcompetition and demands, it is necessary for manufacturing/enterprise systems to increase their responsiveness based on up-to-date knowledge and in-time data gathered from the diverse information and control systems. These have created new challenges for manufacturing sector, and even bigger challenges for collaborative manufacturing. The growing complexity of the information and communication technologies when coping with innovative business services based on collaborative contributions from multiple stakeholders, requires novel and multidisciplinary approaches. Service orientation is a strategic approach to deal with such complexity, and various stakeholders' information systems. Services or more precisely the autonomous computational agents implementing the services, provide an architectural pattern able to cope with the needs of integrated and distributed collaborative solutions. This paper proposes a service-oriented framework, aiming to support a virtual organizations breeding environment that is the basis for establishing short or long term goal-oriented virtual organizations. The notion of integrated business services, where customers receive some value developed through the contribution from a network of companies is a key element.
Resumo:
The advances made in channel-capacity codes, such as turbo codes and low-density parity-check (LDPC) codes, have played a major role in the emerging distributed source coding paradigm. LDPC codes can be easily adapted to new source coding strategies due to their natural representation as bipartite graphs and the use of quasi-optimal decoding algorithms, such as belief propagation. This paper tackles a relevant scenario in distributedvideo coding: lossy source coding when multiple side information (SI) hypotheses are available at the decoder, each one correlated with the source according to different correlation noise channels. Thus, it is proposed to exploit multiple SI hypotheses through an efficient joint decoding technique withmultiple LDPC syndrome decoders that exchange information to obtain coding efficiency improvements. At the decoder side, the multiple SI hypotheses are created with motion compensated frame interpolation and fused together in a novel iterative LDPC based Slepian-Wolf decoding algorithm. With the creation of multiple SI hypotheses and the proposed decoding algorithm, bitrate savings up to 8.0% are obtained for similar decoded quality.
Resumo:
A novel approach to scheduling resolution by combining Autonomic Computing (AC), Multi-Agent Systems (MAS), Case-based Reasoning (CBR), and Bio-Inspired Optimization Techniques (BIT) will be described. AC has emerged as a paradigm aiming at incorporating applications with a management structure similar to the central nervous system. The main intentions are to improve resource utilization and service quality. In this paper we envisage the use of MAS paradigm for supporting dynamic and distributed scheduling in Manufacturing Systems with AC properties, in order to reduce the complexity of managing manufacturing systems and human interference. The proposed CBR based Intelligent Scheduling System was evaluated under different dynamic manufacturing scenarios.
Resumo:
Scheduling resolution requires the intervention of highly skilled human problemsolvers. This is a very hard and challenging domain because current systems are becoming more and more complex, distributed, interconnected and subject to rapidly changing. A natural Autonomic Computing evolution in relation to Current Computing is to provide systems with Self-Managing ability with a minimum human interference. This paper addresses the resolution of complex scheduling problems using cooperative negotiation. A Multi-Agent Autonomic and Meta-heuristics based framework with self-configuring capabilities is proposed.
Resumo:
This chapter addresses the resolution of dynamic scheduling by means of meta-heuristic and multi-agent systems. Scheduling is an important aspect of automation in manufacturing systems. Several contributions have been proposed, but the problem is far from being solved satisfactorily, especially if scheduling concerns real world applications. The proposed multi-agent scheduling system assumes the existence of several resource agents (which are decision-making entities based on meta-heuristics) distributed inside the manufacturing system that interact with other agents in order to obtain optimal or near-optimal global performances.
Resumo:
In competitive electricity markets with deep concerns at the efficiency level, demand response programs gain considerable significance. In the same way, distributed generation has gained increasing importance in the operation and planning of power systems. Grid operators and utilities are taking new initiatives, recognizing the value of demand response and of distributed generation for grid reliability and for the enhancement of organized spot market´s efficiency. Grid operators and utilities become able to act in both energy and reserve components of electricity markets. This paper proposes a methodology for a joint dispatch of demand response and distributed generation to provide energy and reserve by a virtual power player that operates a distribution network. The proposed method has been computationally implemented and its application is illustrated in this paper using a 32 bus distribution network with 32 medium voltage consumers.
Resumo:
All over the world Distributed Generation is seen as a valuable help to get cleaner and more efficient electricity. To get negotiation power and advantages of scale economy, distributed producers can be aggregated giving place to a new concept: the Virtual Power Producer. Virtual Power Producers are multitechnology and multi-site heterogeneous entities. Virtual Power Producers should adopt organization and management methodologies so that they can make Distributed Generation a really profitable activity, able to participate in the market. In this paper we address the development of a multi-agent market simulator – MASCEM – able to study alternative coalitions of distributed producers in order to identify promising Virtual Power Producers in an electricity market.