897 resultados para Robotics Education, Distributed Control, Automonous Robots, Programming, Computer Architecture
Resumo:
The development of versatile bioactive surfaces able to emulate in vivo conditions is of enormous importance to the future of cell and tissue therapy. Tuning cell behaviour on two-dimensional surfaces so that the cells perform as if they were in a natural three-dimensional tissue represents a significant challenge, but one that must be met if the early promise of cell and tissue therapy is to be fully realised. Due to the inherent complexities involved in the manufacture of biomimetic three-dimensional substrates, the scaling up of engineered tissue-based therapies may be simpler if based upon proven two-dimensional culture systems. In this work, we developed new coating materials composed of the self-assembling peptide amphiphiles (PAs) C16G3RGD (RGD) and C16G3RGDS (RGDS) shown to control cell adhesion and tissue architecture while avoiding the use of serum. When mixed with the C16ETTES diluent PA at 13 : 87 (mol mol-1) ratio at 1.25 times 10-3 M, the bioactive {PAs} were shown to support optimal adhesion, maximal proliferation, and prolonged viability of human corneal stromal fibroblasts ({hCSFs)}, while improving the cell phenotype. These {PAs} also provided stable adhesive coatings on highly-hydrophobic surfaces composed of striated polytetrafluoroethylene ({PTFE)}, significantly enhancing proliferation of aligned cells and increasing the complexity of the produced tissue. The thickness and structure of this highly-organised tissue were similar to those observed in vivo, comprising aligned newly-deposited extracellular matrix. As such, the developed coatings can constitute a versatile biomaterial for applications in cell biology, tissue engineering, and regenerative medicine requiring serum-free conditions.
Resumo:
A Execução Condicional Dinâmica (DCE) é uma alternativa para redução dos custos relacionados a desvios previstos incorretamente. A idéia básica é buscar todos os fluxos produzidos por um desvio que obedecem algumas restrições relativas à complexidade e tamanho. Como conseqüência, um número menor de previsões é executado, e assim, um número mais baixo de desvios é incorretamente previsto. Contudo, tal como outras soluções multi-fluxo, o DCE requer uma estrutura de controle mais complexa. Na arquitetura DCE, é observado que várias réplicas da mesma instrução são despachadas para as unidades funcionais, bloqueando recursos que poderiam ser utilizados por outras instruções. Essas réplicas são geradas após o ponto de convergência dos diversos fluxos em execução e são necessárias para garantir a semântica correta entre instruções dependentes de dados. Além disso, o DCE continua produzindo réplicas até que o desvio que gerou os fluxos seja resolvido. Assim, uma seção completa do código pode ser replicado, reduzindo o desempenho. Uma alternativa natural para esse problema é reusar essas seções (ou traços) que são replicadas. O objetivo desse trabalho é analisar e avaliar a efetividade do reuso de valores na arquitetura DCE. Como será apresentado, o princípio do reuso, em diferentes granularidades, pode reduzir efetivamente o problema das réplicas e levar a aumentos de desempenho.
Resumo:
Artificial neural networks are usually applied to solve complex problems. In problems with more complexity, by increasing the number of layers and neurons, it is possible to achieve greater functional efficiency. Nevertheless, this leads to a greater computational effort. The response time is an important factor in the decision to use neural networks in some systems. Many argue that the computational cost is higher in the training period. However, this phase is held only once. Once the network trained, it is necessary to use the existing computational resources efficiently. In the multicore era, the problem boils down to efficient use of all available processing cores. However, it is necessary to consider the overhead of parallel computing. In this sense, this paper proposes a modular structure that proved to be more suitable for parallel implementations. It is proposed to parallelize the feedforward process of an RNA-type MLP, implemented with OpenMP on a shared memory computer architecture. The research consistes on testing and analizing execution times. Speedup, efficiency and parallel scalability are analyzed. In the proposed approach, by reducing the number of connections between remote neurons, the response time of the network decreases and, consequently, so does the total execution time. The time required for communication and synchronization is directly linked to the number of remote neurons in the network, and so it is necessary to investigate which one is the best distribution of remote connections
Resumo:
Nowadays several electronics devices support digital videos. Some examples of these devices are cellphones, digital cameras, video cameras and digital televisions. However, raw videos present a huge amount of data, millions of bits, for their representation as the way they were captured. To store them in its primary form it would be necessary a huge amount of disk space and a huge bandwidth to allow the transmission of these data. The video compression becomes essential to make possible information storage and transmission. Motion Estimation is a technique used in the video coder that explores the temporal redundancy present in video sequences to reduce the amount of data necessary to represent the information. This work presents a hardware architecture of a motion estimation module for high resolution videos according to H.264/AVC standard. The H.264/AVC is the most advanced video coder standard, with several new features which allow it to achieve high compression rates. The architecture presented in this work was developed to provide a high data reuse. The data reuse schema adopted reduces the bandwidth required to execute motion estimation. The motion estimation is the task responsible for the largest share of the gains obtained with the H.264/AVC standard so this module is essential for final video coder performance. This work is included in Rede H.264 project which aims to develop Brazilian technology for Brazilian System of Digital Television
Resumo:
Os sensores inteligentes são dispositivos que se diferenciam dos sensores comuns por apresentar capacidade de processamento sobre os dados monitorados. Eles tipicamente são compostos por uma fonte de alimentação, transdutores (sensores e atuadores), memória, processador e transceptor. De acordo com o padrão IEEE 1451 um sensor inteligente pode ser dividido em módulos TIM e NCAP que devem se comunicar através de uma interface padronizada chamada TII. O módulo NCAP é a parte do sensor inteligente que comporta o processador. Portanto, ele é o responsável por atribuir a característica de inteligência ao sensor. Existem várias abordagens que podem ser utilizadas para o desenvolvimento desse módulo, dentre elas se destacam aquelas que utilizam microcontroladores de baixo custo e/ou FPGA. Este trabalho aborda o desenvolvimento de uma arquitetura hardware/software para um módulo NCAP segundo o padrão IEEE 1451.1. A infra-estrutura de hardware é composta por um driver de interface RS-232, uma memória RAM de 512kB, uma interface TII, o processador embarcado NIOS II e um simulador do módulo TIM. Para integração dos componentes de hardware é utilizada ferramenta de integração automática SOPC Builder. A infra-estrutura de software é composta pelo padrão IEEE 1451.1 e pela aplicação especí ca do NCAP que simula o monitoramento de pressão e temperatura em poços de petróleo com o objetivo de detectar vazamento. O módulo proposto é embarcado em uma FPGA e para a sua prototipação é usada a placa DE2 da Altera que contém a FPGA Cyclone II EP2C35F672C6. O processador embarcado NIOS II é utilizado para dar suporte à infra-estrutura de software do NCAP que é desenvolvido na linguagem C e se baseia no padrão IEEE 1451.1. A descrição do comportamento da infra-estrutura de hardware é feita utilizando a linguagem VHDL
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
This paper describes a data mining environment for knowledge discovery in bioinformatics applications. The system has a generic kernel that implements the mining functions to be applied to input primary databases, with a warehouse architecture, of biomedical information. Both supervised and unsupervised classification can be implemented within the kernel and applied to data extracted from the primary database, with the results being suitably stored in a complex object database for knowledge discovery. The kernel also includes a specific high-performance library that allows designing and applying the mining functions in parallel machines. The experimental results obtained by the application of the kernel functions are reported. © 2003 Elsevier Ltd. All rights reserved.
Resumo:
This paper presents results from an efficient approach to an automatic detection and extraction of human faces from images with any color, texture or objects in background, that consist in find isosceles triangles formed by the eyes and mouth.
Resumo:
Pós-graduação em Ciência da Computação - IBILCE
Resumo:
Huge image collections are becoming available lately. In this scenario, the use of Content-Based Image Retrieval (CBIR) systems has emerged as a promising approach to support image searches. The objective of CBIR systems is to retrieve the most similar images in a collection, given a query image, by taking into account image visual properties such as texture, color, and shape. In these systems, the effectiveness of the retrieval process depends heavily on the accuracy of ranking approaches. Recently, re-ranking approaches have been proposed to improve the effectiveness of CBIR systems by taking into account the relationships among images. The re-ranking approaches consider the relationships among all images in a given dataset. These approaches typically demands a huge amount of computational power, which hampers its use in practical situations. On the other hand, these methods can be massively parallelized. In this paper, we propose to speedup the computation of the RL-Sim algorithm, a recently proposed image re-ranking approach, by using the computational power of Graphics Processing Units (GPU). GPUs are emerging as relatively inexpensive parallel processors that are becoming available on a wide range of computer systems. We address the image re-ranking performance challenges by proposing a parallel solution designed to fit the computational model of GPUs. We conducted an experimental evaluation considering different implementations and devices. Experimental results demonstrate that significant performance gains can be obtained. Our approach achieves speedups of 7x from serial implementation considering the overall algorithm and up to 36x on its core steps.
Resumo:
Techniques of image combination, with extraction of objects to set a final scene, are very used in applications from photos montages to cinematographic productions. These techniques are called digital matting. With them is possible to decrease the cost of productions, because it is not necessary for the actor to be filmed in the location where the final scene occurs. This feature also favors its use in programs made to digital television, which demands a high quality image. Many digital matting algorithms use markings done on the images, to demarcate what is the foreground, the background and the uncertainty areas. This marking is called trimap, which is a triple map containing these three informations. The trimap is done, typically, from manual markings. In this project, methods were created that can be used in digital matting algorithms, with restriction of time and without human interaction, that is, the creation of an algorithm that generates the trimap automatically. This last one can be generated from the difference between a color of an arbitrary background and the foreground, or by using a depth map. It was also created a matting method, based on the Geodesic Matting (BAI; SAPIRO, 2009), which has an inferior processing time then the original one. Aiming to improve the performance of the applications that generates the trimap and of the algorithms that generates the alphamap (map that associates a value to the transparency of each pixel of the image), allowing its use in applications with time restrictions, it was used the CUDA architecture. Taking advantage, this way, of the computational power and the features of the GPGPU, which is massively parallel
Resumo:
This project points out a brief overview of several concepts, as Renewable Energy Resources, Distributed Energy Resources, Distributed Generation, and describes the general architecture of an electrical microgrid, isolated or connected to the Medium Voltage Network. Moreover, the project focuses on a project carried out by GRECDH Department in collaboration with CITCEA Department, both belonging to Universitat Politécnica de Catalunya: it concerns isolated microgrids employing renewable energy resources in two communities in northern Peru. Several solutions found using optimization software regarding different generation systems (wind and photovoltaic) and different energy demand scenarios are commented and analyzed from an electrical point of view. Furthermore, there are some proposals to improve microgrid performances, in particular to increase voltage values for each load connected to the microgrid. The extra costs required by the proposed solutions are calculated and their effect on the total microgrid cost are taken into account; finally there are some considerations about the impact the project has on population and on people's daily life.
Resumo:
I sistemi software nati dall'esigenza di supportare agevolmente ed efficacemente il lavoro cooperativo, in particolare quelli orientati al supporto di azioni di soccorso in scenari di emergenza, appaiono tutt'ora fortemente limitati e frammentati. In molti casi vengono affrontate solamente specifiche dimensioni del problema complessivo, anche se il livello al quale è giunto lo sviluppo tecnologico e i risultati osservati in ambito di ricerca permettono di delineare soluzioni complete e significative per l'impiego in ambiti reali. Tale tipologia di sistemi è stata scelta per il grande interesse che desta sia dal punto di vista accademico, essendo costituita da molteplici sotto--sistemi spesso eterogenei che debbono necessariamente interagire e supportare l'azione umana, sia dal punto di vista industriale, interpretando la necessità crescente di iniettare nel maggior numero possibile di livelli sociali la forte dipendenza (il supporto allo stesso tempo) dalle scienze tecnologiche ed informatiche, per rafforzare e talvolta estendere le possibilità dell'essere umano in quanto tale. Dopo una prima fase in cui verrà delineato un quadro concettuale piuttosto dettagliato circa i principali elementi e problematiche che caratterizzano la classe di sistemi considerati, sarà dato spazio alla validazione di tali principi e considerazioni emerse, confrontandosi con la progettazione e sviluppo in forma prototipale di un sotto--sistema relativo ad un caso di studio reale, significativo per l'ambito di applicazione, nato dalla collaborazione con l'Università degli Studi di Bologna di un'azienda della regione Emilia--Romagna. Il sistema software realizzato vuole essere innanzi tutto la risposta alle esigenze emerse nel caso di studio trattato, in modo tale da potersi sostituire agli attuali limitati supporti alla cooperazione, ma anche un esperimento che possa essere considerato un artefatto centrale da utilizzare come base di conoscenza condivisa, in cui vengano fattorizzati i concetti e meccanismi chiave, fondamentali per sviluppi futuri.
Resumo:
In this paper the software architecture of a framework which simplifies the development of applications in the area of Virtual and Augmented Reality is presented. It is based on VRML/X3D to enable rendering of audio-visual information. We extended our VRML rendering system by a device management system that is based on the concept of a data-flow graph. The aim of the system is to create Mixed Reality (MR) applications simply by plugging together small prefabricated software components, instead of compiling monolithic C++ applications. The flexibility and the advantages of the presented framework are explained on the basis of an exemplary implementation of a classic Augmented Realityapplication and its extension to a collaborative remote expert scenario.
Consolidation of a wsn and minimax method to rapidly neutralise intruders in strategic installations
Resumo:
Due to the sensitive international situation caused by still-recent terrorist attacks, there is a common need to protect the safety of large spaces such as government buildings, airports and power stations. To address this problem, developments in several research fields, such as video and cognitive audio, decision support systems, human interface, computer architecture, communications networks and communications security, should be integrated with the goal of achieving advanced security systems capable of checking all of the specified requirements and spanning the gap that presently exists in the current market. This paper describes the implementation of a decision system for crisis management in infrastructural building security. Specifically, it describes the implementation of a decision system in the management of building intrusions. The positions of the unidentified persons are reported with the help of a Wireless Sensor Network (WSN). The goal is to achieve an intelligent system capable of making the best decision in real time in order to quickly neutralise one or more intruders who threaten strategic installations. It is assumed that the intruders’ behaviour is inferred through sequences of sensors’ activations and their fusion. This article presents a general approach to selecting the optimum operation from the available neutralisation strategies based on a Minimax algorithm. The distances among different scenario elements will be used to measure the risk of the scene, so a path planning technique will be integrated in order to attain a good performance. Different actions to be executed over the elements of the scene such as moving a guard, blocking a door or turning on an alarm will be used to neutralise the crisis. This set of actions executed to stop the crisis is known as the neutralisation strategy. Finally, the system has been tested in simulations of real situations, and the results have been evaluated according to the final state of the intruders. In 86.5% of the cases, the system achieved the capture of the intruders, and in 59.25% of the cases, they were intercepted before they reached their objective.