920 resultados para Engenharia biomedica
Resumo:
É extensa a bibliografia dedicada a potenciais aplicações de materiais com mudança de fase na regulação térmica e no armazenamento de calor ou de frio. No entanto, a baixa condutividade térmica impõe limitações numa grande diversidade de aplicações com exigências críticas em termos de tempo de resposta curto ou com requisitos de elevada potência em ciclos de carga/descarga de calor latente. Foram desenvolvidos códigos numéricos no sentido de obter soluções precisas para descrever a cinética da transferência de calor com mudança de fase, com base em geometrias representativas, i.e. planar e esférica. Foram igualmente propostas soluções aproximadas, sendo identificados correspondentes critérios de validação em função das propriedades dos materiais de mudança de fase e de outros parâmetros relevantes tais como as escalas de tamanho e de tempo, etc. As referidas soluções permitiram identificar com rigor os fatores determinantes daquelas limitações, quantificar os correspondentes efeitos e estabelecer critérios de qualidade adequados para diferentes tipologias de potenciais aplicações. Os referidos critérios foram sistematizados de acordo com metodologias de seleção propostas por Ashby e co-autores, tendo em vista o melhor desempenho dos materiais em aplicações representativas, designadamente com requisitos ao nível de densidade energética, tempo de resposta, potência de carga/descarga e gama de temperaturas de operação. Nesta sistematização foram incluídos alguns dos compósitos desenvolvidos durante o presente trabalho. A avaliação das limitações acima mencionadas deu origem ao desenvolvimento de materiais compósitos para acumulação de calor ou frio, com acentuada melhoria de resposta térmica, mediante incorporação de uma fase com condutividade térmica muito superior à da matriz. Para este efeito, foram desenvolvidos modelos para otimizar a distribuição espacial da fase condutora, de modo a superar os limites de percolação previstos por modelos clássicos de condução em compósitos com distribuição aleatória, visando melhorias de desempenho térmico com reduzidas frações de fase condutora e garantindo que a densidade energética não é significativamente afetada. Os modelos elaborados correspondem a compósitos de tipo core-shell, baseados em microestruturas celulares da fase de elevada condutividade térmica, impregnadas com o material de mudança de fase propriamente dito. Além de visarem a minimização da fração de fase condutora e correspondentes custos, os modelos de compósitos propostos tiveram em conta a adequação a métodos de processamento versáteis, reprodutíveis, preferencialmente com base na emulsificação de líquidos orgânicos em suspensões aquosas ou outros processos de reduzidas complexidade e com base em materiais de baixo custo (material de mudança de fase e fase condutora). O design da distribuição microestrutural também considerou a possibilidade de orientação preferencial de fases condutoras com elevada anisotropia (p.e. grafite), mediante auto-organização. Outros estágios do projeto foram subordinados a esses objetivos de desenvolvimento de compósitos com resposta térmica otimizada, em conformidade com previsões dos modelos de compósitos de tipo core-shell, acima mencionadas. Neste enquadramento, foram preparados 3 tipos de compósitos com organização celular da fase condutora, com as seguintes características e metodologias: i) compósitos celulares parafina-grafite para acumulação de calor, preparados in-situ por emulsificação de uma suspensão de grafite em parafina fundida; ii) compósitos celulares parafina-Al2O3 para acumulação de calor, preparados por impregnação de parafina em esqueleto cerâmico celular de Al2O3; iii) compósitos celulares para acumulação de frio, obtidos mediante impregnação de matrizes celulares de grafite com solução de colagénio, após preparação prévia das matrizes de grafite celular. Os compósitos com esqueleto cerâmico (ii) requereram o desenvolvimento prévio de um método para o seu processamento, baseado na emulsificação de suspensões de Al2O3 em parafina fundida, com adequados aditivos dispersantes, tensioactivos e consolidantes do esqueleto cerâmico, tornando-o auto-suportável durante as fases posteriores de eliminação da parafina, até à queima a alta temperatura, originando cerâmicos celulares com adequada resistência mecânica. Os compósitos desenvolvidos apresentam melhorias significativos de condutividade térmica, atingindo ganhos superiores a 1 ordem de grandeza com frações de fase condutora inferior a 10 % vol. (4 W m-1 K-1), em virtude da organização core-shell e com o contributo adicional da anisotropia da grafite, mediante orientação preferencial. Foram ainda preparados compósitos de armazenamento de frio (iii), com orientação aleatória da fase condutora, obtidos mediante gelificação de suspensões de partículas de grafite em solução aquosa de colagénio. Apesar da estabilidade microestrutural e de forma, conferida por gelificação, estes compósitos confirmaram a esperada limitação dos compósitos com distribuição aleatória, em confronto com os ganhos alcançados com a organização de tipo core-shell.
Resumo:
The expectations of citizens from the Information Technologies (ITs) are increasing as the ITs have become integral part of our society, serving all kinds of activities whether professional, leisure, safety-critical applications or business. Hence, the limitations of the traditional network designs to provide innovative and enhanced services and applications motivated a consensus to integrate all services over packet switching infrastructures, using the Internet Protocol, so as to leverage flexible control and economical benefits in the Next Generation Networks (NGNs). However, the Internet is not capable of treating services differently while each service has its own requirements (e.g., Quality of Service - QoS). Therefore, the need for more evolved forms of communications has driven to radical changes of architectural and layering designs which demand appropriate solutions for service admission and network resources control. This Thesis addresses QoS and network control issues, aiming to improve overall control performance in current and future networks which classify services into classes. The Thesis is divided into three parts. In the first part, we propose two resource over-reservation algorithms, a Class-based bandwidth Over-Reservation (COR) and an Enhanced COR (ECOR). The over-reservation means reserving more bandwidth than a Class of Service (CoS) needs, so the QoS reservation signalling rate is reduced. COR and ECOR allow for dynamically defining over-reservation parameters for CoSs based on network interfaces resource conditions; they aim to reduce QoS signalling and related overhead without incurring CoS starvation or waste of bandwidth. ECOR differs from COR by allowing for optimizing control overhead minimization. Further, we propose a centralized control mechanism called Advanced Centralization Architecture (ACA), that uses a single state-full Control Decision Point (CDP) which maintains a good view of its underlying network topology and the related links resource statistics on real-time basis to control the overall network. It is very important to mention that, in this Thesis, we use multicast trees as the basis for session transport, not only for group communication purposes, but mainly to pin packets of a session mapped to a tree to follow the desired tree. Our simulation results prove a drastic reduction of QoS control signalling and the related overhead without QoS violation or waste of resources. Besides, we provide a generic-purpose analytical model to assess the impact of various parameters (e.g., link capacity, session dynamics, etc.) that generally challenge resource overprovisioning control. In the second part of this Thesis, we propose a decentralization control mechanism called Advanced Class-based resource OverpRovisioning (ACOR), that aims to achieve better scalability than the ACA approach. ACOR enables multiple CDPs, distributed at network edge, to cooperate and exchange appropriate control data (e.g., trees and bandwidth usage information) such that each CDP is able to maintain a good knowledge of the network topology and the related links resource statistics on real-time basis. From scalability perspective, ACOR cooperation is selective, meaning that control information is exchanged dynamically among only the CDPs which are concerned (correlated). Moreover, the synchronization is carried out through our proposed concept of Virtual Over-Provisioned Resource (VOPR), which is a share of over-reservations of each interface to each tree that uses the interface. Thus, each CDP can process several session requests over a tree without requiring synchronization between the correlated CDPs as long as the VOPR of the tree is not exhausted. Analytical and simulation results demonstrate that aggregate over-reservation control in decentralized scenarios keep low signalling without QoS violations or waste of resources. We also introduced a control signalling protocol called ACOR Protocol (ACOR-P) to support the centralization and decentralization designs in this Thesis. Further, we propose an Extended ACOR (E-ACOR) which aggregates the VOPR of all trees that originate at the same CDP, and more session requests can be processed without synchronization when compared with ACOR. In addition, E-ACOR introduces a mechanism to efficiently track network congestion information to prevent unnecessary synchronization during congestion time when VOPRs would exhaust upon every session request. The performance evaluation through analytical and simulation results proves the superiority of E-ACOR in minimizing overall control signalling overhead while keeping all advantages of ACOR, that is, without incurring QoS violations or waste of resources. The last part of this Thesis includes the Survivable ACOR (SACOR) proposal to support stable operations of the QoS and network control mechanisms in case of failures and recoveries (e.g., of links and nodes). The performance results show flexible survivability characterized by fast convergence time and differentiation of traffic re-routing under efficient resource utilization i.e. without wasting bandwidth. In summary, the QoS and architectural control mechanisms proposed in this Thesis provide efficient and scalable support for network control key sub-systems (e.g., QoS and resource control, traffic engineering, multicasting, etc.), and thus allow for optimizing network overall control performance.
Resumo:
O tratamento físico-químico de águas residuais, de origem industrial, mais comum é o tratamento baseado na adição de coagulante e floculante a um efluente. No entanto, o tratamento eletroquímico é um tipo de tratamento que tem vindo a ser explorado e estudado com mais ênfase ao longo dos últimos anos. O tratamento eletroquímico é uma tecnologia AOP (Processos de Oxidação Avançada) e divide-se em eletrólise direta (oxidação e redução) e indireta (eletrocoagulação-flotação e eletrooxidação). A eletrocoagulação e eletrooxidação divergem apenas pelo tipo de reações que ocorrem, devido ao material que constitui os elétrodos. São processos complexos com uma multiplicidade de mecanismos que operam sinergicamente para remover poluentes da água. Tendo em conta a sua complexidade e ainda dificuldade de compreensão, existem na literatura diferentes versões acerca de quais os mecanismos chave, assim como diversas configurações geométricas de reatores e elétrodos. Atualmente, este tipo de tratamento tem vindo a evoluir, tornando-se num método economicamente viável para o tratamento de uma grande variedade de águas residuais, nomeadamente, aquelas que possuem compostos recalcitrantes na sua composição. O presente trabalho foi realizado nas instalações da VentilAQUA S.A. e, tendo em conta a sua área de especialidade, o trabalho exposto focou-se no desenvolvimento de soluções técnicas de AOP, nomeadamente na área eletroquímica (eletrocoagulação e eletrooxidação),para estudo dos parâmetros operacionais numa nova configuração geométrica para os elétrodos. Tendo por base os contributos da revisão bibliográfica, o estudo incidiu num reator tubular, com elétrodos de inox dispostos de forma concêntrica, à mesma distância entre si. Com este reator foram executados variados testes, com diferentes efluentes, que permitiram obter resultados operacionais de otimização de funcionamento, tendo em vista a remoção de poluentes. O estudo financeiro associado permitiu concluir que a eletrooxidação é significativamente mais económica que o tratamento físico-químico, nas condições operacionais e para os efluentes tratados. Relativamente ao Acompanhamento e Gestão de ETAR’s (Capítulo 4) foi possível verificar que todos os casos em estudo apresentam uma boa eficiência de remoção de matéria orgânica, permitindo a descarga do seu efluente com uma carga poluente que cumpre com os requisitos legais de descarga em meio hídrico.
Resumo:
Tem sido frequentemente referido na literatura que os diplomados do ensino superior estão mal preparados para as exigências do mercado de trabalho. Face à situação económica actual, mais do que conhecimentos académicos, os estudantes necessitam desenvolver competências que promovam a sua empregabilidade e aprendizagem ao longo da vida. Isto é particularmente relevante no sector das Telecomunicações, face às constantes mutações tecnológicas e organizacionais. Estas competências são usualmente designadas por competências transversais, englobam o conjunto de competências que são comuns às diferentes actividades profissionais. Trata-se de competências imprescindíveis à empregabilidade, uma vez que permitem aos indivíduos agir numa multiplicidade de tarefas e funções. O presente trabalho propõe uma caracterização da construção das competências transversais no sector das Telecomunicações, tendo por base as percepções dos estudantes de engenharia, mas também as percepções das empresas do sector. No sentido de apoiar a fundamentação de estratégias práticas promotoras do desenvolvimento das competências transversais, o presente trabalho analisou os estilos de aprendizagem dos alunos de engenharia, tendo sido realizado um estudo para a adaptação do Index of Learning Styles (Felder & Soloman, n.d.) para a população portuguesa. No presente trabalho, 337 estudantes avaliaram o seu domínio num conjunto de competências transversais, bem como a sua importância para um futuro profissional. Desta avaliação foram identificados gaps de competências transversais que traduzem necessidades de desenvolvimento e formação. Foram também analisadas as relações entre as competências transversais e os estilos de aprendizagem dos estudantes, e as relações entre as competências transversais e um conjunto de variáveis psicológicas relacionadas com o desempenho competente: auto-eficácia, auto-regulação e optimismo. Representantes de empresas do sector das Telecomunicações avaliaram a importância das mesmas competências transversais para as práticas de negócio. A interpretação dos resultados, bem como as possíveis implicações destes no desenvolvimento de estratégias promotoras da construção das competências transversais, são discutidos à luz da literatura relevante.
Resumo:
Nos últimos anos, o número de vítimas de acidentes de tráfego por milhões de habitantes em Portugal tem sido mais elevado do que a média da União Europeia. Ao nível nacional torna-se premente uma melhor compreensão dos dados de acidentes e sobre o efeito do veículo na gravidade do mesmo. O objetivo principal desta investigação consistiu no desenvolvimento de modelos de previsão da gravidade do acidente, para o caso de um único veículo envolvido e para caso de uma colisão, envolvendo dois veículos. Além disso, esta investigação compreendeu o desenvolvimento de uma análise integrada para avaliar o desempenho do veículo em termos de segurança, eficiência energética e emissões de poluentes. Os dados de acidentes foram recolhidos junto da Guarda Nacional Republicana Portuguesa, na área metropolitana do Porto para o período de 2006-2010. Um total de 1,374 acidentes foram recolhidos, 500 acidentes envolvendo um único veículo e 874 colisões. Para a análise da segurança, foram utilizados modelos de regressão logística. Para os acidentes envolvendo um único veículo, o efeito das características do veículo no risco de feridos graves e/ou mortos (variável resposta definida como binária) foi explorado. Para as colisões envolvendo dois veículos foram criadas duas variáveis binárias adicionais: uma para prever a probabilidade de feridos graves e/ou mortos num dos veículos (designado como veículo V1) e outra para prever a probabilidade de feridos graves e/ou mortos no outro veículo envolvido (designado como veículo V2). Para ultrapassar o desafio e limitações relativas ao tamanho da amostra e desigualdade entre os casos analisados (apenas 5.1% de acidentes graves), foi desenvolvida uma metodologia com base numa estratégia de reamostragem e foram utilizadas 10 amostras geradas de forma aleatória e estratificada para a validação dos modelos. Durante a fase de modelação, foi analisado o efeito das características do veículo, como o peso, a cilindrada, a distância entre eixos e a idade do veículo. Para a análise do consumo de combustível e das emissões, foi aplicada a metodologia CORINAIR. Posteriormente, os dados das emissões foram modelados de forma a serem ajustados a regressões lineares. Finalmente, foi desenvolvido um indicador de análise integrada (denominado “SEG”) que proporciona um método de classificação para avaliar o desempenho do veículo ao nível da segurança rodoviária, consumos e emissões de poluentes.Face aos resultados obtidos, para os acidentes envolvendo um único veículo, o modelo de previsão do risco de gravidade identificou a idade e a cilindrada do veículo como estatisticamente significativas para a previsão de ocorrência de feridos graves e/ou mortos, ao nível de significância de 5%. A exatidão do modelo foi de 58.0% (desvio padrão (D.P.) 3.1). Para as colisões envolvendo dois veículos, ao prever a probabilidade de feridos graves e/ou mortos no veículo V1, a cilindrada do veículo oposto (veículo V2) aumentou o risco para os ocupantes do veículo V1, ao nível de significância de 10%. O modelo para prever o risco de gravidade no veículo V1 revelou um bom desempenho, com uma exatidão de 61.2% (D.P. 2.4). Ao prever a probabilidade de feridos graves e/ou mortos no veículo V2, a cilindrada do veículo V1 aumentou o risco para os ocupantes do veículo V2, ao nível de significância de 5%. O modelo para prever o risco de gravidade no veículo V2 também revelou um desempenho satisfatório, com uma exatidão de 40.5% (D.P. 2.1). Os resultados do indicador integrado SEG revelaram que os veículos mais recentes apresentam uma melhor classificação para os três domínios: segurança, consumo e emissões. Esta investigação demonstra que não existe conflito entre a componente da segurança, a eficiência energética e emissões relativamente ao desempenho dos veículos.
Resumo:
Este trabalho de investigação foi realizado com o objetivo geral de se descobrir quais são os Fatores Críticos de Sucesso e os aspectos chaves de qualidade e sustentabilidade, a serem considerados no ensino de Engenharia de Produção em um IES privada no Brasil. No presente estudo de caso executou-se uma revisão teórica sistemática nas áreas temáticas da pesquisa, que propiciou a criação de um framework que define os 7 Fatores Críticos de Sucesso ao ensino de Engenharia de Produção, apontando sua divisão nas áreas de qualidade e sustentabilidade, e, por fim, a subdivisão em 6 subáreas de ações operacionais a serem consideradas pela gestão de IES para alcançar melhorias de desempenho em suas atividades. Posteriormente ao levantamento teórico, foi realizado o estudo de caso propriamente dito, que dividiu-se em duas etapas, uma qualitativa e uma quantitativa. Na primeira elaborou-se um roteiro de entrevista em profundidade aplicada a um terço do corpo docente do curso em estudo, captando a impressão destes sobre as temáticas em estudo voltadas à execução de suas atividades. Os dados obtidos foram organizados com o auxílio do software WebQDA, gerando ideias do que deve ser trabalhado para melhorar as condições e resultados do referido curso. A etapa quantitativa foi iniciada com a elaboração de um questionário on-line, desenvolvido a partir do enquadramento teórico e da informação obtida na parte qualitativa do estudo, que envolveu respostas de cerca de dois terços dos discentes de Engenharia de Produção. A informação foi recolhida através de um formulário disponibilizado pelo googledocs, e os resultados foram analisados com o auxílio do software SPSS, permitindo conhecer as expectativas e percepções dos alunos e, consequentemente, os gaps provenientes a serem trabalhados. Finalmente, o estudo revelou que para aspectos de qualidade, os pontos de tangibilidade, como infraestrutura geral, laboratórios técnicos específicos, cantina e biblioteca, precisam ser melhorados, e, para aspectos de sustentabilidade, surge a necessidade de melhorar as formas de divulgação do curso, captação de alunos e propostas para a retenção dos mesmos para outros cursos e projetos da IES. Deste modo, o trabalho desenvolvido atingiu seu objetivo e se mostrou válido para aplicações na IES e em outros cursos da mesma (com as devidas adaptações). Fica como recomendação final aplicar o estudo periodicamente para se captar as mudanças situacionais que ocorrerem, manter um controle eficiente dos aspectos estudados e gerar melhorias contínuas no curso estudado e na própria IES.
Resumo:
The increased capabilities (e.g., processing, storage) of portable devices along with the constant need of users to retrieve and send information have introduced a new form of communication. Users can seamlessly exchange data by means of opportunistic contacts among them and this is what characterizes the opportunistic networks (OppNets). OppNets allow users to communicate even when an end-to-end path may not exist between them. Since 2007, there has been a trend to improve the exchange of data by considering social similarity metrics. Social relationships, shared interests, and popularity are examples of such metrics that have been employed successfully: as users interact based on relationships and interests, this information can be used to decide on the best next forwarders of information. This Thesis work combines the features of today's devices found in the regular urban environment with the current social-awareness trend in the context of opportunistic routing. To achieve this goal, this work was divided into di erent tasks that map to a set of speci c objectives, leading to the following contributions: i) an up-to-date opportunistic routing taxonomy; ii) a universal evaluation framework that aids in devising and testing new routing proposals; iii) three social-aware utility functions that consider the dynamic user behavior and can be easily incorporated to other routing proposals; iv) two opportunistic routing proposals based on the users' daily routines and on the content traversing the network and interest of users in such content; and v) a structure analysis of the social-based network formed based on the approaches devised in this work.
Resumo:
This thesis addresses the problem of word learning in computational agents. The motivation behind this work lies in the need to support language-based communication between service robots and their human users, as well as grounded reasoning using symbols relevant for the assigned tasks. The research focuses on the problem of grounding human vocabulary in robotic agent’s sensori-motor perception. Words have to be grounded in bodily experiences, which emphasizes the role of appropriate embodiments. On the other hand, language is a cultural product created and acquired through social interactions. This emphasizes the role of society as a source of linguistic input. Taking these aspects into account, an experimental scenario is set up where a human instructor teaches a robotic agent the names of the objects present in a visually shared environment. The agent grounds the names of these objects in visual perception. Word learning is an open-ended problem. Therefore, the learning architecture of the agent will have to be able to acquire words and categories in an openended manner. In this work, four learning architectures were designed that can be used by robotic agents for long-term and open-ended word and category acquisition. The learning methods used in these architectures are designed for incrementally scaling-up to larger sets of words and categories. A novel experimental evaluation methodology, that takes into account the openended nature of word learning, is proposed and applied. This methodology is based on the realization that a robot’s vocabulary will be limited by its discriminatory capacity which, in turn, depends on its sensors and perceptual capabilities. An extensive set of systematic experiments, in multiple experimental settings, was carried out to thoroughly evaluate the described learning approaches. The results indicate that all approaches were able to incrementally acquire new words and categories. Although some of the approaches could not scale-up to larger vocabularies, one approach was shown to learn up to 293 categories, with potential for learning many more.
Resumo:
Over the years, the increased search and exchange of information lead to an increase of traffic intensity in todays optical communication networks. Coherent communications, using the amplitude and phase of the signal, reappears as one of the transmission techniques to increase the spectral efficiency and throughput of optical channels. In this context, this work present a study on format conversion of modulated signals using MZI-SOAs, based exclusively on all- optical techniques through wavelength conversion. This approach, when applied in interconnection nodes between optical networks with different bit rates and modulation formats, allow a better efficiency and scalability of the network. We start with an experimental characterization of the static and dynamic properties of the MZI-SOA. Then, we propose a semi-analytical model to describe the evolution of phase and amplitude at the output of the MZI-SOA. The model’s coefficients are obtained using a multi-objective genetic algorithm. We validate the model experimentally, by exploring the dependency of the optical signal with the operational parameters of the MZI-SOA. We also propose an all-optical technique for the conversion of amplitude modulation signals to a continuous phase modulation format. Finally, we study the potential of MZI-SOAs for the conversion of amplitude signals to QPSK and QAM signals. We show the dependency of the conversion process with the operational parameters deviation from the optimal values. The technique is experimentally validated for QPSK modulation.
Resumo:
Future emerging market trends head towards positioning based services placing a new perspective on the way we obtain and exploit positioning information. On one hand, innovations in information technology and wireless communication systems enabled the development of numerous location based applications such as vehicle navigation and tracking, sensor networks applications, home automation, asset management, security and context aware location services. On the other hand, wireless networks themselves may bene t from localization information to improve the performances of di erent network layers. Location based routing, synchronization, interference cancellation are prime examples of applications where location information can be useful. Typical positioning solutions rely on measurements and exploitation of distance dependent signal metrics, such as the received signal strength, time of arrival or angle of arrival. They are cheaper and easier to implement than the dedicated positioning systems based on ngerprinting, but at the cost of accuracy. Therefore intelligent localization algorithms and signal processing techniques have to be applied to mitigate the lack of accuracy in distance estimates. Cooperation between nodes is used in cases where conventional positioning techniques do not perform well due to lack of existing infrastructure, or obstructed indoor environment. The objective is to concentrate on hybrid architecture where some nodes have points of attachment to an infrastructure, and simultaneously are interconnected via short-range ad hoc links. The availability of more capable handsets enables more innovative scenarios that take advantage of multiple radio access networks as well as peer-to-peer links for positioning. Link selection is used to optimize the tradeo between the power consumption of participating nodes and the quality of target localization. The Geometric Dilution of Precision and the Cramer-Rao Lower Bound can be used as criteria for choosing the appropriate set of anchor nodes and corresponding measurements before attempting location estimation itself. This work analyzes the existing solutions for node selection in order to improve localization performance, and proposes a novel method based on utility functions. The proposed method is then extended to mobile and heterogeneous environments. Simulations have been carried out, as well as evaluation with real measurement data. In addition, some speci c cases have been considered, such as localization in ill-conditioned scenarios and the use of negative information. The proposed approaches have shown to enhance estimation accuracy, whilst signi cantly reducing complexity, power consumption and signalling overhead.
Resumo:
The rapid evolution and proliferation of a world-wide computerized network, the Internet, resulted in an overwhelming and constantly growing amount of publicly available data and information, a fact that was also verified in biomedicine. However, the lack of structure of textual data inhibits its direct processing by computational solutions. Information extraction is the task of text mining that intends to automatically collect information from unstructured text data sources. The goal of the work described in this thesis was to build innovative solutions for biomedical information extraction from scientific literature, through the development of simple software artifacts for developers and biocurators, delivering more accurate, usable and faster results. We started by tackling named entity recognition - a crucial initial task - with the development of Gimli, a machine-learning-based solution that follows an incremental approach to optimize extracted linguistic characteristics for each concept type. Afterwards, Totum was built to harmonize concept names provided by heterogeneous systems, delivering a robust solution with improved performance results. Such approach takes advantage of heterogenous corpora to deliver cross-corpus harmonization that is not constrained to specific characteristics. Since previous solutions do not provide links to knowledge bases, Neji was built to streamline the development of complex and custom solutions for biomedical concept name recognition and normalization. This was achieved through a modular and flexible framework focused on speed and performance, integrating a large amount of processing modules optimized for the biomedical domain. To offer on-demand heterogenous biomedical concept identification, we developed BeCAS, a web application, service and widget. We also tackled relation mining by developing TrigNER, a machine-learning-based solution for biomedical event trigger recognition, which applies an automatic algorithm to obtain the best linguistic features and model parameters for each event type. Finally, in order to assist biocurators, Egas was developed to support rapid, interactive and real-time collaborative curation of biomedical documents, through manual and automatic in-line annotation of concepts and relations. Overall, the research work presented in this thesis contributed to a more accurate update of current biomedical knowledge bases, towards improved hypothesis generation and knowledge discovery.
Resumo:
This work investigates low cost localization systems (LS) based on received signal strength (RSS) and integrated with different types of antennas with main emphasis on sectorial antennas. The last few years have witnessed an outstanding growth in wireless sensor networks (WSN). Among its various possible applications, the localization field became a major area of research. The localization techniques based on RSS are characterized by simplicity and low cost of integration. The integration of LS based on RSS and sectorial antennas (SA) was proven to provide an effective solution for reducing the number of required nodes of the networks and allows the combination of several techniques, such as RSS and angle of arrival (AoA). This PhD thesis focuses on studying techniques, antennas and protocols that best meet the needs of each LS with main focus on low cost systems based on RSS and AoA. Firstly there are studied localization techniques and system that best suit the requirements of the user and the antennas that are most appropriate according to the nature of the signal. In this step it is intended to provide a fundamental understanding of the undertaken work. Then the developed antennas are presented according to the following categories: sectorial and microstrip antennas. Two sectorial antennas are presented: a narrowband antenna operating at 2.4 to 2.5 GHz and a broadband antenna operating at 800MHz-2.4GHz. The low cost printed antennas were designed to operate at 5 GHz, which may be used for vehicular communication. After presenting the various antennas, several prototypes of indoor/outdoor LS are implemented and analyzed. Localization protocols are also proposed, one based on simplicity and low power, and the other on interoperability with different types of antennas and system requirements.
Resumo:
All-optical solutions for switching and routing packet-based traffic are crucial for realizing a truly transparent network. To meet the increasing requirements for higher bandwidth, such optical packet switched networks may require the implementation of digital functions in the physical layer. This scenario stimulated us to research and develop innovative high-speed all-optical storage memories, focusing mainly on bistables whose state switching is triggered by a pulsed clock signal. In clocked devices, a synchronization signal is responsible for controlling the enabling of the bistable. This thesis also presents novel solutions to implement optical logic gates, which are basic building blocks of any processing system and a fundamental element for the development of complex processing functionalities. Most of the proposed schemes developed in this work are based on SOA-MZI structures due to their inherent characteristics such as, high extinction ratio, high operation speed, high integration capability and compactness. We addressed the experimental implementation of an all-optical packet routing scheme, with contention resolution capability, using interconnected SOAMZIs. The impact on the system performance of the reminiscent power of the blocked packets, from the non ideal switching performed by the SOA-MZIs, was also assessed.
Resumo:
The performance of real-time networks is under continuous improvement as a result of several trends in the digital world. However, these tendencies not only cause improvements, but also exacerbates a series of unideal aspects of real-time networks such as communication latency, jitter of the latency and packet drop rate. This Thesis focuses on the communication errors that appear on such realtime networks, from the point-of-view of automatic control. Specifically, it investigates the effects of packet drops in automatic control over fieldbuses, as well as the architectures and optimal techniques for their compensation. Firstly, a new approach to address the problems that rise in virtue of such packet drops, is proposed. This novel approach is based on the simultaneous transmission of several values in a single message. Such messages can be from sensor to controller, in which case they are comprised of several past sensor readings, or from controller to actuator in which case they are comprised of estimates of several future control values. A series of tests reveal the advantages of this approach. The above-explained approach is then expanded as to accommodate the techniques of contemporary optimal control. However, unlike the aforementioned approach, that deliberately does not send certain messages in order to make a more efficient use of network resources; in the second case, the techniques are used to reduce the effects of packet losses. After these two approaches that are based on data aggregation, it is also studied the optimal control in packet dropping fieldbuses, using generalized actuator output functions. This study ends with the development of a new optimal controller, as well as the function, among the generalized functions that dictate the actuator’s behaviour in the absence of a new control message, that leads to the optimal performance. The Thesis also presents a different line of research, related with the output oscillations that take place as a consequence of the use of classic co-design techniques of networked control. The proposed algorithm has the goal of allowing the execution of such classical co-design algorithms without causing an output oscillation that increases the value of the cost function. Such increases may, under certain circumstances, negate the advantages of the application of the classical co-design techniques. A yet another line of research, investigated algorithms, more efficient than contemporary ones, to generate task execution sequences that guarantee that at least a given number of activated jobs will be executed out of every set composed by a predetermined number of contiguous activations. This algorithm may, in the future, be applied to the generation of message transmission patterns in the above-mentioned techniques for the efficient use of network resources. The proposed task generation algorithm is better than its predecessors in the sense that it is capable of scheduling systems that cannot be scheduled by its predecessor algorithms. The Thesis also presents a mechanism that allows to perform multi-path routing in wireless sensor networks, while ensuring that no value will be counted in duplicate. Thereby, this technique improves the performance of wireless sensor networks, rendering them more suitable for control applications. As mentioned before, this Thesis is centered around techniques for the improvement of performance of distributed control systems in which several elements are connected through a fieldbus that may be subject to packet drops. The first three approaches are directly related to this topic, with the first two approaching the problem from an architectural standpoint, whereas the third one does so from more theoretical grounds. The fourth approach ensures that the approaches to this and similar problems that can be found in the literature that try to achieve goals similar to objectives of this Thesis, can do so without causing other problems that may invalidate the solutions in question. Then, the thesis presents an approach to the problem dealt with in it, which is centered in the efficient generation of the transmission patterns that are used in the aforementioned approaches.
Resumo:
Network virtualisation is seen as a promising approach to overcome the so-called “Internet impasse” and bring innovation back into the Internet, by allowing easier migration towards novel networking approaches as well as the coexistence of complementary network architectures on a shared infrastructure in a commercial context. Recently, the interest from the operators and mainstream industry in network virtualisation has grown quite significantly, as the potential benefits of virtualisation became clearer, both from an economical and an operational point of view. In the beginning, the concept has been mainly a research topic and has been materialized in small-scale testbeds and research network environments. This PhD Thesis aims to provide the network operator with a set of mechanisms and algorithms capable of managing and controlling virtual networks. To this end, we propose a framework that aims to allocate, monitor and control virtual resources in a centralized and efficient manner. In order to analyse the performance of the framework, we performed the implementation and evaluation on a small-scale testbed. To enable the operator to make an efficient allocation, in real-time, and on-demand, of virtual networks onto the substrate network, it is proposed a heuristic algorithm to perform the virtual network mapping. For the network operator to obtain the highest profit of the physical network, it is also proposed a mathematical formulation that aims to maximize the number of allocated virtual networks onto the physical network. Since the power consumption of the physical network is very significant in the operating costs, it is important to make the allocation of virtual networks in fewer physical resources and onto physical resources already active. To address this challenge, we propose a mathematical formulation that aims to minimize the energy consumption of the physical network without affecting the efficiency of the allocation of virtual networks. To minimize fragmentation of the physical network while increasing the revenue of the operator, it is extended the initial formulation to contemplate the re-optimization of previously mapped virtual networks, so that the operator has a better use of its physical infrastructure. It is also necessary to address the migration of virtual networks, either for reasons of load balancing or for reasons of imminent failure of physical resources, without affecting the proper functioning of the virtual network. To this end, we propose a method based on cloning techniques to perform the migration of virtual networks across the physical infrastructure, transparently, and without affecting the virtual network. In order to assess the resilience of virtual networks to physical network failures, while obtaining the optimal solution for the migration of virtual networks in case of imminent failure of physical resources, the mathematical formulation is extended to minimize the number of nodes migrated and the relocation of virtual links. In comparison with our optimization proposals, we found out that existing heuristics for mapping virtual networks have a poor performance. We also found that it is possible to minimize the energy consumption without penalizing the efficient allocation. By applying the re-optimization on the virtual networks, it has been shown that it is possible to obtain more free resources as well as having the physical resources better balanced. Finally, it was shown that virtual networks are quite resilient to failures on the physical network.