986 resultados para Dynamic Software
Resumo:
Dissertação para obtenção do Grau de Doutor em Alterações Climáticas e Políticas de Desenvolvimento Sustentável
Resumo:
Relatório de Estágio apresentado para cumprimento dos requisitos necessários à obtenção do grau de Mestre em Tradução – Especialização em Inglês
Resumo:
Nowadays, existing 3D scanning cameras and microscopes in the market use digital or discrete sensors, such as CCDs or CMOS for object detection applications. However, these combined systems are not fast enough for some application scenarios since they require large data processing resources and can be cumbersome. Thereby, there is a clear interest in exploring the possibilities and performances of analogue sensors such as arrays of position sensitive detectors with the final goal of integrating them in 3D scanning cameras or microscopes for object detection purposes. The work performed in this thesis deals with the implementation of prototype systems in order to explore the application of object detection using amorphous silicon position sensors of 32 and 128 lines which were produced in the clean room at CENIMAT-CEMOP. During the first phase of this work, the fabrication and the study of the static and dynamic specifications of the sensors as well as their conditioning in relation to the existing scientific and technological knowledge became a starting point. Subsequently, relevant data acquisition and suitable signal processing electronics were assembled. Various prototypes were developed for the 32 and 128 array PSD sensors. Appropriate optical solutions were integrated to work together with the constructed prototypes, allowing the required experiments to be carried out and allowing the achievement of the results presented in this thesis. All control, data acquisition and 3D rendering platform software was implemented for the existing systems. All these components were combined together to form several integrated systems for the 32 and 128 line PSD 3D sensors. The performance of the 32 PSD array sensor and system was evaluated for machine vision applications such as for example 3D object rendering as well as for microscopy applications such as for example micro object movement detection. Trials were also performed involving the 128 array PSD sensor systems. Sensor channel non-linearities of approximately 4 to 7% were obtained. Overall results obtained show the possibility of using a linear array of 32/128 1D line sensors based on the amorphous silicon technology to render 3D profiles of objects. The system and setup presented allows 3D rendering at high speeds and at high frame rates. The minimum detail or gap that can be detected by the sensor system is approximately 350 μm when using this current setup. It is also possible to render an object in 3D within a scanning angle range of 15º to 85º and identify its real height as a function of the scanning angle and the image displacement distance on the sensor. Simple and not so simple objects, such as a rubber and a plastic fork, can be rendered in 3D properly and accurately also at high resolution, using this sensor and system platform. The nip structure sensor system can detect primary and even derived colors of objects by a proper adjustment of the integration time of the system and by combining white, red, green and blue (RGB) light sources. A mean colorimetric error of 25.7 was obtained. It is also possible to detect the movement of micrometer objects using the 32 PSD sensor system. This kind of setup offers the possibility to detect if a micro object is moving, what are its dimensions and what is its position in two dimensions, even at high speeds. Results show a non-linearity of about 3% and a spatial resolution of < 2µm.
Resumo:
Dissertation to obtain the degree of Master in Chemical and Biochemical Engineering
Resumo:
This research aims to provide a better understanding on how firms stimulate knowledge sharing through the utilization of collaboration tools, in particular Emergent Social Software Platforms (ESSPs). It focuses on the distinctive applications of ESSPs and on the initiatives contributing to maximize its advantages. In the first part of the research, I have itemized all types of existing collaboration tools and classify them in different categories according to their capabilities, objectives and according to their faculty for promoting knowledge sharing. In the second part, and based on an exploratory case study at Cisco Systems, I have identified the main applications of an existing enterprise social software platform named Webex Social. By combining a qualitative and quantitative approach, as well as combining data collected from survey’s results and from the analysis of the company’s documents, I am expecting to maximize the outcome of this investigation and reduce the risk of bias. Although effects cannot be universalized based on one single case study, some utilization patterns have been underlined from the data collected and potential trends in managing knowledge have been observed. The results of the research have also enabled identifying most of the constraints experienced by the users of the firm’s social software platform. Utterly, this research should provide a primary framework for firms planning to create or implement a social software platform and for firms willing to increase adoption levels and to promote the overall participation of users. It highlights the common traps that should be avoided by developers when designing a social software platform and the capabilities that it should inherently carry to support an effective knowledge management strategy.
Resumo:
Spin-lattice Relaxation, self-Diffusion coefficients and Residual Dipolar Couplings (RDC’s) are the basis of well established Nuclear Magnetic Resonance techniques for the physicochemical study of small molecules (typically organic compounds and natural products with MW < 1000 Da), as they proved to be a powerful and complementary source of information about structural dynamic processes in solution. The work developed in this thesis consists in the application of the earlier-mentioned NMR techniques to explore, analyze and systematize patterns of the molecular dynamic behavior of selected small molecules in particular experimental conditions. Two systems were chosen to investigate molecular dynamic behavior by these techniques: the dynamics of ion-pair formation and ion interaction in ionic liquids (IL) and the dynamics of molecular reorientation when molecules are placed in oriented phases (alignment media). The application of NMR spin-lattice relaxation and self-diffusion measurements was applied to study the rotational and translational molecular dynamics of the IL: 1-butyl-3-methylimidazolium tetrafluoroborate [BMIM][BF4]. The study of the cation-anion dynamics in neat and IL-water mixtures was systematically investigated by a combination of multinuclear NMR relaxation techniques with diffusion data (using by H1, C13 and F19 NMR spectroscopy). Spin-lattice relaxation time (T1), self-diffusion coefficients and nuclear Overhauser effect experiments were combined to determine the conditions that favor the formation of long lived [BMIM][BF4] ion-pairs in water. For this purpose and using the self-diffusion coefficients of cation and anion as a probe, different IL-water compositions were screened (from neat IL to infinite dilution) to find the conditions where both cation and anion present equal diffusion coefficients (8% water fraction at 25 ºC). This condition as well as the neat IL and the infinite dilution were then further studied by 13C NMR relaxation in order to determine correlation times (c) for the molecular reorientational motion using a mathematical iterative procedure and experimental data obtained in a temperature range between 273 and 353 K. The behavior of self-diffusion and relaxation data obtained in our experiments point at the combining parameters of molar fraction 8 % and temperature 298 K as the most favorable condition for the formation of long lived ion-pairs. When molecules are subjected to soft anisotropic motion by being placed in some special media, Residual Dipolar Couplings (RDCs), can be measured, because of the partial alignment induced by this media. RDCs are emerging as a powerful routine tool employed in conformational analysis, as it complements and even outperforms the approaches based on the classical NMR NOE or J3 couplings. In this work, three different alignment media have been characterized and evaluated in terms of integrity using 2H and 1H 1D-NMR spectroscopy, namely the stretched and compressed gel PMMA, and the lyotropic liquid crystals CpCl/n-hexanol/brine and cromolyn/water. The influence that different media and degrees of alignment have on the dynamic properties of several molecules was explored. Different sized sugars were used and their self-diffusion was determined as well as conformation features using RDCs. The results obtained indicate that no influence is felt by the small molecules diffusion and conformational features studied within the alignment degree range studied, which was the 3, 5 and 6 % CpCl/n-hexanol/brine for diffusion, and 5 and 7.5 % CpCl/n-hexanol/brine for conformation. It was also possible to determine that the small molecules diffusion verified in the alignment media presented close values to the ones observed in water, reinforcing the idea of no conditioning of molecular properties in such media.
Resumo:
O peixe-zebra é utilizado como modelo vertebrado para estudos in vivo de diversas patologias de origem genética. Neste trabalho pretendeu-se estudar a discinesia ciliar primária através do estudo in vivo do organizador da esquerda-direita destes peixes, conhecido por vesícula de Kupffer. Em particular, após análise de sequências de imagens captadas a alta velocidade, avaliou-se o com-portamento dinâmico de cílios normais e comparou-se com o de cílios com alte-rações genéticas com o objectivo de melhor compreender os processos que in-fluenciam a localização assimétrica dos órgãos internos, associados a esta doen-ça. De vídeos obtidos por microscopia de alta velocidade do interior da vesí-cula de Kupffer foram analisados 32 cílios, sendo 8 normais e os restantes cílios pertencentes a três alterações genéticas diferentes: subexpressão de Pkd2, so-brexpressão de Arl13b e mutação no gene deltaD. Para cada cílio calculou-se a frequência de batimento e caracterizou-se quantitativamente o movimento. Esta última análise teve como base a segmentação manual do cílio em quatro pontos definindo a base, o meio e a ponta. De seguida, estudou-se a dinâmica de cada uma das rectas constituídas por estas três estruturas ao longo do tempo. Com recurso à análise estatística ANOVA podemos comprovar diferenças no movi-mento entre os cílios alterados quando comparados com os normais. A análise da frequência demonstrou que todos os cílios estudados possu-em uma frequência média 34,9 Hz. Demonstrou-se ainda que, quando compa-rados com cílios normais, os cílios Pkd2 possuem um movimento 60% mais rí-gido, os Alr13b são caracterizados por amplitudes angulares 26% superiores no seu movimento e, por fim, a relação entre as amplitudes base/meio é 23% supe-rior nos cílios Delta D. Para implementar estes estudos, desenvolveu-se uma ferramenta baseada num plugin para ImageJ conjuntamente com códigos em R, que poderá vir a ser utilizada em investigações da discinesia ciliar primária em peixes-zebra.
Resumo:
Os incêndios em edifícios representam um fenómeno que pode ter consequências devastadoras quando não controlado, não só em termos de perdas de vidas humanas, como em termos económicos. No passado, a ocorrência de incêndios de grandes dimensões mostrou os efeitos do fogo descontrolado nos edifícios, assim como a ineficiência dos meios de segurança ativa ao fogo. Nas últimas duas décadas, estas questões motivaram o estudo e compreensão da ação dos incêndios nas estruturas dos edifícios. Neste trabalho estuda-se a modelação da ação do fogo em estruturas metálicas e mistas, com o objetivo de contribuir para a sua melhor caracterização. A presente tese foca-se na validação e compreensão da implementação de análises termo-mecânicas a estruturas mistas no software de elementos finitos OpenSees (Open System for Earthquake Engineering Simulation), contribuindo assim para futuros estudos, não só de análises de estruturas mistas sujeitas a incêndio, mas também de análises de estruturas mistas sujeitas a eventos consecutivos, como sismo seguido de incêndio. Neste trabalho é feita uma breve descrição do fenómeno fogo, assim como dos processos inerentes à dinâmica de um incêndio que constituem uma fonte de incerteza para a modelação de cenários de incêndio num edifício. Posto isto, são abordados alguns modelos de incêndios presentes nos Eurocódigos, assim como o recente modelo de fogos móveis(“Travelling fires”). São realizados exemplos de aplicação no software e dois casos de estudo. O primeiro consiste na modelação de dois ensaios ao fogo realizados na Alemanha em 1986 em estruturas metálicas à escala 1/4. O segundo consiste na modelação de um ensaio ao fogo a uma viga de betão armado simplesmente apoiada, realizado no Instituto Superior Técnico em 2010. Os modelos numéricos desenvolvidos no OpenSees contabilizam as não linearidades físicas e geométricas, com elementos finitos de plasticidade distribuída e com uma formulação baseada em deslocamentos. Os resultados numéricos são então comparados com os experimentais, de modo a validar as análises termo-mecânicas no OpenSees.
Resumo:
The Corporate world is becoming more and more competitive. This leads organisations to adapt to this reality, by adopting more efficient processes, which result in a decrease in cost as well as an increase of product quality. One of these processes consists in making proposals to clients, which necessarily include a cost estimation of the project. This estimation is the main focus of this project. In particular, one of the goals is to evaluate which estimation models fit the Altran Portugal software factory the most, the organization where the fieldwork of this thesis will be carried out. There is no broad agreement about which is the type of estimation model more suitable to be used in software projects. Concerning contexts where there is plenty of objective information available to be used as input to an estimation model, model-based methods usually yield better results than the expert judgment. However, what happens more frequently is not having this volume and quality of information, which has a negative impact in the model-based methods performance, favouring the usage of expert judgement. In practice, most organisations use expert judgment, making themselves dependent on the expert. A common problem found is that the performance of the expert’s estimation depends on his previous experience with identical projects. This means that when new types of projects arrive, the estimation will have an unpredictable accuracy. Moreover, different experts will make different estimates, based on their individual experience. As a result, the company will not directly attain a continuous growing knowledge about how the estimate should be carried. Estimation models depend on the input information collected from previous projects, the size of the project database and the resources available. Altran currently does not store the input information from previous projects in a systematic way. It has a small project database and a team of experts. Our work is targeted to companies that operate in similar contexts. We start by gathering information from the organisation in order to identify which estimation approaches can be applied considering the organization’s context. A gap analysis is used to understand what type of information the company would have to collect so that other approaches would become available. Based on our assessment, in our opinion, expert judgment is the most adequate approach for Altran Portugal, in the current context. We analysed past development and evolution projects from Altran Portugal and assessed their estimates. This resulted in the identification of common estimation deviations, errors, and patterns, which lead to the proposal of metrics to help estimators produce estimates leveraging past projects quantitative and qualitative information in a convenient way. This dissertation aims to contribute to more realistic estimates, by identifying shortcomings in the current estimation process and supporting the self-improvement of the process, by gathering as much relevant information as possible from each finished project.
Resumo:
A discinesia ciliar primária (DCP) resulta de disfunção ciliar no ser humano, estando associada a um conjunto de sintomas muito diversificados. É uma doença respiratória rara caracterizada por infecções respiratórias, situs inversus, infertilidade e hidrocefalia. Em Portugal não existe nenhum centro de diagnóstico da doença. Mas a inten-ção de criar um surgiu, seguindo o método de centros de diagnóstico para DCP utilizado noutros países. Este diagnóstico consiste em recolher amostras dos cílios do nariz, através do método de escovagem nasal e obter a gravação do batimento das células ciliadas por uma câmara de alta velocidade acoplada a um microscópio com objectivas de alta resolução. É possível estudar a DCP através da análise do comportamento físico dos cílios, e, para uma melhor abordagem, foi desenvolvido um programa executável, em C#, para análise destas amostras. Este, após a escolha de uma zona de interesse da sequência de imagens pelo utilizador (ROI), detecta as frequências do bati-mento ciliar, indicando uma lista com as percentagens das frequências obtidas e cria um mapa de frequências do ROI. A ferramenta permite ainda calcular o comprimento do cílio e realizar um estudo do movimento do mesmo, algo que ainda não foi abordado por outros programas. O código desenvolvido permitirá, assim, obter um diagnóstico de DCP em Por-tugal, rápido e nalguns casos com um melhor desempenho do que a inspecção visual seguida noutros centros de diagnóstico.
Resumo:
RESUMO: O cancro de mama e o mais frequente diagnoticado a indiv duos do sexo feminino. O conhecimento cientifico e a tecnologia tem permitido a cria ção de muitas e diferentes estrat egias para tratar esta patologia. A Radioterapia (RT) est a entre as diretrizes atuais para a maioria dos tratamentos de cancro de mama. No entanto, a radia ção e como uma arma de dois canos: apesar de tratar, pode ser indutora de neoplasias secund arias. A mama contralateral (CLB) e um orgão susceptivel de absorver doses com o tratamento da outra mama, potenciando o risco de desenvolver um tumor secund ario. Nos departamentos de radioterapia tem sido implementadas novas tecnicas relacionadas com a radia ção, com complexas estrat egias de administra ção da dose e resultados promissores. No entanto, algumas questões precisam de ser devidamente colocadas, tais como: E seguro avançar para tecnicas complexas para obter melhores indices de conformidade nos volumes alvo, em radioterapia de mama? O que acontece aos volumes alvo e aos tecidos saudaveis adjacentes? Quão exata e a administração de dose? Quais são as limitações e vantagens das técnicas e algoritmos atualmente usados? A resposta a estas questões e conseguida recorrendo a m etodos de Monte Carlo para modelar com precisão os diferentes componentes do equipamento produtor de radia ção(alvos, ltros, colimadores, etc), a m de obter uma descri cão apropriada dos campos de radia cão usados, bem como uma representa ção geometrica detalhada e a composição dos materiais que constituem os orgãos e os tecidos envolvidos. Este trabalho visa investigar o impacto de tratar cancro de mama esquerda usando diferentes tecnicas de radioterapia f-IMRT (intensidade modulada por planeamento direto), IMRT por planeamento inverso (IMRT2, usando 2 feixes; IMRT5, com 5 feixes) e DCART (arco conformacional dinamico) e os seus impactos em irradia ção da mama e na irradia ção indesejada dos tecidos saud aveis adjacentes. Dois algoritmos do sistema de planeamento iPlan da BrainLAB foram usados: Pencil Beam Convolution (PBC) e Monte Carlo comercial iMC. Foi ainda usado um modelo de Monte Carlo criado para o acelerador usado (Trilogy da VARIAN Medical Systems), no c odigo EGSnrc MC, para determinar as doses depositadas na mama contralateral. Para atingir este objetivo foi necess ario modelar o novo colimador multi-laminas High- De nition que nunca antes havia sido simulado. O modelo desenvolvido est a agora disponí vel no pacote do c odigo EGSnrc MC do National Research Council Canada (NRC). O acelerador simulado foi validado com medidas realizadas em agua e posteriormente com c alculos realizados no sistema de planeamento (TPS).As distribui ções de dose no volume alvo (PTV) e a dose nos orgãos de risco (OAR) foram comparadas atrav es da an alise de histogramas de dose-volume; an alise estati stica complementar foi realizadas usando o software IBM SPSS v20. Para o algoritmo PBC, todas as tecnicas proporcionaram uma cobertura adequada do PTV. No entanto, foram encontradas diferen cas estatisticamente significativas entre as t ecnicas, no PTV, nos OAR e ainda no padrão da distribui ção de dose pelos tecidos sãos. IMRT5 e DCART contribuem para maior dispersão de doses baixas pelos tecidos normais, mama direita, pulmão direito, cora cão e at e pelo pulmão esquerdo, quando comparados com as tecnicas tangenciais (f-IMRT e IMRT2). No entanto, os planos de IMRT5 melhoram a distribuição de dose no PTV apresentando melhor conformidade e homogeneidade no volume alvo e percentagens de dose mais baixas nos orgãos do mesmo lado. A t ecnica de DCART não apresenta vantagens comparativamente com as restantes t ecnicas investigadas. Foram tamb em identi cadas diferen cas entre os algoritmos de c alculos: em geral, o PBC estimou doses mais elevadas para o PTV, pulmão esquerdo e cora ção, do que os algoritmos de MC. Os algoritmos de MC, entre si, apresentaram resultados semelhantes (com dferen cas at e 2%). Considera-se que o PBC não e preciso na determina ção de dose em meios homog eneos e na região de build-up. Nesse sentido, atualmente na cl nica, a equipa da F sica realiza medi ções para adquirir dados para outro algoritmo de c alculo. Apesar de melhor homogeneidade e conformidade no PTV considera-se que h a um aumento de risco de cancro na mama contralateral quando se utilizam t ecnicas não-tangenciais. Os resultados globais dos estudos apresentados confirmam o excelente poder de previsão com precisão na determinação e c alculo das distribui ções de dose nos orgãos e tecidos das tecnicas de simulação de Monte Carlo usados.---------ABSTRACT:Breast cancer is the most frequent in women. Scienti c knowledge and technology have created many and di erent strategies to treat this pathology. Radiotherapy (RT) is in the actual standard guidelines for most of breast cancer treatments. However, radiation is a two-sword weapon: although it may heal cancer, it may also induce secondary cancer. The contralateral breast (CLB) is a susceptible organ to absorb doses with the treatment of the other breast, being at signi cant risk to develop a secondary tumor. New radiation related techniques, with more complex delivery strategies and promising results are being implemented and used in radiotherapy departments. However some questions have to be properly addressed, such as: Is it safe to move to complex techniques to achieve better conformation in the target volumes, in breast radiotherapy? What happens to the target volumes and surrounding healthy tissues? How accurate is dose delivery? What are the shortcomings and limitations of currently used treatment planning systems (TPS)? The answers to these questions largely rely in the use of Monte Carlo (MC) simulations using state-of-the-art computer programs to accurately model the di erent components of the equipment (target, lters, collimators, etc.) and obtain an adequate description of the radiation elds used, as well as the detailed geometric representation and material composition of organs and tissues. This work aims at investigating the impact of treating left breast cancer using di erent radiation therapy (RT) techniques f-IMRT (forwardly-planned intensity-modulated), inversely-planned IMRT (IMRT2, using 2 beams; IMRT5, using 5 beams) and dynamic conformal arc (DCART) RT and their e ects on the whole-breast irradiation and in the undesirable irradiation of the surrounding healthy tissues. Two algorithms of iPlan BrainLAB TPS were used: Pencil Beam Convolution (PBC)and commercial Monte Carlo (iMC). Furthermore, an accurate Monte Carlo (MC) model of the linear accelerator used (a Trilogy R VARIANR) was done with the EGSnrc MC code, to accurately determine the doses that reach the CLB. For this purpose it was necessary to model the new High De nition multileaf collimator that had never before been simulated. The model developed was then included on the EGSnrc MC package of National Research Council Canada (NRC). The linac was benchmarked with water measurements and later on validated against the TPS calculations. The dose distributions in the planning target volume (PTV) and the dose to the organs at risk (OAR) were compared analyzing dose-volume histograms; further statistical analysis was performed using IBM SPSS v20 software. For PBC, all the techniques provided adequate coverage of the PTV. However, statistically significant dose di erences were observed between the techniques, in the PTV, OAR and also in the pattern of dose distribution spreading into normal tissues. IMRT5 and DCART spread low doses into greater volumes of normal tissue, right breast, right lung, heart and even the left lung than tangential techniques (f-IMRT and IMRT2). However,IMRT5 plans improved distributions for the PTV, exhibiting better conformity and homogeneity in target and reduced high dose percentages in ipsilateral OAR. DCART did not present advantages over any of the techniques investigated. Di erences were also found comparing the calculation algorithms: PBC estimated higher doses for the PTV, ipsilateral lung and heart than the MC algorithms predicted. The MC algorithms presented similar results (within 2% di erences). The PBC algorithm was considered not accurate in determining the dose in heterogeneous media and in build-up regions. Therefore, a major e ort is being done at the clinic to acquire data to move from PBC to another calculation algorithm. Despite better PTV homogeneity and conformity there is an increased risk of CLB cancer development, when using non-tangential techniques. The overall results of the studies performed con rm the outstanding predictive power and accuracy in the assessment and calculation of dose distributions in organs and tissues rendered possible by the utilization and implementation of MC simulation techniques in RT TPS.
Resumo:
Nesta Tese foi desenvolvida uma plataforma online multiutilizador, que tem como objetivo principal comparar algoritmos de análise de imagens para de-terminar o seu grau de eficácia. Um exemplo de aplicação é a comparação de algoritmos de análise de imagens da retina para deteção de drusas. A compa-ração é feita considerando um dos algoritmos como referência padrão e sobre o este são avaliados os restantes. O funcionamento da plataforma é semelhante à de um fórum, onde é possível a um utilizador criar tópicos publicando imagens e seu descritivo. Após a cria-ção do tópico qualquer utilizador pode visualizar o mesmo, dando a hipótese de comentar ou de acrescentar imagens processadas com os seus próprios al-goritmos de análise. Com o aumento de imagens processadas, obtém-se uma base de dados de algoritmos de análise de imagens sobre a qual é possível avaliar o seu grau de eficácia. A plataforma pretende também criar comunidades onde os utilizadores pos-sam interagir uns com os outros comentando sobre os tópicos, contribuindo assim para o melhoramento dos algoritmos. Deste modo, além de uma base de dados que qualquer utilizador pode usar, obtém-se uma fonte de informação disponibilizada por outros profissionais da área.
Resumo:
Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disease characterized by motor neurons degeneration, which reduces muscular force, being very difficult to diagnose. Mathematical methods are used in order to analyze the surface electromiographic signal’s dynamic behavior (Fractal Dimension (FD) and Multiscale Entropy (MSE)), evaluate different muscle group’s synchronization (Coherence and Phase Locking Factor (PLF)) and to evaluate the signal’s complexity (Lempel-Ziv (LZ) techniques and Detrended Fluctuation Analysis (DFA)). Surface electromiographic signal acquisitions were performed in upper limb muscles, being the analysis executed for instants of contraction for ipsilateral acquisitions for patients and control groups. Results from LZ, DFA and MSE analysis present capability to distinguish between the patient group and the control group, whereas coherence, PLF and FD algorithms present results very similar for both groups. LZ, DFA and MSE algorithms appear then to be a good measure of corticospinal pathways integrity. A classification algorithm was applied to the results in combination with extracted features from the surface electromiographic signal, with an accuracy percentage higher than 70% for 118 combinations for at least one classifier. The classification results demonstrate capability to distinguish members between patients and control groups. These results can demonstrate a major importance in the disease diagnose, once surface electromyography (sEMG) may be used as an auxiliary diagnose method.
Resumo:
Most of today’s systems, especially when related to the Web or to multi-agent systems, are not standalone or independent, but are part of a greater ecosystem, where they need to interact with other entities, react to complex changes in the environment, and act both over its own knowledge base and on the external environment itself. Moreover, these systems are clearly not static, but are constantly evolving due to the execution of self updates or external actions. Whenever actions and updates are possible, the need to ensure properties regarding the outcome of performing such actions emerges. Originally purposed in the context of databases, transactions solve this problem by guaranteeing atomicity, consistency, isolation and durability of a special set of actions. However, current transaction solutions fail to guarantee such properties in dynamic environments, since they cannot combine transaction execution with reactive features, or with the execution of actions over domains that the system does not completely control (thus making rolling back a non-viable proposition). In this thesis, we investigate what and how transaction properties can be ensured over these dynamic environments. To achieve this goal, we provide logic-based solutions, based on Transaction Logic, to precisely model and execute transactions in such environments, and where knowledge bases can be defined by arbitrary logic theories.
Resumo:
Contém CD com áudio como anexo