968 resultados para Machine à vecteurs de support
Resumo:
In this work the identification and diagnosis of various stages of chronic liver disease is addressed. The classification results of a support vector machine, a decision tree and a k-nearest neighbor classifier are compared. Ultrasound image intensity and textural features are jointly used with clinical and laboratorial data in the staging process. The classifiers training is performed by using a population of 97 patients at six different stages of chronic liver disease and a leave-one-out cross-validation strategy. The best results are obtained using the support vector machine with a radial-basis kernel, with 73.20% of overall accuracy. The good performance of the method is a promising indicator that it can be used, in a non invasive way, to provide reliable information about the chronic liver disease staging.
Resumo:
In this work liver contour is semi-automatically segmented and quantified in order to help the identification and diagnosis of diffuse liver disease. The features extracted from the liver contour are jointly used with clinical and laboratorial data in the staging process. The classification results of a support vector machine, a Bayesian and a k-nearest neighbor classifier are compared. A population of 88 patients at five different stages of diffuse liver disease and a leave-one-out cross-validation strategy are used in the classification process. The best results are obtained using the k-nearest neighbor classifier, with an overall accuracy of 80.68%. The good performance of the proposed method shows a reliable indicator that can improve the information in the staging of diffuse liver disease.
Resumo:
Introduction: A major focus of data mining process - especially machine learning researches - is to automatically learn to recognize complex patterns and help to take the adequate decisions strictly based on the acquired data. Since imaging techniques like MPI – Myocardial Perfusion Imaging on Nuclear Cardiology, can implicate a huge part of the daily workflow and generate gigabytes of data, there could be advantages on Computerized Analysis of data over Human Analysis: shorter time, homogeneity and consistency, automatic recording of analysis results, relatively inexpensive, etc.Objectives: The aim of this study relates with the evaluation of the efficacy of this methodology on the evaluation of MPI Stress studies and the process of decision taking concerning the continuation – or not – of the evaluation of each patient. It has been pursued has an objective to automatically classify a patient test in one of three groups: “Positive”, “Negative” and “Indeterminate”. “Positive” would directly follow to the Rest test part of the exam, the “Negative” would be directly exempted from continuation and only the “Indeterminate” group would deserve the clinician analysis, so allowing economy of clinician’s effort, increasing workflow fluidity at the technologist’s level and probably sparing time to patients. Methods: WEKA v3.6.2 open source software was used to make a comparative analysis of three WEKA algorithms (“OneR”, “J48” and “Naïve Bayes”) - on a retrospective study using the comparison with correspondent clinical results as reference, signed by nuclear cardiologist experts - on “SPECT Heart Dataset”, available on University of California – Irvine, at the Machine Learning Repository. For evaluation purposes, criteria as “Precision”, “Incorrectly Classified Instances” and “Receiver Operating Characteristics (ROC) Areas” were considered. Results: The interpretation of the data suggests that the Naïve Bayes algorithm has the best performance among the three previously selected algorithms. Conclusions: It is believed - and apparently supported by the findings - that machine learning algorithms could significantly assist, at an intermediary level, on the analysis of scintigraphic data obtained on MPI, namely after Stress acquisition, so eventually increasing efficiency of the entire system and potentially easing both roles of Technologists and Nuclear Cardiologists. In the actual continuation of this study, it is planned to use more patient information and significantly increase the population under study, in order to allow improving system accuracy.
Resumo:
Na Fábrica de Papel da Ponte Redonda fabricam-se sacos de papel multi-folhas e papel reciclado do tipo Kraft. Tendo em consideração a primeira actividade, é de grande interesse optimizar o processo de fabrico de papel com vista a incorporara a máxima taxa de papel produzido internamente nas diferentes camadas dos sacos de papel. Os papéis de maior interesse são os do tipo Fluting e Liners, tendo sido produzidos em 2010 um total de 4,9 mil toneladas, ou seja 90% de todo o papel fabricado em 2010, correspondendo a a 4 mil toneladas de papéis do tipo Liners e 0,9 mil toneladas para os papéis do tipo Fluting. Nos papéis do tipo Liners incluem-se os papéis do tipo Test-Liner e Kraft-Liner, representando em termos produtivos valores idênticos. No âmbito deste trabalho, em que se pretendeu controlar as águas do processo e optimizar a produção de papel, foram introduzidos uma unidade de flutuação e um sistema que permitisse regular a consistência da suspensão fibrosa à entrada da máquina do papel, e foram ainda estudadas as possibilidades de adição de produtos químicos para melhorar as características da pasta assim como um tratamento microbiológico mais eficaz para todo o processo. Para se avaliar se as medidas implementadas teriam um impacto positivo na qualidade desses dois tipos de papéis, desenvolveu-se o trabalho em duas fases: a primeira envolve a introdução de um sistema de flutuação e de um sistema de controlo de consistência da pasta, assim como a selecção de produtos químicos a adicionar ao processo. A segunda fase consistiu na avaliação do efeito destas medidas nas características do papel fabricado. Para o efeito foram escolhidos dois tipos de papel e de diferentes gramagens, nomeadamente Test-Liner de 80 g/m2 e Fluting de 110 g/m2. Introduziu-se um flutuador com o objectivo de tratar parte das águas do processo de fabrico com vista a serem reutilizadas em determinadas aplicações possíveis para a qualidade da água obtida (lavagens e água do processo), de modo a conseguir-se uma poupança de água, assim como aproveitar-se as lamas resultantes, ricas em fibra de celulose, para utilizá-las como matéria-prima. Foi introduzido um regulador de consistência no processo da Ponte Redonda com o objectivo de alimentar de uma forma constante a consistência da pasta à entrada da máquina do papel proporcionando uma melhor formação da folha, devido à ligação entre fibras, na direcção máquina e direcção transversal. Esse sistema inovador é um Regulador de Consistência que vem proporcionar à máquina do papel uma alimentação em fibra mais constante. O fabrico de papel apenas a partir de fibras de celulose não permitirá obter um papel com as características desejadas para a sua utilização. Para corrigir estas deficiências, são adicionados produtos químicos para atribuir ou melhorar as propriedades dos papéis. Desta forma considerou-se interessante introduzir no processo um agente de retenção numa fase posterior à preparação da pasta e antes da chegada à máquina de papel, de forma a melhorar as características da suspensão fibrosa. Assim foi implementado um sistema cuja eficácia foi avaliada. Concluiu-se que com a sua implementação a máquina de papel apresentou melhores resultados na drenagem e na turbidez da água removida, significando uma água com menor teor de matéria suspensa e dissolvida, devido à melhor agregação das fibras dispersas na suspensão fibrosa, levando a um aumento da drenagem e consequentemente melhor eficiência das prensas e secaria. Foi também elaborado um estudo para introdução de um sistema de tratamento microbiológico em todo o processo de fabrico de papel, devido à existência de microorganismos prejudiciais ao seu fabrico. Concluiu-se que a água clarificada proveniente do flutuador apresentou qualidade aceitável para os objectivos pretendidos. No entanto, considerando a eficiência de 26,5% na remoção de sólidos suspensos será necessário mais algum tempo de utilização da água clarificada, cerca de um ano, para avaliar se esta terá algum efeito prejudicial nos equipamentos. Verificou-se que devido à existência de microrganismos em todo o processo de fabrico de papel será necessário efectuar lavagens aos tinões, tanques e circuitos com alguma regularidade, aproveitando-se as paragens do processo assim como implementar um sistema de tratamento microbiológico mais eficaz. Em resultado das medidas implementadas concluiu-se que os papéis produzidos apresentaram melhorias, tendo-se obtido melhores resultados em todos os ensaios de resistência. No papel do tipo Test-Liner destacam-se os bons resultados nos ensaios de superfície, Cobb60 e rebentamento. No caso do parâmetro do Cobb60, foi um resultado surpreendente visto que por norma este tipo de papéis reciclados não suportam este ensaio. Concluiu-se também que as medidas implementadas proporcionaram uma melhor agregação e ligação entre fibras, e melhor formação da folha na máquina do papel proporcionando aos papéis propriedades físico-mecânicas mais interessantes.
Resumo:
CISTI'2015 - 10ª Conferência Ibérica de Sistemas e Tecnologias de Informação, 17 a 20 de junho de 2015, Águeda, Aveiro, Portugal.
Resumo:
Dissertação de Mestrado, Ciências Biomédicas, 3 de Fevereiro de 2016, Universidade dos Açores.
Resumo:
The purpose of this study is to analyse the interlimb relation and the influence of mechanical energy on metabolic energy expenditure during gait. In total, 22 subjects were monitored as to electromyographic activity, ground reaction forces and VO2 consumption (metabolic power) during gait. The results demonstrate a moderate negative correlation between the activity of tibialis anterior, biceps femoris and vastus medialis of the trailing limb during the transition between midstance and double support and that of the leading limb during double support for the same muscles, and between these and gastrocnemius medialis and soleus of the trailing limb during double support. Trailing limb soleus during the transition between mid-stance and double support was positively correlated to leading limb tibialis anterior, vastus medialis and biceps femoris during double support. Also, the trailing limb centre of mass mechanical work was strongly influenced by the leading limbs, although only the mechanical power related to forward progression of both limbs was correlated to metabolic power. These findings demonstrate a consistent interlimb relation in terms of electromyographic activity and centre of mass mechanical work, being the relations occurred in the plane of forward progression the more important to gait energy expenditure.
Resumo:
In almost all industrialized countries, the energy sector has suffered a severe restructuring that originated a greater complexity in market players’ interactions. The complexity that these changes brought made way for the creation of decision support tools that facilitate the study and understanding of these markets. MASCEM – “Multiagent Simulator for Competitive Electricity Markets” arose in this context providing a framework for evaluating new rules, new behaviour, and new participants in deregulated electricity markets. MASCEM uses game theory, machine learning techniques, scenario analysis and optimisation techniques to model market agents and to provide them with decision-support. ALBidS is a multiagent system created to provide decision support to market negotiating players. Fully integrated with MASCEM it considers several different methodologies based on very distinct approaches. The Six Thinking Hats is a powerful technique used to look at decisions from different perspectives. This tool’s goal is to force the thinker to move outside his habitual thinking style. It was developed to be used mainly at meetings in order to “run better meetings, make faster decisions”. This dissertation presents a study about the applicability of the Six Thinking Hats technique in Decision Support Systems, particularly with the multiagent paradigm like the MASCEM simulator. As such this work’s proposal is of a new agent, a meta-learner based on STH technique that organizes several different ALBidS’ strategies and combines the distinct answers into a single one that, expectedly, out-performs any of them.
Resumo:
Este trabalho é baseado no simulador de redes PST2200 do Laboratório de Sistemas de Energia (LSE) pois está avariado com vários problemas conhecidos, designadamente: - Defeito de isolamento (disparo de diferencial), - Desregulação da velocidade da máquina primária (motor DC), - Circuito de excitação da máquina síncrona inoperacional, - Inexistência de esquemas elétricos dos circuitos do simulador, - Medidas desreguladas e com canais de medida com circuito impresso queimado. O trabalho executado foi: - O levantamento e desenho de raiz (não existe qualquer manual) dos esquemas dos 10 módulos do simulador, designadamente naqueles com avaria ou com desempenho problemático a fim de que se possa ter uma visão mais pormenorizada dos circuitos e seus problemas, por forma a intervir para os minimizar e resolver, - Foi realizado o diagnóstico de avaria do simulador e foram propostas soluções para os mesmos, - Realizaram-se as intervenções propostas e aprovadas. Nas intervenções realizadas, os princípios orientadores foram: - Aumentar a robustez do equipamento por forma a garantir a sua integridade a utilizações menos apropriados e manobras 'exóticas' próprias de alunos, que pela sua condição, estão em fase de aprendizagem, - Atualizar o equipamento, colocando-o em sintonia com o 'estado da arte', - Como fator de valorização suplementar, foi concebida e aplicada a supervisão remota do funcionamento do simulador através da rede informática. Foram detetados inúmeros erros: - Má ligação do motor de corrente continua ao variador, resultando a falta de controlo da frequência da rede do sistema, - Ligações entre painéis trocadas resultando em avarias diversas das fontes de alimentação, - Cartas eletrónicas de medidas avariadas e que além de se reparar, foram também calibradas. Devido ao mecenato da empresa Schnitt + Sohn participando monetariamente, fez-se o projeto de alteração e respetiva execução de grande parte do simulador aumentando a fiabilidade do mesmo, diminuindo assim a frequência das avarias naturais mais as que acontecem involuntariamente devido a este ser um instrumento didático. Além do trabalho elétrico, foi feito muito trabalho de chaparia para alteração de estrutura e suporte do material com diferenças de posicionamento. Neste trabalho dá-se também alguns exemplos de cálculo e simulação das redes de transporte que se pode efetuar no simulador como estudo e simulação de avarias num sistema produtivo real. Realizou-se a monitorização de dois aparelhos indicadores de parâmetros de energia (Janitza UMG96S) através duma rede com dois protocolos ethernet e profibus utilizando o plc (Omron CJ2M) como valorização do trabalho.
Resumo:
OBJECTIVES: Estimate the frequency of online searches on the topic of smoking and analyze the quality of online resources available to smokers interested in giving up smoking. METHODS: Search engines were used to revise searches and online resources related to stopping smoking in Brazil in 2010. The number of searches was determined using analytical tools available on Google Ads; the number and type of sites were determined by replicating the search patterns of internet users. The sites were classified according to content (advertising, library of articles and other). The quality of the sites was analyzed using the Smoking Treatment Scale- Content (STS-C) and the Smoking Treatment Scale - Rating (STS-R). RESULTS: A total of 642,446 searches was carried out. Around a third of the 113 sites encountered were of the 'library' type, i.e. they only contained articles, followed by sites containing clinical advertising (18.6) and professional education (10.6). Thirteen of the sites offered advice on quitting directed at smokers. The majority of the sites did not contain evidence-based information, were not interactive and did not have the possibility of communicating with users after the first contact. Other limitations we came across were a lack of financial disclosure as well as no guarantee of privacy concerning information obtained and no distinction made between editorial content and advertisements. CONCLUSIONS: There is a disparity between the high demand for online support in giving up smoking and the scarcity of quality online resources for smokers. It is necessary to develop interactive, customized online resources based on evidence and random clinical testing in order to improve the support available to Brazilian smokers.
Resumo:
In this paper we address the ability of WorldFIP to cope with the real-time requirements of distributed computer-controlled systems (DCCS). Typical DCCS include process variables that must be transferred between network devices both in a periodic and sporadic (aperiodic) basis. The WorldFIP protocol is designed to support both types of traffic. WorldFIP can easily guarantee the timing requirements for the periodic traffic. However, for the aperiodic traffic more complex analysis must be made in order to guarantee its timing requirements. This paper describes work that is being carried out to extend previous relevant work, in order to include the actual schedule for the periodic traffic in the worst-case response time analysis of sporadic traffic in WorldFIP networks
Resumo:
This paper presents an architecture (Multi-μ) being implemented to study and develop software based fault tolerant mechanisms for Real-Time Systems, using the Ada language (Ada 95) and Commercial Off-The-Shelf (COTS) components. Several issues regarding fault tolerance are presented and mechanisms to achieve fault tolerance by software active replication in Ada 95 are discussed. The Multi-μ architecture, based on a specifically proposed Fault Tolerance Manager (FTManager), is then described. Finally, some considerations are made about the work being done and essential future developments.
Resumo:
In this paper, we analyse the ability of P-NET [1] fieldbus to cope with the timing requirements of a Distributed Computer Control System (DCCS), where messages associated to discrete events should be made available within a maximum bound time. The main objective of this work is to analyse how the network access and queueing delays, imposed by P-NET’s virtual token Medium Access Control (MAC) mechanism, affect the realtime behaviour of the supported DCCS.
Resumo:
In the past few years Tabling has emerged as a powerful logic programming model. The integration of concurrent features into the implementation of Tabling systems is demanded by need to use recently developed tabling applications within distributed systems, where a process has to respond concurrently to several requests. The support for sharing of tables among the concurrent threads of a Tabling process is a desirable feature, to allow one of Tabling’s virtues, the re-use of computations by other threads and to allow efficient usage of available memory. However, the incremental completion of tables which are evaluated concurrently is not a trivial problem. In this dissertation we describe the integration of concurrency mechanisms, by the way of multi-threading, in a state of the art Tabling and Prolog system, XSB. We begin by reviewing the main concepts for a formal description of tabled computations, called SLG resolution and for the implementation of Tabling under the SLG-WAM, the abstract machine supported by XSB. We describe the different scheduling strategies provided by XSB and introduce some new properties of local scheduling, a scheduling strategy for SLG resolution. We proceed to describe our implementation work by describing the process of integrating multi-threading in a Prolog system supporting Tabling, without addressing the problem of shared tables. We describe the trade-offs and implementation decisions involved. We then describe an optimistic algorithm for the concurrent sharing of completed tables, Shared Completed Tables, which allows the sharing of tables without incurring in deadlocks, under local scheduling. This method relies on the execution properties of local scheduling and includes full support for negation. We provide a theoretical framework and discuss the implementation’s correctness and complexity. After that, we describe amethod for the sharing of tables among threads that allows parallelism in the computation of inter-dependent subgoals, which we name Concurrent Completion. We informally argue for the correctness of Concurrent Completion. We give detailed performance measurements of the multi-threaded XSB systems over a variety of machines and operating systems, for both the Shared Completed Tables and the Concurrent Completion implementations. We focus our measurements inthe overhead over the sequential engine and the scalability of the system. We finish with a comparison of XSB with other multi-threaded Prolog systems and we compare our approach to concurrent tabling with parallel and distributed methods for the evaluation of tabling. Finally, we identify future research directions.