1000 resultados para Algoritmo computacional alg-619
Resumo:
The Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disease characterized by progressive muscle weakness that leads the patient to death, usually due to respiratory complications. Thus, as the disease progresses the patient will require noninvasive ventilation (NIV) and constant monitoring. This paper presents a distributed architecture for homecare monitoring of nocturnal NIV in patients with ALS. The implementation of this architecture used single board computers and mobile devices placed in patient’s homes, to display alert messages for caregivers and a web server for remote monitoring by the healthcare staff. The architecture used a software based on fuzzy logic and computer vision to capture data from a mechanical ventilator screen and generate alert messages with instructions for caregivers. The monitoring was performed on 29 patients for 7 con-tinuous hours daily during 5 days generating a total of 126000 samples for each variable monitored at a sampling rate of one sample per second. The system was evaluated regarding the rate of hits for character recognition and its correction through an algorithm for the detection and correction of errors. Furthermore, a healthcare team evaluated regarding the time intervals at which the alert messages were generated and the correctness of such messages. Thus, the system showed an average hit rate of 98.72%, and in the worst case 98.39%. As for the message to be generated, the system also agreed 100% to the overall assessment, and there was disagreement in only 2 cases with one of the physician evaluators.
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
This work proposes a new autonomous navigation strategy assisted by genetic algorithm with dynamic planning for terrestrial mobile robots, called DPNA-GA (Dynamic Planning Navigation Algorithm optimized with Genetic Algorithm). The strategy was applied in environments - both static and dynamic - in which the location and shape of the obstacles is not known in advance. In each shift event, a control algorithm minimizes the distance between the robot and the object and maximizes the distance from the obstacles, rescheduling the route. Using a spatial location sensor and a set of distance sensors, the proposed navigation strategy is able to dynamically plan optimal collision-free paths. Simulations performed in different environments demonstrated that the technique provides a high degree of flexibility and robustness. For this, there were applied several variations of genetic parameters such as: crossing rate, population size, among others. Finally, the simulation results successfully demonstrate the effectiveness and robustness of DPNA-GA technique, validating it for real applications in terrestrial mobile robots.
Resumo:
This work proposes a new autonomous navigation strategy assisted by genetic algorithm with dynamic planning for terrestrial mobile robots, called DPNA-GA (Dynamic Planning Navigation Algorithm optimized with Genetic Algorithm). The strategy was applied in environments - both static and dynamic - in which the location and shape of the obstacles is not known in advance. In each shift event, a control algorithm minimizes the distance between the robot and the object and maximizes the distance from the obstacles, rescheduling the route. Using a spatial location sensor and a set of distance sensors, the proposed navigation strategy is able to dynamically plan optimal collision-free paths. Simulations performed in different environments demonstrated that the technique provides a high degree of flexibility and robustness. For this, there were applied several variations of genetic parameters such as: crossing rate, population size, among others. Finally, the simulation results successfully demonstrate the effectiveness and robustness of DPNA-GA technique, validating it for real applications in terrestrial mobile robots.
Resumo:
Nell'elaborato viene introdotto l'ambito della Computer Vision e come l'algoritmo SIFT si inserisce nel suo panorama. Viene inoltre descritto SIFT stesso, le varie fasi di cui si compone e un'applicazione al problema dell'object recognition. Infine viene presentata un'implementazione di SIFT in linguaggio Python creata per ottenere un'applicazione didattica interattiva e vengono mostrati esempi di questa applicazione.
Resumo:
La misurazione del gradiente pressorio venoso (HVPG) viene utilizzata per diagnosticare la gravità della malattia del fegato ma si tratta di una pratica invasiva e che non può essere effettuata in tutti i pazienti. Per questo motivo sono state studiate nuove metodiche per riuscire ad analizzare la cirrosi, una tra le quali l’indagine ecografica. Un progetto in fase di svolgimento (progetto CLEVER) è stato avviato per riuscire a sviluppare, validare e trasferire nella pratica clinica un nuovo sistema basato sull'analisi di immagini ecografiche con mezzo di contrasto al fine di valutare la gravità della degenerazione della rete vascolare causata dalla cirrosi. L'obiettivo principale della ricerca è quello di sviluppare uno strumento completamente automatico per l'analisi in tempo reale della rete vascolare del fegato in grado di valutare la connettività vascolare e quantificare il grado di squilibrio della rete.
Resumo:
Site 619, located in the Pigmy Basin off the coast of Louisiana, penetrated the late Quaternary Ericson Zones X, Y, and Z. The penetrated section can be divided into four intervals. The lower interval (below 157 m sub-bottom) comprises 51 m of displaced sediments which probably originated from the Louisiana continental shelf. The upper three intervals (above 157 m) are dominated by pelagic/hemipelagic sedimentation associated with a closed basin. These are divided on the basis of planktonic foraminifers into Zones X, Y, and Z. These warm-cool water intervals are identified mainly by using the Globorotalia menardii complex (warm) and G. inflata (cool). The intervals correlate with published curves taken from piston core samples in the western Gulf of Mexico.
Resumo:
The objective of this work is to use algorithms known as Boltzmann Machine to rebuild and classify patterns as images. This algorithm has a similar structure to that of an Artificial Neural Network but network nodes have stochastic and probabilistic decisions. This work presents the theoretical framework of the main Artificial Neural Networks, General Boltzmann Machine algorithm and a variation of this algorithm known as Restricted Boltzmann Machine. Computer simulations are performed comparing algorithms Artificial Neural Network Backpropagation with these algorithms Boltzmann General Machine and Machine Restricted Boltzmann. Through computer simulations are analyzed executions times of the different described algorithms and bit hit percentage of trained patterns that are later reconstructed. Finally, they used binary images with and without noise in training Restricted Boltzmann Machine algorithm, these images are reconstructed and classified according to the bit hit percentage in the reconstruction of the images. The Boltzmann machine algorithms were able to classify patterns trained and showed excellent results in the reconstruction of the standards code faster runtime and thus can be used in applications such as image recognition.
Greenow: um algoritmo de roteamento orientado a workspace para uma arquitetura de Internet do futuro
Resumo:
Current and future applications pose new requirements that Internet architecture is not able to satisfy, like Mobility, Multicast, Multihoming, Bandwidth Guarantee and so on. The Internet architecture has some limitations which do not allow all future requirements to be covered. New architectures were proposed considering these requirements when a communication is established. ETArch (Entity Title Architecture) is a new Internet architecture, clean slate, able to use application’s requirements on each communication, and flexible to work with several layers. The Routing has an important role on Internet, because it decides the best way to forward primitives through the network. In Future Internet, all requirements depend on the routing. Routing is responsible for deciding the best path and, in the future, a better route can consider Mobility aspects or Energy Consumption, for instance. In the dawn of ETArch, the Routing has not been defined. This work provides intra and inter-domain routing algorithms to be used in the ETArch. It is considered that the route should be defined completely before the data start to traffic, to ensure that the requirements are met. In the Internet, the Routing has two distinct functions: (i) run specific algorithms to define the best route; and (ii) to forward data primitives to the correct link. In traditional Internet architecture, the two Routing functions are performed in all routers everytime that a packet arrives. This work allows that the complete route is defined before the communication starts, like in the telecommunication systems. This work determined the Routing for ETArch and experiments were performed to demonstrate the control plane routing viability. The initial setup before a communication takes longer, then only forwarding of primitives is performed, saving processing time.
Resumo:
The search for mitigation solutions, with respect to the effects of overvoltages linked to the energization and reclosing maneuvers of transmission lines include a challenging subject with strong impact on the insulation coordination of electrical systems. Although the recognition of classical and commercial proposals to mitigate these phenomena, other possibilities are certainly still worthwhile highlighting and investigations. In this context, the present work is grounded in the exposure of the physical and mathematical foundations of a strategy based on controlled switchings, whose moments to the line reclosing are pre-programmed. The computational evaluation of the effectiveness of the methodology is made using the ATP simulator, which are studies in a typical electrical system subjected to the action of short-circuits fallowed by shutdowns and subsequent reclosing, under the action of technology here focused and lack thereof.
Resumo:
In this work are considered two bidimensional systems, with distints chacacteristcs and applicabilitys. Is studied the adsorption of transition metals (MT) Fe, Co, Mn and Ru in extended defects, formed by graphene grain boundaries. First in pristine graphene The hollow site of carbon hexagon, in pristine graphene, are the most stable for MT adsorption. The Dirac cone in eletronic structure of graphene was manteined with the presence of MT. For the considered grain boundaries the MT has a greater stability for absorption in the grain boundaries sites in comparison with pristine graphene. Through the energy barrier values, are observed diffusion chanels for MT localized on the grain boundaries. This diffusion chanels indicate a possible formation of nanolines of MT in graphene. For the first stage of the nanolines, ate observed a better stability for the system with greater MT concentration, due to MT-MT interactions. Also, due to the magnetic moment of the MT, the nanolines show a magnetization. For the most stable configurations the system are metallics, particularly for Fe the band structure indicates an anisotropic spin current. In a second study, are considereted the retention capacity for metallic contaminants Cd and Hg in clayminerals, kaolinite (KAO) and montmorillonite (MMT). Through the adsorption energies of contaminantes in the clayminerals, was observed a increase in stability with the increase of contaminants concentration, due to the interaction Cd-Cd and Hg-Hg. Also, was observed that KAO has a strong interaction beteween monolayers than MMT. In this sence, for the adsoption process of contaminantes in the natural form of KAO and MMT, the latter has a better retention capacity, due to the small net work for contaminant intercalation. However, when the modification of clayminerals, with molecules that increase the spacing between monolayers, exist a optimal condition, which the contaminant absorption are more stable in KAO system than in MMT. In the Langmuir adsorption model for the clayminerals in the optimal monolayer spacing, the retention capacity for Cd and Hg in KAO system are 21% greater than in MMT system. Also, for the X-ray Absorption Near Edge Spectroscopy (XANES) for the K edge of Cd and Hg, are found a positive shift of absorption edge with the decreasing of monolayer spacing. This result indicates a possible way to determine the concentration of adsorbed contaminats in relation to unabsorbed ones, from the decomposition of experimental XANES in the obteined spectras.
Resumo:
A lo largo de la historia, nuestro planeta ha atravesado numerosas y diferentes etapas. Sin embargo, desde finales del cretácico no se vivía un cambio tan rápido como el actual. Y a la cabeza del cambio, nosotros, el ser humano. De igual manera que somos la causa, debemos ser también la solución, y el análisis a gran escala de la tierra está siendo un punto de interés para la comunidad científica en los últimos años. Prueba de ello es que, cada vez con más frecuencia, se lanzan gran cantidad de satélites cuya finalidad es el análisis, mediante fotografías, de la superficie terrestre. Una de las técnicas más versátiles para este análisis es la toma de imágenes hiperespectrales, donde no solo se captura el espectro visible, sino numerosas longitudes de onda. Suponen, eso sí un reto tecnológico, pues los sensores consumen más energía y las imágenes más memoria, ambos recursos escasos en el espacio. Dado que el análisis se hace en tierra firme, es importante una transmisión de datos eficaz y rápida. Por ello creemos que la compresión en tiempo real mediante FPGAs es la solución idónea, combinando un bajo consumo con una alta tasa de compresión, posibilitando el análisis ininterrumpido del astro en el que vivimos. En este trabajo de fin de grado se ha realizado una implementación sobre FPGA, utilizando VHDL, del estándar CCSDS 123. Este está diseñado para la compresión sin pérdida de imágenes hiperespectrales, y permite una amplia gama de configuraciones para adaptarse de manera óptima a cualquier tipo de imagen. Se ha comprobado exitosamente la validez de la implementación comparando los resultados obtenidos con otras implementaciones (software) existentes. Las principales ventajas que presentamos aquí es que se posibilita la compresión en tiempo real, obteniendo además un rendimiento energético muy prometedor. Estos resultados mejoran notablemente los de una implementación software del algoritmo, y permitirán la compresión de las imágenes a bordo de los satélites que las toman.
Resumo:
Esta dissertação apresenta um trabalho sobre codificação de vídeo 3D compatível com vídeo 2D. Tem por base o desenvolvimento de um método para melhorar, no descodificador, a reconstrução de uma vista subamostrada resultante de uma transmissão simulcast usando a norma de codificação de vídeo H.265 (informalmente denominada de High Efficiency Video Coding (HEVC)). Apesar de manter a compatibilidade com vídeo 2D a transmissão simulcast normalmente requer uma taxa de transmissão elevada. Na ausência de ferramentas de codificação 3D adequadas é possível reduzir a taxa de transmissão utilizando compressão assimétrica do vídeo, onde a vista base é codificada com a resolução espacial original, enquanto que a vista auxiliar é codificada com uma resolução espacial menor, sendo sobreamostrada no descodificador. O método desenvolvido visa melhorar a vista auxiliar sobreamostrada no descodificador utilizando informação dos detalhes da vista base, ou seja, as componentes de alta frequência. Este processo depende de transformadas Afim para realizar um mapeamento geométrico entre a informação de alta frequência da vista base de resolução completa e a vista auxiliar de menor resolução. Adicionalmente, de modo a manter a continuidade do conteúdo da imagem entre regiões, evitando artefatos de blocos, o mapeamento utiliza uma malha de triangulação da vista auxiliar aplicado à imagem de detalhes obtida a partir da vista base. A técnica proposta é comparada com um método de estimação de disparidade por correspondência de blocos, sendo que os resultados mostram que para algumas sequências a técnica desenvolvida melhora não só a qualidade objetiva (PSNR) até 2.2 dB, mas também a qualidade subjetiva, para a mesma taxa de compressão global.
Resumo:
Hoje em dia, a área de codificação de dados é transversal a diversos tipos de engenharias devido à sua grande importância. Com o aumento exponencial na criação de dados digitais, o campo da compressão de dados ganhou uma grande visibilidade nesta área. São constantemente desenvolvidos e melhorados algoritmos de compressão por forma a obter a maior compressão de dados possível seja com ou sem perda de dados, permitindo sustentar o rápido e constante crescimento dos mesmos. Um dos grandes problemas deste tipo de algoritmos deve-se ao grande poder computacional que por vezes é necessário para obter uma boa taxa de compressão mantendo a qualidade dos dados quando descompactados. Este documento descreve uma estratégia para tentar reduzir o impacto do poder computacional necessário à codificação de imagens utilizando uma implementação heterogénea. O objetivo é tentar efetuar a paralelização das secções que requerem elevado poder computacional reduzindo assim o tempo necessário à compressão de dados. Este documento baseia-se na implementação desta estratégia para o algoritmo de codificação de imagens MMP-Intra. Utilizando inicialmente uma análise teórica, demonstramos que é viável efetuar a paralelização do algoritmo, sendo possível obter elevados ganhos de desempenho. Por forma a provar que o algoritmo MMP-Intra era paralelizavel e identificar os ganhos reais foi desenvolvido um protótipo inicial, o qual obteve um desempenho muito inferiore ao do algoritmo original, necessitando de muito mais tempo para obter os mesmo resultados. Utilizando um processo de otimização iterativo o protótipo passou por várias etapas de refinação. O protótipo refinado final obteve resultados muito superiores ao algoritmo sequencial no qual o mesmo foi baseado chegando a obter desempenhos quatro vezes superior ao original.
Resumo:
Preliminary data on dissolved organic carbon (DOC) and dissolved sugars in interstitial water samples collected at Sites 618, 619, and 623 of Deep Sea Drilling Project Leg 96 are presented. At Site 618 in Orca Basin, the DOC content of the interstitial water peaks in the hypersaline sulfate reduction zone. The sugar content reaches a maximum and the DOC content begins to decrease at the depth of methane gas generation. Below that depth, the sugar and DOC contents are about constant. At Site 619 in Pigmy Basin, the DOC content increases slightly with depth in the sulfate reduction and the methane fermentation zones. The sugar content is lower in the sulfate reduction zone than in the methane fermentation zone; sugar concentration increases and fluctuates with methane gas percentages within the methane fermentation zone. At Site 623 in the lower fan region of the Mississippi Fan, there is no sulfate reduction zone. The DOC and sugar contents of the interstitial water are almost constant with depth.