960 resultados para Similarity queries
Resumo:
Earthquakes are associated with negative events, such as large number of casualties, destruction of buildings and infrastructures, or emergence of tsunamis. In this paper, we apply the Multidimensional Scaling (MDS) analysis to earthquake data. MDS is a set of techniques that produce spatial or geometric representations of complex objects, such that, objects perceived to be similar/distinct in some sense are placed nearby/distant on the MDS maps. The interpretation of the charts is based on the resulting clusters since MDS produces a different locus for each similarity measure. In this study, over three million seismic occurrences, covering the period from January 1, 1904 up to March 14, 2012 are analyzed. The events, characterized by their magnitude and spatiotemporal distributions, are divided into groups, either according to the Flinn–Engdahl seismic regions of Earth or using a rectangular grid based in latitude and longitude coordinates. Space-time and Space-frequency correlation indices are proposed to quantify the similarities among events. MDS has the advantage of avoiding sensitivity to the non-uniform spatial distribution of seismic data, resulting from poorly instrumented areas, and is well suited for accessing dynamics of complex systems. MDS maps are proven as an intuitive and useful visual representation of the complex relationships that are present among seismic events, which may not be perceived on traditional geographic maps. Therefore, MDS constitutes a valid alternative to classic visualization tools, for understanding the global behavior of earthquakes.
Resumo:
The members of the subfamily Triatominae (Hemiptera : Reduviidae) comprise a great number of species of medical importance in the transmission of the T. cruzi (American trypanosomiasis). The aim of this study was to contribute to the knowledge about the chemical composition in proteins, lipids, lipoproteins, and carbohydrates of vectors of Chagas' disease corresponding to twelve members of the subfamily Triatominae. This study was carried out in ninphs of the fifth instar and adult males of the species: T. delpontei, T. dimidiata, T. guasayana, T. infestans, T. mazzotti, T. pallidipennis, T. patagonica, T. platensis, T. rubrovaria, T. sordida of the Triatoma genus, and D. maximus and P. megistus of the Dipatalogaster and Panstrongylus genera respectively. The results show on one hand, qualitative differences in the protein composition, and on the other hand, similarity in the lipoprotein profiles. Lipids, proteins, and carbohydrates did not show significant differences between species or/and stages.
Resumo:
A chromatographic separation of active ingredients of Combivir, Epivir, Kaletra, Norvir, Prezista, Retrovir, Trivizir, Valcyte, and Viramune is performed on thin layer chromatography. The spectra of these nine drugs were recorded using the Fourier transform infrared spectroscopy. This information is then analyzed by means of the cosine correlation. The comparison of the infrared spectra in the perspective of the adopted similarity measure is possible to visualize with present day computer tools, and the emerging clusters provide additional information about the similarities of the investigated set of complex drugs.
Resumo:
In the last decade, local image features have been widely used in robot visual localization. In order to assess image similarity, a strategy exploiting these features compares raw descriptors extracted from the current image with those in the models of places. This paper addresses the ensuing step in this process, where a combining function must be used to aggregate results and assign each place a score. Casting the problem in the multiple classifier systems framework, in this paper we compare several candidate combiners with respect to their performance in the visual localization task. For this evaluation, we selected the most popular methods in the class of non-trained combiners, namely the sum rule and product rule. A deeper insight into the potential of these combiners is provided through a discriminativity analysis involving the algebraic rules and two extensions of these methods: the threshold, as well as the weighted modifications. In addition, a voting method, previously used in robot visual localization, is assessed. Furthermore, we address the process of constructing a model of the environment by describing how the model granularity impacts upon performance. All combiners are tested on a visual localization task, carried out on a public dataset. It is experimentally demonstrated that the sum rule extensions globally achieve the best performance, confirming the general agreement on the robustness of this rule in other classification problems. The voting method, whilst competitive with the product rule in its standard form, is shown to be outperformed by its modified versions.
Resumo:
Mice transcutaneously infected with about 400 cercariae were submitted to treatment with oxamniquine (400 mg/kg), 24 hours after infection. The recovery of schistosomules, at 4, 24, 48 and 72 hours and 35 days after treatment, showed the activity of the drug on the parasites, thus practically preventing their migration from the skin to the lungs. Worm recovery performed in the lungs (96 hours after treatment) showed recovery means of 0.6 worms/mouse in the treated group and 53.8 in the control group (untreated). The perfusion of the portal system carried out at 35 days after treatment clearly showed the elimination of all the parasites in the treated group, whereas a recovery mean of 144.7 worms/mouse was detected in the control group (untreated). These findings confirm the efficacy of oxamniquine at the skin phase of infection, and also show similarity with the immunization method that uses irradiated cercariae. The practical application of these findings in the medical clinic is discussed too
Resumo:
The Evidence Accumulation Clustering (EAC) paradigm is a clustering ensemble method which derives a consensus partition from a collection of base clusterings obtained using different algorithms. It collects from the partitions in the ensemble a set of pairwise observations about the co-occurrence of objects in a same cluster and it uses these co-occurrence statistics to derive a similarity matrix, referred to as co-association matrix. The Probabilistic Evidence Accumulation for Clustering Ensembles (PEACE) algorithm is a principled approach for the extraction of a consensus clustering from the observations encoded in the co-association matrix based on a probabilistic model for the co-association matrix parameterized by the unknown assignments of objects to clusters. In this paper we extend the PEACE algorithm by deriving a consensus solution according to a MAP approach with Dirichlet priors defined for the unknown probabilistic cluster assignments. In particular, we study the positive regularization effect of Dirichlet priors on the final consensus solution with both synthetic and real benchmark data.
Resumo:
Arguably, the most difficult task in text classification is to choose an appropriate set of features that allows machine learning algorithms to provide accurate classification. Most state-of-the-art techniques for this task involve careful feature engineering and a pre-processing stage, which may be too expensive in the emerging context of massive collections of electronic texts. In this paper, we propose efficient methods for text classification based on information-theoretic dissimilarity measures, which are used to define dissimilarity-based representations. These methods dispense with any feature design or engineering, by mapping texts into a feature space using universal dissimilarity measures; in this space, classical classifiers (e.g. nearest neighbor or support vector machines) can then be used. The reported experimental evaluation of the proposed methods, on sentiment polarity analysis and authorship attribution problems, reveals that it approximates, sometimes even outperforms previous state-of-the-art techniques, despite being much simpler, in the sense that they do not require any text pre-processing or feature engineering.
Resumo:
Mestrado em Engenharia Mecânica- Energia
Resumo:
Na atualidade, existe uma quantidade de dados criados diariamente que ultrapassam em muito as mais otimistas espectativas estabelecidas na década anterior. Estes dados têm origens bastante diversas e apresentam-se sobre várias formas. Este novo conceito que dá pelo nome de Big Data está a colocar novos e rebuscados desafios ao seu armazenamento, tratamento e manipulação. Os tradicionais sistemas de armazenamento não se apresentam como a solução indicada para este problema. Estes desafios são alguns dos mais analisados e dissertados temas informáticos do momento. Várias tecnologias têm emergido com esta nova era, das quais se salienta um novo paradigma de armazenamento, o movimento NoSQL. Esta nova filosofia de armazenamento visa responder às necessidades de armazenamento e processamento destes volumosos e heterogéneos dados. Os armazéns de dados são um dos componentes mais importantes do âmbito Business Intelligence e são, maioritariamente, utilizados como uma ferramenta de apoio aos processos de tomada decisão, levados a cabo no dia-a-dia de uma organização. A sua componente histórica implica que grandes volumes de dados sejam armazenados, tratados e analisados tendo por base os seus repositórios. Algumas organizações começam a ter problemas para gerir e armazenar estes grandes volumes de informação. Esse facto deve-se, em grande parte, à estrutura de armazenamento que lhes serve de base. Os sistemas de gestão de bases de dados relacionais são, há algumas décadas, considerados como o método primordial de armazenamento de informação num armazém de dados. De facto, estes sistemas começam a não se mostrar capazes de armazenar e gerir os dados operacionais das organizações, sendo consequentemente cada vez menos recomendada a sua utilização em armazéns de dados. É intrinsecamente interessante o pensamento de que as bases de dados relacionais começam a perder a luta contra o volume de dados, numa altura em que um novo paradigma de armazenamento surge, exatamente com o intuito de dominar o grande volume inerente aos dados Big Data. Ainda é mais interessante o pensamento de que, possivelmente, estes novos sistemas NoSQL podem trazer vantagens para o mundo dos armazéns de dados. Assim, neste trabalho de mestrado, irá ser estudada a viabilidade e as implicações da adoção de bases de dados NoSQL, no contexto de armazéns de dados, em comparação com a abordagem tradicional, implementada sobre sistemas relacionais. Para alcançar esta tarefa, vários estudos foram operados tendo por base o sistema relacional SQL Server 2014 e os sistemas NoSQL, MongoDB e Cassandra. Várias etapas do processo de desenho e implementação de um armazém de dados foram comparadas entre os três sistemas, sendo que três armazéns de dados distintos foram criados tendo por base cada um dos sistemas. Toda a investigação realizada neste trabalho culmina no confronto da performance de consultas, realizadas nos três sistemas.
Resumo:
In the present study we report the results of an analysis, based on serotyping, multilocus enzyme electrophoresis (MEE), and ribotyping of N. meningitidis serogroup C strains isolated from patients with meningococcal disease (MD) in Rio Grande do Sul (RS) and Santa Catarina (SC) States, Brazil, as the Center of Epidemiology Control of Ministry of Health detected an increasing of MD cases due to this serogroup in the last two years (1992-1993). We have demonstrated that the MD due to N.meningitidis serogroup C strains in RS and SC States occurring in the last 4 years were caused mainly by one clone of strains (ET 40), with isolates indistinguishable by serogroup, serotype, subtype and even by ribotyping. One small number of cases that were not due to an ET 40 strains, represent closely related clones that probably are new lineages generated from the ET 40 clone referred as ET 11A complex. We have also analyzed N.meningitidis serogroup C strains isolated in the greater São Paulo in 1976 as representative of the first post epidemic year in that region. The ribotyping method, as well as MEE, could provide useful information about the clonal characteristics of those isolates and also of strains isolated in south Brazil. The strains from 1976 have more similarity with the actual endemic than epidemic strains, by the ribotyping, sulfonamide sensitivity, and MEE results. In conclusion, serotyping with monoclonal antibodies (C:2b:P1.3), MEE (ET 11 and ET 11A complex), and ribotyping by using ClaI restriction enzyme (Rb2), were useful to characterize these epidemic strains of N.meningitidis related to the increased incidence of MD in different States of south Brazil. It is mostly probable that these N.meningitidis serogroup C strains have poor or no genetic corelation with 1971-1975 epidemic serogroup C strains. The genetic similarity of members of the ET 11 and ET 11A complex were confirmed by the ribotyping method by using three restriction endonucleases.
Resumo:
We show here a simplified RT-PCR for identification of dengue virus types 1 and 2. Five dengue virus strains, isolated from Brazilian patients, and yellow fever vaccine 17DD as a negative control, were used in this study. C6/36 cells were infected and supernatants were collected after 7 days. The RT-PCR, done in a single reaction vessel, was carried out following a 1/10 dilution of virus in distilled water or in a detergent mixture containing Nonidet P40. The 50 µl assay reaction mixture included 50 pmol of specific primers amplifying a 482 base pair sequence for dengue type 1 and 210 base pair sequence for dengue type 2. In other assays, we used dengue virus consensus primers having maximum sequence similarity to the four serotypes, amplifying a 511 base pair sequence. The reaction mixture also contained 0.1 mM of the four deoxynucleoside triphosphates, 7.5 U of reverse transcriptase, 1U of thermostable Taq DNA polymerase. The mixture was incubated for 5 minutes at 37ºC for reverse transcription followed by 30 cycles of two-step PCR amplification (92ºC for 60 seconds, 53ºC for 60 seconds) with slow temperature increment. The PCR products were subjected to 1.7% agarose gel electrophoresis and visualized by UV light after staining with ethidium bromide solution. Low virus titer around 10 3, 6 TCID50/ml was detected by RT-PCR for dengue type 1. Specific DNA amplification was observed with all the Brazilian dengue strains by using dengue virus consensus primers. As compared to other RT-PCRs, this assay is less laborious, done in a shorter time, and has reduced risk of contamination
Resumo:
Thesis submitted in the fulfillment of the requirements for the Degree of Master in Biomedical Engineering
Resumo:
No dia-a-dia existe regularmente a necessidade de rotular um item com informação adicional de forma a poder ser mais facilmente recuperado ou identificado posteriormente. Diversas plataformas permitem que os utilizadores rotulem recursos com tags que habitualmente são partilhadas com outros utilizadores. Assim, ao longo do tempo foram propostas várias formas de visualização das tags associados aos recursos, com o intuito de não só facilitar aos utilizadores a pesquisa dos mesmos, mas também permitir a visualização do tag space. A nuvem de tags destaca-se como a forma mais comum de visualização. Este documento apresenta um estudo efetuado sobre formas de visualização de tags, as suas vantagens e limitações, e propõe uma forma de visualização alternativa. Sugere-se também uma nova interpretação sobre como pesquisar e visualizar recursos com tags associadas: o sistema Molecule, uma solução viável e inovadora, para vários dos problemas associados à tradicional nuvem de tags que, para além de permitir aos seus utilizadores associem tags aos s recursos, proporciona uma abordagem multivista para os mesmos navegarem no tag space e pesquisarem informação.
Resumo:
Systematics is the study of diversity of the organisms and their relationships comprising classification, nomenclature and identification. The term classification or taxonomy means the arrangement of the organisms in groups (rate) and the nomenclature is the attribution of correct international scientific names to organisms and identification is the inclusion of unknown strains in groups derived from classification. Therefore, classification for a stable nomenclature and a perfect identification are required previously. The beginning of the new bacterial systematics era can be remembered by the introduction and application of new taxonomic concepts and techniques, from the 50s and 60s. Important progress were achieved using numerical taxonomy and molecular taxonomy. Molecular taxonomy, brought into effect after the emergence of the Molecular Biology resources, provided knowledge that comprises systematics of bacteria, in which occurs great evolutionary interest, or where is observed the necessity of eliminating any environmental interference. When you study the composition and disposition of nucleotides in certain portions of the genetic material, you study searching their genome, much less susceptible to environmental alterations than proteins, codified based on it. In the molecular taxonomy, you can research both DNA and RNA, and the main techniques that have been used in the systematics comprise the build of restriction maps, DNA-DNA hybridization, DNA-RNA hybridization, sequencing of DNA sequencing of sub-units 16S and 23S of rRNA, RAPD, RFLP, PFGE etc. Techniques such as base sequencing, though they are extremely sensible and greatly precise, are relatively onerous and impracticable to the great majority of the bacterial taxonomy laboratories. Several specialized techniques have been applied to taxonomic studies of microorganisms. In the last years, these have included preliminary electrophoretic analysis of soluble proteins and isoenzymes, and subsequently determination of deoxyribonucleic acid base composition and assessment of base sequence homology by means of DNA-RNA hybrid experiments beside others. These various techniques, as expected, have generally indicated a lack of taxonomic information in microbial systematics. There are numberless techniques and methodologies that make bacteria identification and classification study possible, part of them described here, allowing establish different degrees of subspecific and interspecific similarity through phenetic-genetic polymorphism analysis. However, was pointed out the necessity of using more than one technique for better establish similarity degrees within microorganisms. Obtaining data resulting from application of a sole technique isolatedly may not provide significant information from Bacterial Systematics viewpoint
Resumo:
Plos Genetics, 5(7): ARTe1000566