37 resultados para subset comparisons
em Repositório Científico do Instituto Politécnico de Lisboa - Portugal
Resumo:
A crucial method for investigating patients with coronary artery disease (CAD) is the calculation of the left ventricular ejection fraction (LVEF). It is, consequently, imperative to precisely estimate the value of LVEF--a process that can be done with myocardial perfusion scintigraphy. Therefore, the present study aimed to establish and compare the estimation performance of the quantitative parameters of the reconstruction methods filtered backprojection (FBP) and ordered-subset expectation maximization (OSEM). Methods: A beating-heart phantom with known values of end-diastolic volume, end-systolic volume, and LVEF was used. Quantitative gated SPECT/quantitative perfusion SPECT software was used to obtain these quantitative parameters in a semiautomatic mode. The Butterworth filter was used in FBP, with the cutoff frequencies between 0.2 and 0.8 cycles per pixel combined with the orders of 5, 10, 15, and 20. Sixty-three reconstructions were performed using 2, 4, 6, 8, 10, 12, and 16 OSEM subsets, combined with several iterations: 2, 4, 6, 8, 10, 12, 16, 32, and 64. Results: With FBP, the values of end-diastolic, end-systolic, and the stroke volumes rise as the cutoff frequency increases, whereas the value of LVEF diminishes. This same pattern is verified with the OSEM reconstruction. However, with OSEM there is a more precise estimation of the quantitative parameters, especially with the combinations 2 iterations × 10 subsets and 2 iterations × 12 subsets. Conclusion: The OSEM reconstruction presents better estimations of the quantitative parameters than does FBP. This study recommends the use of 2 iterations with 10 or 12 subsets for OSEM and a cutoff frequency of 0.5 cycles per pixel with the orders 5, 10, or 15 for FBP as the best estimations for the left ventricular volumes and ejection fraction quantification in myocardial perfusion scintigraphy.
Resumo:
Feature selection is a central problem in machine learning and pattern recognition. On large datasets (in terms of dimension and/or number of instances), using search-based or wrapper techniques can be cornputationally prohibitive. Moreover, many filter methods based on relevance/redundancy assessment also take a prohibitively long time on high-dimensional. datasets. In this paper, we propose efficient unsupervised and supervised feature selection/ranking filters for high-dimensional datasets. These methods use low-complexity relevance and redundancy criteria, applicable to supervised, semi-supervised, and unsupervised learning, being able to act as pre-processors for computationally intensive methods to focus their attention on smaller subsets of promising features. The experimental results, with up to 10(5) features, show the time efficiency of our methods, with lower generalization error than state-of-the-art techniques, while being dramatically simpler and faster.
Resumo:
A organização automática de mensagens de correio electrónico é um desafio actual na área da aprendizagem automática. O número excessivo de mensagens afecta cada vez mais utilizadores, especialmente os que usam o correio electrónico como ferramenta de comunicação e trabalho. Esta tese aborda o problema da organização automática de mensagens de correio electrónico propondo uma solução que tem como objectivo a etiquetagem automática de mensagens. A etiquetagem automática é feita com recurso às pastas de correio electrónico anteriormente criadas pelos utilizadores, tratando-as como etiquetas, e à sugestão de múltiplas etiquetas para cada mensagem (top-N). São estudadas várias técnicas de aprendizagem e os vários campos que compõe uma mensagem de correio electrónico são analisados de forma a determinar a sua adequação como elementos de classificação. O foco deste trabalho recai sobre os campos textuais (o assunto e o corpo das mensagens), estudando-se diferentes formas de representação, selecção de características e algoritmos de classificação. É ainda efectuada a avaliação dos campos de participantes através de algoritmos de classificação que os representam usando o modelo vectorial ou como um grafo. Os vários campos são combinados para classificação utilizando a técnica de combinação de classificadores Votação por Maioria. Os testes são efectuados com um subconjunto de mensagens de correio electrónico da Enron e um conjunto de dados privados disponibilizados pelo Institute for Systems and Technologies of Information, Control and Communication (INSTICC). Estes conjuntos são analisados de forma a perceber as características dos dados. A avaliação do sistema é realizada através da percentagem de acerto dos classificadores. Os resultados obtidos apresentam melhorias significativas em comparação com os trabalhos relacionados.
Resumo:
A maioria das infra-estruturas de transportes, nomeadamente os pavimentos rodoviários e aeroportuários, são constituídas por misturas betuminosas, o que permite um bom desempenho e uma adequada durabilidade, nas condições usuais de serviço. As misturas betuminosas são ainda amplamente utilizadas na construção de zonas de estacionamento de veículos, tendo-se verificado recentemente a sua aplicação também em infra-estruturas ferroviárias. Face à necessidade de melhorar o desempenho das vias-férreas, permitindo uma concepção mais durável de linhas de alta velocidade e uma redução dos custos da sua manutenção, tem-se vindo a desenvolver diversos estudos para promover a utilização de novos materiais, principalmente através da incorporação de misturas betuminosas. O presente trabalho tem como objectivo a caracterização do comportamento mecânico de misturas betuminosas a aplicar em infra-estruturas de transportes. Como metodologia para o estudo do comportamento mecânico das misturas betuminosas foram realizados em laboratório ensaios de cargas repetidas, nomeadamente, ensaios de flexão em quatro pontos para determinação da rigidez e da resistência à fadiga e ensaios de compressão triaxiais cíclicos para avaliação do comportamento à deformação permanente. A resistência à fadiga das misturas betuminosas em estudo foi avaliada através do ensaio de flexão em quatro pontos, com extensão controlada, e aplicação de um carregamento sinusoidal com diferentes frequências, de acordo com o procedimento de ensaio da norma europeia EN 12697-24 (2004 + A1: 2007). A resistência à deformação permanente das misturas betuminosas foi analisada através de ensaios de compressão triaxiais cíclicos, submetendo-as a uma tensão de confinamento estática pela aplicação parcial de vácuo e a uma pressão axial cíclica sob a forma rectangular, de acordo com a norma europeia EN 12697-25 (2004). O conhecimento destas propriedades mecânicas assume particular importância ao nível da formulação das misturas betuminosas, do dimensionamento de uma estrutura ou do estabelecimento de uma adequada solução para uma obra de reabilitação duma infra-estrutura de transportes. Para este estudo foi utilizado um modelo físico construído numa fossa no LNEC, com o propósito de serem testadas três substruturas ferroviárias não convencionais, utilizando sub-balastro betuminoso. A selecção das substruturas foi efectuada após uma análise de várias secções de estruturas já testadas e aplicadas noutros países, de forma a proporcionar comparações fiáveis entre elas. Os resultados obtidos mostraram que a mistura betuminosa AC20 base 50/70 (MB) aplicada na camada de sub-balastro é adequada para ser aplicada nas infra-estruturas de transportes pois apresenta um bom desempenho à fadiga e à deformação permanente. Através dos ensaios efectuados foi ainda possível entender a importante influência das características volumétricas, principalmente da porosidade para o bom comportamento da mistura betuminosa.
Resumo:
A replicate evaluation of increased micronucleus (MN) frequencies in peripheral lymphocytes of workers occupationally exposed to formaldehyde (FA) was undertaken to verify the observed effect and to determine scoring variability. May–Grünwald–Giemsa-stained slides were obtained from a previously performed cytokinesis-block micronucleus test (CBMNT) with 56 workers in anatomy and pathology laboratories and 85 controls. The first evaluation by one scorer (scorer 1) had led to a highly significant difference between workers and controls (3.96 vs 0.81 MN per 1000 cells). The slides were coded before re-evaluation and the code was broken after the complete re-evaluation of the study. A total of 1000 binucleated cells (BNC) were analysed per subject and the frequency of MN (in ‰) was determined. Slides were distributed equally and randomly between two scorers, so that the scorers had no knowledge of the exposure status. Scorer 2 (32 exposed, 36 controls) measured increased MN frequencies in exposed workers (9.88 vs 6.81). Statistical analysis with the two-sample Wilcoxon test indicated that this difference was not significant (p = 0.17). Scorer 3 (20 exposed, 46 controls) obtained a similar result, but slightly higher values for the comparison of exposed and controls (19.0 vs 12.89; p = 0.089). Combining the results of the two scorers (13.38 vs 10.22), a significant difference between exposed and controls (p = 0.028) was obtained when the stratified Wilcoxon test with the scorers as strata was applied. Interestingly, the re-evaluation of the slides led to clearly higher MN frequencies for exposed and controls compared with the first evaluation. Bland–Altman plots indicated that the agreement between the measurements of the different scorers was very poor, as shown by mean differences of 5.9 between scorer 1 and scorer 2 and 13.0 between scorer 1 and scorer 3. Calculation of the intra-class correlation coefficient (ICC) revealed that all scorer comparisons in this study were far from acceptable for the reliability of this assay. Possible implications for the use of the CBMNT in human biomonitoring studies are discussed.
Resumo:
Objective - To describe and validate the simulation of the basic features of GE Millennium MG gamma camera using the GATE Monte Carlo platform. Material and methods - Crystal size and thickness, parallel-hole collimation and a realistic energy acquisition window were simulated in the GATE platform. GATE results were compared to experimental data in the following imaging conditions: a point source of 99mTc at different positions during static imaging and tomographic acquisitions using two different energy windows. The accuracy between the events expected and detected by simulation was obtained with the Mann–Whitney–Wilcoxon test. Comparisons were made regarding the measurement of sensitivity and spatial resolution, static and tomographic. Simulated and experimental spatial resolutions for tomographic data were compared with the Kruskal–Wallis test to assess simulation accuracy for this parameter. Results - There was good agreement between simulated and experimental data. The number of decays expected when compared with the number of decays registered, showed small deviation (≤0.007%). The sensitivity comparisons between static acquisitions for different distances from source to collimator (1, 5, 10, 20, 30cm) with energy windows of 126–154 keV and 130–158 keV showed differences of 4.4%, 5.5%, 4.2%, 5.5%, 4.5% and 5.4%, 6.3%, 6.3%, 5.8%, 5.3%, respectively. For the tomographic acquisitions, the mean differences were 7.5% and 9.8% for the energy window 126–154 keV and 130–158 keV. Comparison of simulated and experimental spatial resolutions for tomographic data showed no statistically significant differences with 95% confidence interval. Conclusions - Adequate simulation of the system basic features using GATE Monte Carlo simulation platform was achieved and validated.
Resumo:
Beaches worldwide provide recreational opportunities to hundreds of millions of people and serve as important components of coastal economies. Beach water is often monitored for microbiological quality to detect the presence of indicators of human sewage contamination so as to prevent public health outbreaks associated with water contact. However, growing evidence suggests that beach sand can harbor microbes harmful to human health, often in concentrations greater than the beach water. Currently, there are no standards for monitoring, sampling, analyzing, or managing beach sand quality. In addition to indicator microbes, growing evidence has identified pathogenic bacteria, viruses, and fungi in a variety of beach sands worldwide. The public health threat associated with these populations through direct and indirect contact is unknown because so little research has been conducted relating to health outcomes associated with sand quality. In this manuscript, we present the consensus findings of a workshop of experts convened in Lisbon, Portugal to discuss the current state of knowledge on beach sand microbiological quality and to develop suggestions for standardizing the evaluation of sand at coastal beaches. The expert group at the "Microareias 2012" workshop recommends that 1) beach sand should be screened for a variety of pathogens harmful to human health, and sand monitoring should then be initiated alongside regular water monitoring; 2) sampling and analysis protocols should be standardized to allow proper comparisons among beach locations; and 3) further studies are needed to estimate human health risk with exposure to contaminated beach sand. Much of the manuscript is focused on research specific to Portugal, but similar results have been found elsewhere, and the findings have worldwide implications.
Resumo:
The general transcription factor TFIIB, encoded by SUA7 in Saccharomyces cerevisiae, is required for transcription activation but apparently of a specific subset of genes, for example, linked with mitochondrial activity and hence with oxidative environments. Therefore, studying SUA7/TFIIB as a potential target of oxidative stress is fundamental. We found that controlled SUA7 expression under oxidative conditions occurs at transcriptional and mRNA stability levels. Both regulatory events are associated with the transcription activator Yap1 in distinct ways: Yap1 affects SUA7 transcription up regulation in exponentially growing cells facing oxidative signals; the absence of this activator per se contributes to increase SUA7 mRNA stability. However, unlike SUA7 mRNA, TFIIB abundance is not altered on oxidative signals. The biological impact of this preferential regulation of SUA7 mRNA pool is revealed by the partial suppression of cellular oxidative sensitivity by SUA7 overexpression, and supported by the insights on the existence of a novel RNA-binding factor, acting as an oxidative sensor, which regulates mRNA stability. Taken together the results point out a primarily cellular commitment to guarantee SUA7 mRNA levels under oxidative environments.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Electrónica e Telecomunicações
Resumo:
Purpose – Quantitative instruments to assess patient safety culture have been developed recently and a few review articles have been published. Measuring safety culture enables healthcare managers and staff to improve safety behaviours and outcomes for patients and staff. The study aims to determine the AHRQ Hospital Survey on Patient Safety Culture (HSPSC) Portuguese version's validity and reliability. Design/methodology/approach – A missing-value analysis and item analysis was performed to identify problematic items. Reliability analysis, inter-item correlations and inter-scale correlations were done to check internal consistency, composite scores. Inter-correlations were examined to assess construct validity. A confirmatory factor analysis was performed to investigate the observed data's fit to the dimensional structure proposed in the AHRQ HSPSC Portuguese version. To analyse differences between hospitals concerning composites scores, an ANOVA analysis and multiple comparisons were done. Findings – Eight of 12 dimensions had Cronbach's alphas higher than 0.7. The instrument as a whole achieved a high Cronbach's alpha (0.91). Inter-correlations showed that there is no dimension with redundant items, however dimension 10 increased its internal consistency when one item is removed. Originality/value – This study is the first to evaluate an American patient safety culture survey using Portuguese data. The survey has satisfactory reliability and construct validity.
Resumo:
Signal subspace identification is a crucial first step in many hyperspectral processing algorithms such as target detection, change detection, classification, and unmixing. The identification of this subspace enables a correct dimensionality reduction, yielding gains in algorithm performance and complexity and in data storage. This paper introduces a new minimum mean square error-based approach to infer the signal subspace in hyperspectral imagery. The method, which is termed hyperspectral signal identification by minimum error, is eigen decomposition based, unsupervised, and fully automatic (i.e., it does not depend on any tuning parameters). It first estimates the signal and noise correlation matrices and then selects the subset of eigenvalues that best represents the signal subspace in the least squared error sense. State-of-the-art performance of the proposed method is illustrated by using simulated and real hyperspectral images.
Resumo:
Chpater in Book Proceedings with Peer Review Second Iberian Conference, IbPRIA 2005, Estoril, Portugal, June 7-9, 2005, Proceedings, Part II
Resumo:
Research on the problem of feature selection for clustering continues to develop. This is a challenging task, mainly due to the absence of class labels to guide the search for relevant features. Categorical feature selection for clustering has rarely been addressed in the literature, with most of the proposed approaches having focused on numerical data. In this work, we propose an approach to simultaneously cluster categorical data and select a subset of relevant features. Our approach is based on a modification of a finite mixture model (of multinomial distributions), where a set of latent variables indicate the relevance of each feature. To estimate the model parameters, we implement a variant of the expectation-maximization algorithm that simultaneously selects the subset of relevant features, using a minimum message length criterion. The proposed approach compares favourably with two baseline methods: a filter based on an entropy measure and a wrapper based on mutual information. The results obtained on synthetic data illustrate the ability of the proposed expectation-maximization method to recover ground truth. An application to real data, referred to official statistics, shows its usefulness.
Resumo:
In data clustering, the problem of selecting the subset of most relevant features from the data has been an active research topic. Feature selection for clustering is a challenging task due to the absence of class labels for guiding the search for relevant features. Most methods proposed for this goal are focused on numerical data. In this work, we propose an approach for clustering and selecting categorical features simultaneously. We assume that the data originate from a finite mixture of multinomial distributions and implement an integrated expectation-maximization (EM) algorithm that estimates all the parameters of the model and selects the subset of relevant features simultaneously. The results obtained on synthetic data illustrate the performance of the proposed approach. An application to real data, referred to official statistics, shows its usefulness.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Mecânica