18 resultados para SUBSET
em Repositório Científico do Instituto Politécnico de Lisboa - Portugal
Resumo:
A crucial method for investigating patients with coronary artery disease (CAD) is the calculation of the left ventricular ejection fraction (LVEF). It is, consequently, imperative to precisely estimate the value of LVEF--a process that can be done with myocardial perfusion scintigraphy. Therefore, the present study aimed to establish and compare the estimation performance of the quantitative parameters of the reconstruction methods filtered backprojection (FBP) and ordered-subset expectation maximization (OSEM). Methods: A beating-heart phantom with known values of end-diastolic volume, end-systolic volume, and LVEF was used. Quantitative gated SPECT/quantitative perfusion SPECT software was used to obtain these quantitative parameters in a semiautomatic mode. The Butterworth filter was used in FBP, with the cutoff frequencies between 0.2 and 0.8 cycles per pixel combined with the orders of 5, 10, 15, and 20. Sixty-three reconstructions were performed using 2, 4, 6, 8, 10, 12, and 16 OSEM subsets, combined with several iterations: 2, 4, 6, 8, 10, 12, 16, 32, and 64. Results: With FBP, the values of end-diastolic, end-systolic, and the stroke volumes rise as the cutoff frequency increases, whereas the value of LVEF diminishes. This same pattern is verified with the OSEM reconstruction. However, with OSEM there is a more precise estimation of the quantitative parameters, especially with the combinations 2 iterations × 10 subsets and 2 iterations × 12 subsets. Conclusion: The OSEM reconstruction presents better estimations of the quantitative parameters than does FBP. This study recommends the use of 2 iterations with 10 or 12 subsets for OSEM and a cutoff frequency of 0.5 cycles per pixel with the orders 5, 10, or 15 for FBP as the best estimations for the left ventricular volumes and ejection fraction quantification in myocardial perfusion scintigraphy.
Resumo:
A organização automática de mensagens de correio electrónico é um desafio actual na área da aprendizagem automática. O número excessivo de mensagens afecta cada vez mais utilizadores, especialmente os que usam o correio electrónico como ferramenta de comunicação e trabalho. Esta tese aborda o problema da organização automática de mensagens de correio electrónico propondo uma solução que tem como objectivo a etiquetagem automática de mensagens. A etiquetagem automática é feita com recurso às pastas de correio electrónico anteriormente criadas pelos utilizadores, tratando-as como etiquetas, e à sugestão de múltiplas etiquetas para cada mensagem (top-N). São estudadas várias técnicas de aprendizagem e os vários campos que compõe uma mensagem de correio electrónico são analisados de forma a determinar a sua adequação como elementos de classificação. O foco deste trabalho recai sobre os campos textuais (o assunto e o corpo das mensagens), estudando-se diferentes formas de representação, selecção de características e algoritmos de classificação. É ainda efectuada a avaliação dos campos de participantes através de algoritmos de classificação que os representam usando o modelo vectorial ou como um grafo. Os vários campos são combinados para classificação utilizando a técnica de combinação de classificadores Votação por Maioria. Os testes são efectuados com um subconjunto de mensagens de correio electrónico da Enron e um conjunto de dados privados disponibilizados pelo Institute for Systems and Technologies of Information, Control and Communication (INSTICC). Estes conjuntos são analisados de forma a perceber as características dos dados. A avaliação do sistema é realizada através da percentagem de acerto dos classificadores. Os resultados obtidos apresentam melhorias significativas em comparação com os trabalhos relacionados.
Resumo:
The general transcription factor TFIIB, encoded by SUA7 in Saccharomyces cerevisiae, is required for transcription activation but apparently of a specific subset of genes, for example, linked with mitochondrial activity and hence with oxidative environments. Therefore, studying SUA7/TFIIB as a potential target of oxidative stress is fundamental. We found that controlled SUA7 expression under oxidative conditions occurs at transcriptional and mRNA stability levels. Both regulatory events are associated with the transcription activator Yap1 in distinct ways: Yap1 affects SUA7 transcription up regulation in exponentially growing cells facing oxidative signals; the absence of this activator per se contributes to increase SUA7 mRNA stability. However, unlike SUA7 mRNA, TFIIB abundance is not altered on oxidative signals. The biological impact of this preferential regulation of SUA7 mRNA pool is revealed by the partial suppression of cellular oxidative sensitivity by SUA7 overexpression, and supported by the insights on the existence of a novel RNA-binding factor, acting as an oxidative sensor, which regulates mRNA stability. Taken together the results point out a primarily cellular commitment to guarantee SUA7 mRNA levels under oxidative environments.
Resumo:
Signal subspace identification is a crucial first step in many hyperspectral processing algorithms such as target detection, change detection, classification, and unmixing. The identification of this subspace enables a correct dimensionality reduction, yielding gains in algorithm performance and complexity and in data storage. This paper introduces a new minimum mean square error-based approach to infer the signal subspace in hyperspectral imagery. The method, which is termed hyperspectral signal identification by minimum error, is eigen decomposition based, unsupervised, and fully automatic (i.e., it does not depend on any tuning parameters). It first estimates the signal and noise correlation matrices and then selects the subset of eigenvalues that best represents the signal subspace in the least squared error sense. State-of-the-art performance of the proposed method is illustrated by using simulated and real hyperspectral images.
Resumo:
Chpater in Book Proceedings with Peer Review Second Iberian Conference, IbPRIA 2005, Estoril, Portugal, June 7-9, 2005, Proceedings, Part II
Resumo:
Research on the problem of feature selection for clustering continues to develop. This is a challenging task, mainly due to the absence of class labels to guide the search for relevant features. Categorical feature selection for clustering has rarely been addressed in the literature, with most of the proposed approaches having focused on numerical data. In this work, we propose an approach to simultaneously cluster categorical data and select a subset of relevant features. Our approach is based on a modification of a finite mixture model (of multinomial distributions), where a set of latent variables indicate the relevance of each feature. To estimate the model parameters, we implement a variant of the expectation-maximization algorithm that simultaneously selects the subset of relevant features, using a minimum message length criterion. The proposed approach compares favourably with two baseline methods: a filter based on an entropy measure and a wrapper based on mutual information. The results obtained on synthetic data illustrate the ability of the proposed expectation-maximization method to recover ground truth. An application to real data, referred to official statistics, shows its usefulness.
Resumo:
In data clustering, the problem of selecting the subset of most relevant features from the data has been an active research topic. Feature selection for clustering is a challenging task due to the absence of class labels for guiding the search for relevant features. Most methods proposed for this goal are focused on numerical data. In this work, we propose an approach for clustering and selecting categorical features simultaneously. We assume that the data originate from a finite mixture of multinomial distributions and implement an integrated expectation-maximization (EM) algorithm that estimates all the parameters of the model and selects the subset of relevant features simultaneously. The results obtained on synthetic data illustrate the performance of the proposed approach. An application to real data, referred to official statistics, shows its usefulness.
Resumo:
Dissertação apresentada à Escola Superior de Educação de Lisboa para obtenção de grau de Mestre em Intervenção Precoce
Resumo:
A unified architecture for fast and efficient computation of the set of two-dimensional (2-D) transforms adopted by the most recent state-of-the-art digital video standards is presented in this paper. Contrasting to other designs with similar functionality, the presented architecture is supported on a scalable, modular and completely configurable processing structure. This flexible structure not only allows to easily reconfigure the architecture to support different transform kernels, but it also permits its resizing to efficiently support transforms of different orders (e. g. order-4, order-8, order-16 and order-32). Consequently, not only is it highly suitable to realize high-performance multi-standard transform cores, but it also offers highly efficient implementations of specialized processing structures addressing only a reduced subset of transforms that are used by a specific video standard. The experimental results that were obtained by prototyping several configurations of this processing structure in a Xilinx Virtex-7 FPGA show the superior performance and hardware efficiency levels provided by the proposed unified architecture for the implementation of transform cores for the Advanced Video Coding (AVC), Audio Video coding Standard (AVS), VC-1 and High Efficiency Video Coding (HEVC) standards. In addition, such results also demonstrate the ability of this processing structure to realize multi-standard transform cores supporting all the standards mentioned above and that are capable of processing the 8k Ultra High Definition Television (UHDTV) video format (7,680 x 4,320 at 30 fps) in real time.
Resumo:
The methods of molecular biology applied in epidemiological research lead us to the realm of molecular epidemiology, where there is immense potential for the establishment of associations between cancer and exposure to risk factors in lifestyle, profession, or pollution. Human biomonitoring consists, on the one hand, in research and identification of hazardous environmental conditions and, on the other hand, in the assessment of cancer risk following exposure to such conditions. Since carcinogenesis is a lengthy process, the biomarkers used to recognize biological abnormalities are selected and developed in the realm of molecular epidemiology. Such biomarkers are quantifiable and allow for the recognition of progression from normal to abnormal biological conditions at the molecular level. They can be categorized in biomarkers of exposure, effect, and genetic susceptibility. Genotoxicity biomarkers are a particular subset of effect biomarkers and are used to assess genomic instability caused by environmental or occupational exposure, being considered useful carcinogenesis predictors.
Resumo:
Let X be a finite or infinite chain and let be the monoid of all endomorphisms of X. In this paper, we describe the largest regular subsemigroup of and Green's relations on. In fact, more generally, if Y is a nonempty subset of X and is the subsemigroup of of all elements with range contained in Y, we characterize the largest regular subsemigroup of and Green's relations on. Moreover for finite chains, we determine when two semigroups of the type are isomorphic and calculate their ranks.
Resumo:
Feature selection is a central problem in machine learning and pattern recognition. On large datasets (in terms of dimension and/or number of instances), using search-based or wrapper techniques can be cornputationally prohibitive. Moreover, many filter methods based on relevance/redundancy assessment also take a prohibitively long time on high-dimensional. datasets. In this paper, we propose efficient unsupervised and supervised feature selection/ranking filters for high-dimensional datasets. These methods use low-complexity relevance and redundancy criteria, applicable to supervised, semi-supervised, and unsupervised learning, being able to act as pre-processors for computationally intensive methods to focus their attention on smaller subsets of promising features. The experimental results, with up to 10(5) features, show the time efficiency of our methods, with lower generalization error than state-of-the-art techniques, while being dramatically simpler and faster.
Resumo:
Brain dopamine transporters imaging by Single Photon Emission Tomography (SPECT) with 123I-FP-CIT has become an important tool in the diagnosis and evaluation of parkinsonian syndromes, since this radiopharmaceutical exhibits high affinity for membrane transporters responsible for cellular reabsorption of dopamine on the striatum. However, Ordered Subset Expectation Maximization (OSEM) is the method recommended in the literature for imaging reconstruction. Filtered Back Projection (FBP) is still used due to its fast processing, even if it presents some disadvantages. The aim of this work is to investigate the influence of reconstruction parameters for FBP in semiquantification of Brain Studies with 123I-FPCIT compared with those obtained with OSEM recommended reconstruction.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Electrónica e telecomunicações
Resumo:
A imagem de transportadores cerebrais da dopamina com recurso à tomografia por emissão de fotão único com 123I-FP-CIT tornou-se uma ferramenta importante no diagnóstico e avaliação de síndromes parkinsonianos. Embora o algoritmo de reconstrução de imagem Ordered Subset Expectation Maximization (OSEM) seja o método mais recomendado na literatura para reconstrução da imagem, o Filtered Back Projection (FBP) é ainda usado devido à sua rapidez. O objetivo deste trabalho é investigar a influência dos parâmetros de reconstrução para FBP na semiquantificação em estudos cerebrais com 123I-FPCIT em comparação com os obtidos com a reconstrução recomendada por OSEM.