908 resultados para Pre-processing


Relevância:

60.00% 60.00%

Publicador:

Resumo:

An approach for seismic damage identification of a single-storey steel concentrically braced frame (CBF) structure is presented through filtering and double integration of a recorded acceleration signal. A band-pass filter removes noise from the acceleration signal followed by baseline correction being used to reduce the drift in velocity and displacement during numerical integration. The pre-processing achieves reliable numerical integration that predicts the displacement response accurately when compared to the measured lateral in-plane displacement of the CBF structure. The lateral displacement of the CBF structure is used to infer buckling and yielding of bracing members through seismic tests. The level of interstorey drift of the CBF during a seismic excitation allows the yield and buckling of the bracing members to be identified and indirectly detects damage based on exceedance of calculated displacement limits. The calculated buckling and yielding displacement threshold limits used to identify damage are demonstrated to accurately identify initial buckling and yielding in the bracing members.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Chili powder is a globally traded commodity which has been found to be adulterated with Sudan dyes from 2003 onwards. In this study, chili powders were adulterated with varying quantities of Sudan I dye (0.1-5%) and spectra were generated using near infrared reflectance spectroscopy (NIRS) and Raman
spectroscopy (on a spectrometer with a sample compartment modified as part of the study). Chemometrics were applied to the spectral data to produce quantitative and qualitative calibration models and prediction statistics. For the quantitative models coefficients of determination (R2) were found to be
0.891-0.994 depending on which spectral data (NIRS/Raman) was processed, the mathematical algorithm used and the data pre-processing applied. The corresponding values for the root mean square error of calibration (RMSEC) and root mean square error of prediction (RMSEP) were found to be 0.208-0.851%
and 0.141-0.831% respectively, once again depending on the spectral data and the chemometric treatment applied to the data. Indications are that the NIR spectroscopy based models are superior to the models produced from Raman spectral data based on a comparison of the values of the chemometric
parameters. The limit of detection (LOD) based on analysis of 20 blank chili powders against each calibration model gave 0.25% and 0.88% for the NIR and Raman data, respectively. In addition, adopting a qualitative approach with the spectral data and applying PCA or PLS-DA, it was possible to discriminate
between adulterated chili powders from non-adulterated chili powders.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Increasingly semiconductor manufacturers are exploring opportunities for virtual metrology (VM) enabled process monitoring and control as a means of reducing non-value added metrology and achieving ever more demanding wafer fabrication tolerances. However, developing robust, reliable and interpretable VM models can be very challenging due to the highly correlated input space often associated with the underpinning data sets. A particularly pertinent example is etch rate prediction of plasma etch processes from multichannel optical emission spectroscopy data. This paper proposes a novel input-clustering based forward stepwise regression methodology for VM model building in such highly correlated input spaces. Max Separation Clustering (MSC) is employed as a pre-processing step to identify a reduced srt of well-conditioned, representative variables that can then be used as inputs to state-of-the-art model building techniques such as Forward Selection Regression (FSR), Ridge regression, LASSO and Forward Selection Ridge Regression (FCRR). The methodology is validated on a benchmark semiconductor plasma etch dataset and the results obtained are compared with those achieved when the state-of-art approaches are applied directly to the data without the MSC pre-processing step. Significant performance improvements are observed when MSC is combined with FSR (13%) and FSRR (8.5%), but not with Ridge Regression (-1%) or LASSO (-32%). The optimal VM results are obtained using the MSC-FSR and MSC-FSRR generated models. © 2012 IEEE.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A new approach for global detection of seismic damage in a single-storey steel concentrically braced frame (CBF) structure is presented. The filtered lateral in-plane acceleration response of the CBF structure is integrated twice to provide the lateral in-plane displacement which is used to infer buckling and yielding damage. The level of interstorey drift of the CBF during a seismic excitation allows the yield and buckling of the bracing members to be identified and indirectly detects damage based on exceedance of calculated lateral in-plane displacement limits. A band-pass filter removes noise from the acceleration signal followed by baseline correction being used to reduce the drift in velocity and displacement during numerical integration. This pre-processing results in reliable numerical integration of the frame acceleration that predicts the displacement response accurately when compared to the measured lateral displacement of the CBF structure. Importantly, the structural damage is not assumed through removal of bracing members, rather damage is induced through actual seismic loading. The buckling and yielding displacement threshold limits used to identify damage are demonstrated to accurately identify the initiation of buckling and yielding.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Masked implementations of cryptographic algorithms are often used in commercial embedded cryptographic devices to increase their resistance to side channel attacks. In this work we show how neural networks can be used to both identify the mask value, and to subsequently identify the secret key value with a single attack trace with high probability. We propose the use of a pre-processing step using principal component analysis (PCA) to significantly increase the success of the attack. We have developed a classifier that can correctly identify the mask for each trace, hence removing the security provided by that mask and reducing the attack to being equivalent to an attack against an unprotected implementation. The attack is performed on the freely available differential power analysis (DPA) contest data set to allow our work to be easily reproducible. We show that neural networks allow for a robust and efficient classification in the context of side-channel attacks.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

We address the problem of mining interesting phrases from subsets of a text corpus where the subset is specified using a set of features such as keywords that form a query. Previous algorithms for the problem have proposed solutions that involve sifting through a phrase dictionary based index or a document-based index where the solution is linear in either the phrase dictionary size or the size of the document subset. We propose the usage of an independence assumption between query keywords given the top correlated phrases, wherein the pre-processing could be reduced to discovering phrases from among the top phrases per each feature in the query. We then outline an indexing mechanism where per-keyword phrase lists are stored either in disk or memory, so that popular aggregation algorithms such as No Random Access and Sort-merge Join may be adapted to do the scoring at real-time to identify the top interesting phrases. Though such an approach is expected to be approximate, we empirically illustrate that very high accuracies (of over 90%) are achieved against the results of exact algorithms. Due to the simplified list-aggregation, we are also able to provide response times that are orders of magnitude better than state-of-the-art algorithms. Interestingly, our disk-based approach outperforms the in-memory baselines by up to hundred times and sometimes more, confirming the superiority of the proposed method.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Cryptographic algorithms have been designed to be computationally secure, however it has been shown that when they are implemented in hardware, that these devices leak side channel information that can be used to mount an attack that recovers the secret encryption key. In this paper an overlapping window power spectral density (PSD) side channel attack, targeting an FPGA device running the Advanced Encryption Standard is proposed. This improves upon previous research into PSD attacks by reducing the amount of pre-processing (effort) required. It is shown that the proposed overlapping window method requires less processing effort than that of using a sliding window approach, whilst overcoming the issues of sampling boundaries. The method is shown to be effective for both aligned and misaligned data sets and is therefore recommended as an improved approach in comparison with existing time domain based correlation attacks.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper describes a novel idea to identify the total number of red blood cells (RBCs) as well as their location in a Giemsa stained thin blood film image. This work is being undertaken as a part of developing an automated malaria parasite detection system by scanning a photograph of thin blood film in order to evaluate the parasitemia of the blood. Not only will this method eliminates the segmentation procedures that are normally used to segment the cells in the microscopic image, but also avoids any image pre-processing to deal with non uniform illumination prior to cell detection. The method utilizes basic knowledge on cell structure and brightness of the components due to Giemsa staining of the sample and detects and locates the RBCs in the image.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Food product safety is one of the most promising areas for the application of electronic noses. The performance of a portable electronic nose has been evaluated in monitoring the spoilage of beef fillet stored aerobically at different storage temperatures (0, 4, 8, 12, 16 and 20°C). This paper proposes a fuzzy-wavelet neural network model which incorporates a clustering pre-processing stage for the definition of fuzzy rules. The dual purpose of the proposed modeling approach is not only to classify beef samples in the respective quality class (i.e. fresh, semi-fresh and spoiled), but also to predict their associated microbiological population directly from volatile compounds fingerprints. Comparison results indicated that the proposed modeling scheme could be considered as a valuable detection methodology in food microbiology

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Freshness and safety of muscle foods are generally considered as the most important parameters for the food industry. To address the rapid detection of meat spoilage microorganisms during aerobic or modified atmosphere storage, an electronic nose with the aid of fuzzy wavelet network has been considered in this research. The proposed model incorporates a clustering pre-processing stage for the definition of fuzzy rules. The dual purpose of the proposed modelling approach is not only to classify beef samples in the respective quality class (i.e. fresh, semi-fresh and spoiled), but also to predict their associated microbiological population directly from volatile compounds fingerprints. Comparison results against neural networks and neurofuzzy systems indicated that the proposed modelling scheme could be considered as a valuable detection methodology in food microbiology

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The present research paper presents five different clustering methods to identify typical load profiles of medium voltage (MV) electricity consumers. These methods are intended to be used in a smart grid environment to extract useful knowledge about customer’s behaviour. The obtained knowledge can be used to support a decision tool, not only for utilities but also for consumers. Load profiles can be used by the utilities to identify the aspects that cause system load peaks and enable the development of specific contracts with their customers. The framework presented throughout the paper consists in several steps, namely the pre-processing data phase, clustering algorithms application and the evaluation of the quality of the partition, which is supported by cluster validity indices. The process ends with the analysis of the discovered knowledge. To validate the proposed framework, a case study with a real database of 208 MV consumers is used.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents a methodology supported on the data base knowledge discovery process (KDD), in order to find out the failure probability of electrical equipments’, which belong to a real electrical high voltage network. Data Mining (DM) techniques are used to discover a set of outcome failure probability and, therefore, to extract knowledge concerning to the unavailability of the electrical equipments such us power transformers and high-voltages power lines. The framework includes several steps, following the analysis of the real data base, the pre-processing data, the application of DM algorithms, and finally, the interpretation of the discovered knowledge. To validate the proposed methodology, a case study which includes real databases is used. This data have a heavy uncertainty due to climate conditions for this reason it was used fuzzy logic to determine the set of the electrical components failure probabilities in order to reestablish the service. The results reflect an interesting potential of this approach and encourage further research on the topic.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Relatório do Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Electrónica e Telecomunicações

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Discrete data representations are necessary, or at least convenient, in many machine learning problems. While feature selection (FS) techniques aim at finding relevant subsets of features, the goal of feature discretization (FD) is to find concise (quantized) data representations, adequate for the learning task at hand. In this paper, we propose two incremental methods for FD. The first method belongs to the filter family, in which the quality of the discretization is assessed by a (supervised or unsupervised) relevance criterion. The second method is a wrapper, where discretized features are assessed using a classifier. Both methods can be coupled with any static (unsupervised or supervised) discretization procedure and can be used to perform FS as pre-processing or post-processing stages. The proposed methods attain efficient representations suitable for binary and multi-class problems with different types of data, being competitive with existing methods. Moreover, using well-known FS methods with the features discretized by our techniques leads to better accuracy than with the features discretized by other methods or with the original features. (C) 2013 Elsevier B.V. All rights reserved.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper proposes an FPGA-based architecture for onboard hyperspectral unmixing. This method based on the Vertex Component Analysis (VCA) has several advantages, namely it is unsupervised, fully automatic, and it works without dimensionality reduction (DR) pre-processing step. The architecture has been designed for a low cost Xilinx Zynq board with a Zynq-7020 SoC FPGA based on the Artix-7 FPGA programmable logic and tested using real hyperspectral datasets. Experimental results indicate that the proposed implementation can achieve real-time processing, while maintaining the methods accuracy, which indicate the potential of the proposed platform to implement high-performance, low cost embedded systems.