959 resultados para algoritmi non evolutivi pattern recognition analisi dati avanzata metodi matematici intelligenza artificiale non evolutive algorithms artificial intelligence


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The identification of gasoline adulteration by organic solvents is not an easy task, because compounds that constitute the solvents are already in gasoline composition. In this work, the combination of Hydrogen Nuclear Magnetic Resonance ((1)H NMR) spectroscopic fingerprintings with pattern-recognition multivariate Soft Independent Modeling of Class Analogy (SIMCA) chemometric analysis provides an original and alternative approach to screening Brazilian commercial gasoline quality in a Monitoring Program for Quality Control of Automotive Fuels. SIMCA was performed on spectroscopic fingerprints to classify the quality of representative commercial gasoline samples selected by Hierarchical Cluster Analysis (HCA) and collected over a 6-month period from different gas stations in the São Paulo state, Brazil. Following optimized the (1)H NMR-SIMCA algorithm, it was possible to correctly classify 92.0% of commercial gasoline samples, which is considered acceptable. The chemometric method is recommended for routine applications in Quality-Control Monitoring Programs, since its measurements are fast and can be easily automated. Also, police laboratories could employ this method for rapid screening analysis to discourage adulteration practices. (C) 2010 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A body of research has developed within the context of nonlinear signal and image processing that deals with the automatic, statistical design of digital window-based filters. Based on pairs of ideal and observed signals, a filter is designed in an effort to minimize the error between the ideal and filtered signals. The goodness of an optimal filter depends on the relation between the ideal and observed signals, but the goodness of a designed filter also depends on the amount of sample data from which it is designed. In order to lessen the design cost, a filter is often chosen from a given class of filters, thereby constraining the optimization and increasing the error of the optimal filter. To a great extent, the problem of filter design concerns striking the correct balance between the degree of constraint and the design cost. From a different perspective and in a different context, the problem of constraint versus sample size has been a major focus of study within the theory of pattern recognition. This paper discusses the design problem for nonlinear signal processing, shows how the issue naturally transitions into pattern recognition, and then provides a review of salient related pattern-recognition theory. In particular, it discusses classification rules, constrained classification, the Vapnik-Chervonenkis theory, and implications of that theory for morphological classifiers and neural networks. The paper closes by discussing some design approaches developed for nonlinear signal processing, and how the nature of these naturally lead to a decomposition of the error of a designed filter into a sum of the following components: the Bayes error of the unconstrained optimal filter, the cost of constraint, the cost of reducing complexity by compressing the original signal distribution, the design cost, and the contribution of prior knowledge to a decrease in the error. The main purpose of the paper is to present fundamental principles of pattern recognition theory within the framework of active research in nonlinear signal processing.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Structural health monitoring (SHM) is related to the ability of monitoring the state and deciding the level of damage or deterioration within aerospace, civil and mechanical systems. In this sense, this paper deals with the application of a two-step auto-regressive and auto-regressive with exogenous inputs (AR-ARX) model for linear prediction of damage diagnosis in structural systems. This damage detection algorithm is based on the. monitoring of residual error as damage-sensitive indexes, obtained through vibration response measurements. In complex structures there are. many positions under observation and a large amount of data to be handed, making difficult the visualization of the signals. This paper also investigates data compression by using principal component analysis. In order to establish a threshold value, a fuzzy c-means clustering is taken to quantify the damage-sensitive index in an unsupervised learning mode. Tests are made in a benchmark problem, as proposed by IASC-ASCE with different damage patterns. The diagnosis that was obtained showed high correlation with the actual integrity state of the structure. Copyright © 2007 by ABCM.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The results obtained through biological research usually need to be analyzed using computational tools, since manual analysis becomes unfeasible due to the complexity and size of these results. For instance, the study of quasispecies frequently demands the analysis of several, very lengthy sequences of nucleotides and amino acids. Therefore, bioinformatics tools for the study of quasispecies are constantly being developed due to different problems found by biologists. In the present study, we address the development of a software tool for the evaluation of population diversity in quasispecies. Special attention is paid to the localization of genome regions prone to changes, as well as of possible hot spots.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this project, the main focus is to apply image processing techniques in computer vision through an omnidirectional vision system to agricultural mobile robots (AMR) used for trajectory navigation problems, as well as localization matters. To carry through this task, computational methods based on the JSEG algorithm were used to provide the classification and the characterization of such problems, together with Artificial Neural Networks (ANN) for pattern recognition. Therefore, it was possible to run simulations and carry out analyses of the performance of JSEG image segmentation technique through Matlab/Octave platforms, along with the application of customized Back-propagation algorithm and statistical methods as structured heuristics methods in a Simulink environment. Having the aforementioned procedures been done, it was practicable to classify and also characterize the HSV space color segments, not to mention allow the recognition of patterns in which reasonably accurate results were obtained. ©2010 IEEE.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This is a preliminary theoretical discussion on the computational requirements of the state of the art smoothed particle hydrodynamics (SPH) from the optics of pattern recognition and artificial intelligence. It is pointed out in the present paper that, when including anisotropy detection to improve resolution on shock layer, SPH is a very peculiar case of unsupervised machine learning. On the other hand, the free particle nature of SPH opens an opportunity for artificial intelligence to study particles as agents acting in a collaborative framework in which the timed outcomes of a fluid simulation forms a large knowledge base, which might be very attractive in computational astrophysics phenomenological problems like self-propagating star formation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This study evaluated the expression of pattern recognition receptors (PRRs) and activation factors associated with salivary and blood neutrophils from different aged patients diagnosed with Candida-related denture stomatitis (DS). Expression of neutrophil PRRs was determined by flow cytometry and immunofluorescence, and the levels of selected cytokines that influence immune activation were determined by ELISA. The salivary (but not the serum derived) neutrophils of individuals with DS were found to have an increased expression of CD69 regardless of the age of the patient compared to patients without DS. However, these salivary neutrophils had a lower expression of CD66b and CD64. Expression of TLR2 was lower on the salivary-and serum-derived neutrophils from elderly individuals compared to the neutrophils of younger subjects, regardless of whether the individual had DS. Salivary interleukin (IL)-4 was elevated in both of the elderly subject groups (with or without DS). Only elderly DS patients were observed to have increased serum IL-4 levels and reduced salivary IL-12 levels. Younger DS patients showed an increase in salivary IL-10 levels, and both the saliva and the serum levels of IFN-gamma were increased in all of the younger subjects. Our data demonstrated that changes in both the oral immune cells and the protein components could be associated with DS. Furthermore, changes in the blood-derived factors were more associated with age than DS status. (C) 2012 Elsevier Inc. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

One of the problems in the analysis of nucleus-nucleus collisions is to get information on the value of the impact parameter b. This work consists in the application of pattern recognition techniques aimed at associating values of b to groups of events. To this end, a support vec- tor machine (SVM) classifier is adopted to analyze multifragmentation reactions. This method allows to backtracing the values of b through a particular multidimensional analysis. The SVM classification con- sists of two main phase. In the first one, known as training phase, the classifier learns to discriminate the events that are generated by two different model:Classical Molecular Dynamics (CMD) and Heavy- Ion Phase-Space Exploration (HIPSE) for the reaction: 58Ni +48 Ca at 25 AMeV. To check the classification of events in the second one, known as test phase, what has been learned is tested on new events generated by the same models. These new results have been com- pared to the ones obtained through others techniques of backtracing the impact parameter. Our tests show that, following this approach, the central collisions and peripheral collisions, for the CMD events, are always better classified with respect to the classification by the others techniques of backtracing. We have finally performed the SVM classification on the experimental data measured by NUCL-EX col- laboration with CHIMERA apparatus for the previous reaction.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The study is aimed to calculate an innovative numerical index for bit performance evaluation called Bit Index (BI), applied on a new type of bit database named Formation Drillability Catalogue (FDC). A dedicated research programme (developed by Eni E&P and the University of Bologna) studied a drilling model for bit performance evaluation named BI, derived from data recorded while drilling (bit records, master log, wireline log, etc.) and dull bit evaluation. This index is calculated with data collected inside the FDC, a novel classification of Italian formations aimed to the geotechnical and geomechanical characterization and subdivisions of the formations, called Minimum Interval (MI). FDC was conceived and prepared at Eni E&P Div., and contains a large number of significant drilling parameters. Five wells have been identified inside the FDC and have been tested for bit performance evaluation. The values of BI are calculated for each bit run and are compared with the values of the cost per metre. The case study analyzes bits of the same type, diameters and run in the same formation. The BI methodology implemented on MI classification of FDC can improve consistently the bit performances evaluation, and it helps to identify the best performer bits. Moreover, FDC turned out to be functional to BI, since it discloses and organizes formation details that are not easily detectable or usable from bit records or master logs, allowing for targeted bit performance evaluations. At this stage of development, the BI methodology proved to be economic and reliable. The quality of bit performance analysis obtained with BI seems also more effective than the traditional “quick look” analysis, performed on bit records, or on the pure cost per metre evaluation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Recenti studi hanno evidenziato come il cammino in ambiente acquatico possa portare a notevoli benefici nell’ambito di un processo riabilitativo: il cammino in acqua è infatti oggi considerato una delle principali terapie per pazienti con disturbi nella deambulazione, oltre ad essere impiegato per migliorare il recupero a seguito di interventi ed infortuni. Una caratterizzazione biomeccanica del cammino umano in acqua permetterebbe tuttavia di giungere a una conoscenza più approfondita degli effetti di quest’attività sul processo riabilitativo, e dunque a una sua prescrizione più mirata come parte delle terapie. Nonostante il crescente interesse, uno dei motivi per cui ancora pochi studi sono stati condotti in questo senso risiede nell’inadeguatezza di molti dei tradizionali sistemi di Motion Capture rispetto all’impiego subacqueo. La nuova branca della Markerless Motion Capture potrebbe invece in questo senso rappresentare una soluzione. In particolare, ci si occuperà in questo lavoro di tesi della tecnica markerless basata sulla ricostruzione del visual hull per retroproiezione delle silhouette. Il processo iniziale che permette di ottenere le silhouette dai video delle acquisizioni è detto segmentazione, la quale è anche una fase particolarmente importante per ottenere una buona accuratezza finale nella ricostruzione della cinematica articolare. Si sono pertanto sviluppati e caratterizzati in questo lavoro di tesi sette algoritmi di segmentazione, nati specificamente nell’ottica dell’analisi del cammino in acqua con tecnica markerless. Si mostrerà inoltre come determinate caratteristiche degli algoritmi influenzino la qualità finale della segmentazione, e sarà infine presentato un ulteriore algoritmo di post-processing per il miglioramento della qualità delle immagini segmentate.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Progettazione e implementazione dei moduli di visualizzazione, memorizzazione e analisi di un sistema software di acquisizione dati in real-time da dispositivi prodotti da Elements s.r.l. La tesi mostra tutte le fasi di analisi, progettazione, implementazione e testing dei moduli sviluppati.