26 resultados para Wavelet Packet and Support Vector Machine
em Universidad Politécnica de Madrid
Resumo:
El análisis de imágenes hiperespectrales permite obtener información con una gran resolución espectral: cientos de bandas repartidas desde el espectro infrarrojo hasta el ultravioleta. El uso de dichas imágenes está teniendo un gran impacto en el campo de la medicina y, en concreto, destaca su utilización en la detección de distintos tipos de cáncer. Dentro de este campo, uno de los principales problemas que existen actualmente es el análisis de dichas imágenes en tiempo real ya que, debido al gran volumen de datos que componen estas imágenes, la capacidad de cómputo requerida es muy elevada. Una de las principales líneas de investigación acerca de la reducción de dicho tiempo de procesado se basa en la idea de repartir su análisis en diversos núcleos trabajando en paralelo. En relación a esta línea de investigación, en el presente trabajo se desarrolla una librería para el lenguaje RVC – CAL – lenguaje que está especialmente pensado para aplicaciones multimedia y que permite realizar la paralelización de una manera intuitiva – donde se recogen las funciones necesarias para implementar el clasificador conocido como Support Vector Machine – SVM. Cabe mencionar que este trabajo complementa el realizado en [1] y [2] donde se desarrollaron las funciones necesarias para implementar una cadena de procesado que utiliza el método unmixing para procesar la imagen hiperespectral. En concreto, este trabajo se encuentra dividido en varias partes. La primera de ellas expone razonadamente los motivos que han llevado a comenzar este Trabajo de Investigación y los objetivos que se pretenden conseguir con él. Tras esto, se hace un amplio estudio del estado del arte actual y, en él, se explican tanto las imágenes hiperespectrales como sus métodos de procesado y, en concreto, se detallará el método que utiliza el clasificador SVM. Una vez expuesta la base teórica, nos centraremos en la explicación del método seguido para convertir una versión en Matlab del clasificador SVM optimizado para analizar imágenes hiperespectrales; un punto importante en este apartado es que se desarrolla la versión secuencial del algoritmo y se asientan las bases para una futura paralelización del clasificador. Tras explicar el método utilizado, se exponen los resultados obtenidos primero comparando ambas versiones y, posteriormente, analizando por etapas la versión adaptada al lenguaje RVC – CAL. Por último, se aportan una serie de conclusiones obtenidas tras analizar las dos versiones del clasificador SVM en cuanto a bondad de resultados y tiempos de procesado y se proponen una serie de posibles líneas de actuación futuras relacionadas con dichos resultados. ABSTRACT. Hyperspectral imaging allows us to collect high resolution spectral information: hundred of bands covering from infrared to ultraviolet spectrum. These images have had strong repercussions in the medical field; in particular, we must highlight its use in cancer detection. In this field, the main problem we have to deal with is the real time analysis, because these images have a great data volume and they require a high computational power. One of the main research lines that deals with this problem is related with the analysis of these images using several cores working at the same time. According to this investigation line, this document describes the development of a RVC – CAL library – this language has been widely used for working with multimedia applications and allows an optimized system parallelization –, which joins all the functions needed to implement the Support Vector Machine – SVM - classifier. This research complements the research conducted in [1] and [2] where the necessary functions to implement the unmixing method to analyze hyperspectral images were developed. The document is divided in several chapters. The first of them introduces the motivation of the Master Thesis and the main objectives to achieve. After that, we study the state of the art of some technologies related with this work, like hyperspectral images, their processing methods and, concretely, the SVM classifier. Once we have exposed the theoretical bases, we will explain the followed methodology to translate a Matlab version of the SVM classifier optimized to process an hyperspectral image to RVC – CAL language; one of the most important issues in this chapter is that a sequential implementation is developed and the bases of a future parallelization of the SVM classifier are set. At this point, we will expose the results obtained in the comparative between versions and then, the results of the different steps that compose the SVM in its RVC – CAL version. Finally, we will extract some conclusions related with algorithm behavior and time processing. In the same way, we propose some future research lines according to the results obtained in this document.
Resumo:
To develop a Support Vector Machine (SVM) algorithm as a predictive tool for diagnostic outcome in patients with FE-EOP, based on clinical and biomedical data at the emergence of the illness.
Resumo:
The application of thematic maps obtained through the classification of remote images needs the obtained products with an optimal accuracy. The registered images from the airplanes display a very satisfactory spatial resolution, but the classical methods of thematic classification not always give better results than when the registered data from satellite are used. In order to improve these results of classification, in this work, the LIDAR sensor data from first return (Light Detection And Ranging) registered simultaneously with the spectral sensor data from airborne are jointly used. The final results of the thematic classification of the scene object of study have been obtained, quantified and discussed with and without LIDAR data, after applying different methods: Maximum Likehood Classification, Support Vector Machine with four different functions kernel and Isodata clustering algorithm (ML, SVM-L, SVM-P, SVM-RBF, SVM-S, Isodata). The best results are obtained for SVM with Sigmoide kernel. These allow the correlation with others different physical parameters with great interest like Manning hydraulic coefficient, for their incorporation in a GIS and their application in hydraulic modeling.
Resumo:
In this work we propose an image acquisition and processing methodology (framework) developed for performance in-field grapes and leaves detection and quantification, based on a six step methodology: 1) image segmentation through Fuzzy C-Means with Gustafson Kessel (FCM-GK) clustering; 2) obtaining of FCM-GK outputs (centroids) for acting as seeding for K-Means clustering; 3) Identification of the clusters generated by K-Means using a Support Vector Machine (SVM) classifier. 4) Performance of morphological operations over the grapes and leaves clusters in order to fill holes and to eliminate small pixels clusters; 5)Creation of a mosaic image by Scale-Invariant Feature Transform (SIFT) in order to avoid overlapping between images; 6) Calculation of the areas of leaves and grapes and finding of the centroids in the grape bunches. Image data are collected using a colour camera fixed to a mobile platform. This platform was developed to give a stabilized surface to guarantee that the images were acquired parallel to de vineyard rows. In this way, the platform avoids the distortion of the images that lead to poor estimation of the areas. Our preliminary results are promissory, although they still have shown that it is necessary to implement a camera stabilization system to avoid undesired camera movements, and also a parallel processing procedure in order to speed up the mosaicking process.
Resumo:
This paper outlines an automatic computervision system for the identification of avena sterilis which is a special weed seed growing in cereal crops. The final goal is to reduce the quantity of herbicide to be sprayed as an important and necessary step for precision agriculture. So, only areas where the presence of weeds is important should be sprayed. The main problems for the identification of this kind of weed are its similar spectral signature with respect the crops and also its irregular distribution in the field. It has been designed a new strategy involving two processes: image segmentation and decision making. The image segmentation combines basic suitable image processing techniques in order to extract cells from the image as the low level units. Each cell is described by two area-based attributes measuring the relations among the crops and weeds. The decision making is based on the SupportVectorMachines and determines if a cell must be sprayed. The main findings of this paper are reflected in the combination of the segmentation and the SupportVectorMachines decision processes. Another important contribution of this approach is the minimum requirements of the system in terms of memory and computation power if compared with other previous works. The performance of the method is illustrated by comparative analysis against some existing strategies.
Resumo:
New forms of natural interactions between human operators and UAVs (Unmanned Aerial Vehicle) are demanded by the military industry to achieve a better balance of the UAV control and the burden of the human operator. In this work, a human machine interface (HMI) based on a novel gesture recognition system using depth imagery is proposed for the control of UAVs. Hand gesture recognition based on depth imagery is a promising approach for HMIs because it is more intuitive, natural, and non-intrusive than other alternatives using complex controllers. The proposed system is based on a Support Vector Machine (SVM) classifier that uses spatio-temporal depth descriptors as input features. The designed descriptor is based on a variation of the Local Binary Pattern (LBP) technique to efficiently work with depth video sequences. Other major consideration is the especial hand sign language used for the UAV control. A tradeoff between the use of natural hand signs and the minimization of the inter-sign interference has been established. Promising results have been achieved in a depth based database of hand gestures especially developed for the validation of the proposed system.
Resumo:
Among those damage identification methods, the Wavelet Packet Energy Curvature Difference (WPECD) Method is an effective one. However, most of the existing methods rely on numerical simulation and are unverified via experiment, and very few of them have been applied to practice. In this paper, the validity of WPECD in structural damage identification is verified by a numerical example. A damage simulation experiment is taken on a real replaced girder at the Ziya River New Bridge in Cangzhou. Two damage cases are applied and the acceleration responses at the measuring points are obtained, based on which the damages are identified with the WPECD Method, and the influence of wavelet function and decomposition level is studied. The results show that the WPECD Method can identify structure damage efficiently and can be put into practice.
Implementation of the disruption predictor APODIS in JET Real Time Network using the MARTe framework
Resumo:
Disruptions in tokamaks devices are unavoidable, and they can have a significant impact on machine integrity. So it is very important have mechanisms to predict this phenomenon. Disruption prediction is a very complex task, not only because it is a multi-dimensional problem, but also because in order to be effective, it has to detect well in advance the actual disruptive event, in order to be able to use successful mitigation strategies. With these constraints in mind a real-time disruption predictor has been developed to be used in JET tokamak. The predictor has been designed to run in the Multithreaded Application Real-Time executor (MARTe) framework. The predictor ?Advanced Predictor Of DISruptions? (APODIS) is based on Support Vector Machine (SVM).
Resumo:
This paper presents a multi-stage algorithm for the dynamic condition monitoring of a gear. The algorithm provides information referred to the gear status (fault or normal condition) and estimates the mesh stiffness per shaft revolution in case that any abnormality is detected. In the first stage, the analysis of coefficients generated through discrete wavelet transformation (DWT) is proposed as a fault detection and localization tool. The second stage consists in establishing the mesh stiffness reduction associated with local failures by applying a supervised learning mode and coupled with analytical models. To do this, a multi-layer perceptron neural network has been configured using as input features statistical parameters sensitive to torsional stiffness decrease and derived from wavelet transforms of the response signal. The proposed method is applied to the gear condition monitoring and results show that it can update the mesh dynamic properties of the gear on line.
Resumo:
For the decades to come can be foreseen that electricity and water will keep be playing a key role in the countries development, both can be considered the most important energy vectors and its control can be crucial for governments, companies and leaders in general. Energy is essential for all human activities and its availability is critical to economic and social development. In particular, electricity, a form of energy, is required to produce goods, to provide medical assistance and basic civic services in education, to assure availability of clean water, to create conducive environment for prosperity and improvement, and to keep an acceptable quality of life. The way in which electricity is generated from different resources varies through the different countries. Nuclear energy controlled within reactors to steam production, gas, fuel-oil and coal fired in power stations, water, solar and wind energy among others are employed, sometimes not very efficiently, to produce electricity. The so call energy mix of an individual country is formed up by the contribution of each resource or form of energy to the electricity generation market of the so country. During the last decade the establishment of proper energy mixes for countries has gained much importance, and energy drivers should enforce long term plans and policies. Hints, reports and guides giving tracks on energy resources contribution are been developed by noticeable organisations like the IEA (International Energy Agency) or the IAEA (International Atomic Energy Agency) and the WEC (World Energy Council). This paper evaluates energy issues the market and countries are facing today regarding energy mix scheduling and panorama. This paper revises and seeks to improve methodology available that are applicable on energy mix plan definition. Key Factors are identified, established and assessed through this paper for the common implementation, the themes driving the future energy mix methodology proposal. Those have a clear influence and are closely related to future environmental policies. Key Factors take into consideration sustainability, energy security, social and economic growth, climate change, air quality and social stability. The strength of the Key Factors application on energy system planning to different countries is contingent on country resources, location, electricity demand and electricity generation industry, technology available, economic situation and prospects, energy policy and regulation
Resumo:
In this dissertation, after testing that neither the definition of Agile methodologies, nor the current tools that support them, such as Scrum or XP, gave guidance for stages of software development prior to the definition of the first interaction of development; we proceeded to study the state of the art of Inception techniques, that is, techniques to deal with this early phase of the project, that would help guide its development. From the analysis of these Inception techniques, we defined what we considered as the essential properties of an Inception framework. With that list at hand, it was found that no current Inception framework supported all the features, also, we found that it did not exist, either, any software application on the market that did it. Finally, after checking the above gaps, we defined the Inception framework "Agile Incepti-ON", with all the practices necessary to meet the requirements specified above. In addition to this, a software application was developed to support the practices defined in the Inception framework, called "Agile Dojo".
Resumo:
La rápida adopción de dispositivos electrónicos en el automóvil, ha contribuido a mejorar en gran medida la seguridad y el confort. Desde principios del siglo 20, la investigación en sistemas de seguridad activa ha originado el desarrollo de tecnologías como ABS (Antilock Brake System), TCS (Traction Control System) y ESP (Electronic Stability Program). El coste de despliegue de estos sistemas es crítico: históricamente, sólo han sido ampliamente adoptados cuando el precio de los sensores y la electrónica necesarios para su construcción ha caído hasta un valor marginal. Hoy en día, los vehículos a motor incluyen un amplio rango de sensores para implementar las funciones de seguridad. La incorporación de sistemas que detecten la presencia de agua, hielo o nieve en la vía es un factor adicional que podría ayudar a evitar situaciones de riesgo. Existen algunas implementaciones prácticas capaces de detectar carreteras mojadas, heladas y nevadas, aunque con limitaciones importantes. En esta tesis doctoral, se propone una aproximación novedosa al problema, basada en el análisis del ruido de rodadura generado durante la conducción. El ruido de rodadura es capturado y preprocesado. Después es analizado utilizando un clasificador basado en máquinas de vectores soporte (SVM), con el fin de generar una estimación del estado del firme. Todas estas operaciones se realizan en el propio vehículo. El sistema propuesto se ha desarrollado y evaluado utilizando Matlabr, mostrando tasas de aciertos de más del 90%. Se ha realizado una implementación en tiempo real, utilizando un prototipo basado en DSP. Después se han introducido varias optimizaciones para permitir que el sistema sea realizable usando un microcontrolador de propósito general. Finalmente se ha realizado una implementación hardware basada en un microcontrolador, integrándola estrechamente con las ECU del vehículo, pudiendo obtener datos capturados por los sensores del mismo y enviar las estimaciones del estado del firme. El sistema resultante ha sido patentado, y destaca por su elevada tasa de aciertos con un tamaño, consumo y coste reducidos. ABSTRACT Proliferation of automotive electronics, has greatly improved driving safety and comfort. Since the beginning of the 20th century, investigation in active safety systems has resulted in the development of technologies such as ABS (Antilock Brake System), TCS (Traction Control System) and ESP (Electronic Stability Program). Deployment cost of these systems is critical: historically, they have been widely adopted only when the price of the sensors and electronics needed to build them has been cut to a marginal value. Nowadays, motor vehicles include a wide range of sensors to implement the safety functions. Incorporation of systems capable of detecting water, ice or snow on the road is an additional factor that could help avoiding risky situations. There are some implementations capable of detecting wet, icy and snowy roads, although with important limitations. In this PhD Thesis, a novel approach is proposed, based on the analysis of the tyre/road noise radiated during driving. Tyre/road noise is captured and pre-processed. Then it is analysed using a Support Vector Machine (SVM) based classifier, to output an estimation of the road status. All these operations are performed on-board. Proposed system is developed and evaluated using Matlabr, showing success rates greater than 90%. A real time implementation is carried out using a DSP based prototype. Several optimizations are introduced enabling the system to work using a low-cost general purpose microcontroller. Finally a microcontroller based hardware implementation is developed. This implementation is tightly integrated with the vehicle ECUs, allowing it to obtain data captured by its sensors, and to send the road status estimations. Resulting system has been patented, and is notable because of its high hit rate, small size, low power consumption and low cost.
Resumo:
This work proposes an optimization of a semi-supervised Change Detection methodology based on a combination of Change Indices (CI) derived from an image multitemporal data set. For this purpose, SPOT 5 Panchromatic images with 2.5 m spatial resolution have been used, from which three Change Indices have been calculated. Two of them are usually known indices; however the third one has been derived considering the Kullbak-Leibler divergence. Then, these three indices have been combined forming a multiband image that has been used in as input for a Support Vector Machine (SVM) classifier where four different discriminant functions have been tested in order to differentiate between change and no_change categories. The performance of the suggested procedure has been assessed applying different quality measures, reaching in each case highly satisfactory values. These results have demonstrated that the simultaneous combination of basic change indices with others more sophisticated like the Kullback-Leibler distance, and the application of non-parametric discriminant functions like those employees in the SVM method, allows solving efficiently a change detection problem.
Resumo:
The application of the Electro-Mechanical Impedance (EMI) method for damage detection in Structural Health Monitoring has noticeable increased in recent years. EMI method utilizes piezoelectric transducers for directly measuring the mechanical properties of the host structure, obtaining the so called impedance measurement, highly influenced by the variations of dynamic parameters of the structure. These measurements usually contain a large number of frequency points, as well as a high number of dimensions, since each frequency range swept can be considered as an independent variable. That makes this kind of data hard to handle, increasing the computational costs and being substantially time-consuming. In that sense, the Principal Component Analysis (PCA)-based data compression has been employed in this work, in order to enhance the analysis capability of the raw data. Furthermore, a Support Vector Machine (SVM), which has been widespread used in machine learning and pattern recognition fields, has been applied in this study in order to model any possible existing pattern in the PCAcompress data, using for that just the first two Principal Components. Different known non-damaged and damaged measurements of an experimental tested beam were used as training input data for the SVM algorithm, using as test input data the same amount of cases measured in beams with unknown structural health conditions. Thus, the purpose of this work is to demonstrate how, with a few impedance measurements of a beam as raw data, its healthy status can be determined based on pattern recognition procedures.
Resumo:
A depth-based face recognition algorithm specially adapted to high range resolution data acquired by the new Microsoft Kinect 2 sensor is presented. A novel descriptor called Depth Local Quantized Pattern descriptor has been designed to make use of the extended range resolution of the new sensor. This descriptor is a substantial modification of the popular Local Binary Pattern algorithm. One of the main contributions is the introduction of a quantification step, increasing its capacity to distinguish different depth patterns. The proposed descriptor has been used to train and test a Support Vector Machine classifier, which has proven to be able to accurately recognize different people faces from a wide range of poses. In addition, a new depth-based face database acquired by the new Kinect 2 sensor have been created and made public to evaluate the proposed face recognition system.