835 resultados para Image-based detector
Resumo:
Successful classification, information retrieval and image analysis tools are intimately related with the quality of the features employed in the process. Pixel intensities, color, texture and shape are, generally, the basis from which most of the features are Computed and used in such fields. This papers presents a novel shape-based feature extraction approach where an image is decomposed into multiple contours, and further characterized by Fourier descriptors. Unlike traditional approaches we make use of topological knowledge to generate well-defined closed contours, which are efficient signatures for image retrieval. The method has been evaluated in the CBIR context and image analysis. The results have shown that the multi-contour decomposition, as opposed to a single shape information, introduced a significant improvement in the discrimination power. (c) 2008 Elsevier B.V. All rights reserved,
Resumo:
Traditional content-based image retrieval (CBIR) systems use low-level features such as colors, shapes, and textures of images. Although, users make queries based on semantics, which are not easily related to such low-level characteristics. Recent works on CBIR confirm that researchers have been trying to map visual low-level characteristics and high-level semantics. The relation between low-level characteristics and image textual information has motivated this article which proposes a model for automatic classification and categorization of words associated to images. This proposal considers a self-organizing neural network architecture, which classifies textual information without previous learning. Experimental results compare the performance results of the text-based approach to an image retrieval system based on low-level features. (c) 2008 Wiley Periodicals, Inc.
Resumo:
This paper presents an automatic method to detect and classify weathered aggregates by assessing changes of colors and textures. The method allows the extraction of aggregate features from images and the automatic classification of them based on surface characteristics. The concept of entropy is used to extract features from digital images. An analysis of the use of this concept is presented and two classification approaches, based on neural networks architectures, are proposed. The classification performance of the proposed approaches is compared to the results obtained by other algorithms (commonly considered for classification purposes). The obtained results confirm that the presented method strongly supports the detection of weathered aggregates.
Resumo:
A method is developed to search for air showers initiated by photons using data recorded by the surface detector of the Auger Observatory. The approach is based on observables sensitive to the longitudinal shower development, the signal risetime and the curvature of the shower front. Applying this method to the data, tipper limits on the flux of photons of 3.8 x 10(-3), 2.5 x 10(-3), and 2.2 x 10(-3) km(-2) sr(-1) yr(-1) above 10(19) eV, 2 x 10(19) eV, and 4 x 10(19) eV are derived, with corresponding limits on the fraction of photons being 2.0%, 5.1%, and 31% (all limits at 95% c.l.). These photon limits disfavor certain exotic models of sources of cosmic rays. The results also show that the approach adopted by the Auger Observatory to calibrate the shower energy is not strongly biased by a contamination from photons. (C) 2008 Elsevier B.V. All rights reserved.
Resumo:
The protective shielding design of a mammography facility requires the knowledge of the scattered radiation by the patient and image receptor components. The shape and intensity of secondary x-ray beams depend on the kVp applied to the x-ray tube, target/filter combination, primary x-ray field size, and scattering angle. Currently, shielding calculations for mammography facilities are performed based on scatter fraction data for Mo/Mo target/filter, even though modern mammography equipment is designed with different anode/filter combinations. In this work we present scatter fraction data evaluated based on the x-ray spectra produced by a Mo/Mo, Mo/Rh and W/Rh target/filter, for 25, 30 and 35 kV tube voltages and scattering angles between 30 and 165 degrees. Three mammography phantoms were irradiated and the scattered radiation was measured with a CdZnTe detector. The primary x-ray spectra were computed with a semiempirical model based on the air kerma and HVL measured with an ionization chamber. The results point out that the scatter fraction values are higher for W/Rh than for Mo/Mo and Mo/Rh, although the primary and scattered air kerma are lower for W/Rh than for Mo/Mo and Mo/Rh target/filter combinations. The scatter fractions computed in this work were applied in a shielding design calculation in order to evaluate shielding requirements for each of these target/filter combinations. Besides, shielding requirements have been evaluated converting the scattered air kerma from mGy/week to mSv/week adopting initially a conversion coefficient from air kerma to effective dose as 1 Sv/Gy and then a mean conversion coefficient specific for the x-ray beam considered. Results show that the thickest barrier should be provided for Mo/Mo target/filter combination. They also point out that the use of the conversion coefficient from air kerma to effective dose as 1 Sv/Gy is conservatively high in the mammography energy range and overestimate the barrier thickness. (c) 2008 American Association of Physicists in Medicine.
Resumo:
Texture is an important visual attribute used to describe the pixel organization in an image. As well as it being easily identified by humans, its analysis process demands a high level of sophistication and computer complexity. This paper presents a novel approach for texture analysis, based on analyzing the complexity of the surface generated from a texture, in order to describe and characterize it. The proposed method produces a texture signature which is able to efficiently characterize different texture classes. The paper also illustrates a novel method performance on an experiment using texture images of leaves. Leaf identification is a difficult and complex task due to the nature of plants, which presents a huge pattern variation. The high classification rate yielded shows the potential of the method, improving on traditional texture techniques, such as Gabor filters and Fourier analysis.
Resumo:
This paper introduces a novel methodology to shape boundary characterization, where a shape is modeled into a small-world complex network. It uses degree and joint degree measurements in a dynamic evolution network to compose a set of shape descriptors. The proposed shape characterization method has all efficient power of shape characterization, it is robust, noise tolerant, scale invariant and rotation invariant. A leaf plant classification experiment is presented on three image databases in order to evaluate the method and compare it with other descriptors in the literature (Fourier descriptors, Curvature, Zernike moments and multiscale fractal dimension). (C) 2008 Elsevier Ltd. All rights reserved.
Resumo:
Recently, the deterministic tourist walk has emerged as a novel approach for texture analysis. This method employs a traveler visiting image pixels using a deterministic walk rule. Resulting trajectories provide clues about pixel interaction in the image that can be used for image classification and identification tasks. This paper proposes a new walk rule for the tourist which is based on contrast direction of a neighborhood. The yielded results using this approach are comparable with those from traditional texture analysis methods in the classification of a set of Brodatz textures and their rotated versions, thus confirming the potential of the method as a feasible texture analysis methodology. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
This paper describes a visual stimulus generator (VSImG) capable of displaying a gray-scale, 256 x 256 x 8 bitmap image with a frame rate of 500 Hz using a boustrophedonic scanning technique. It is designed for experiments with motion-sensitive neurons of the fly`s visual system, where the flicker fusion frequency of the photoreceptors can reach up to 500 Hz. Devices with such a high frame rate are not commercially available, but are required, if sensory systems with high flicker fusion frequency are to be studied. The implemented hardware approach gives us complete real-time control of the displacement sequence and provides all the signals needed to drive an electrostatic deflection display. With the use of analog signals, very small high-resolution displacements, not limited by the image`s pixel size can be obtained. Very slow image displacements with visually imperceptible steps can also be generated. This can be of interest for other vision research experiments. Two different stimulus files can be used simultaneously, allowing the system to generate X-Y displacements on one display or independent movements on two displays as long as they share the same bitmap image. (C) 2011 Elsevier B.V. All rights reserved.
Resumo:
The design of translation invariant and locally defined binary image operators over large windows is made difficult by decreased statistical precision and increased training time. We present a complete framework for the application of stacked design, a recently proposed technique to create two-stage operators that circumvents that difficulty. We propose a novel algorithm, based on Information Theory, to find groups of pixels that should be used together to predict the Output Value. We employ this algorithm to automate the process of creating a set of first-level operators that are later combined in a global operator. We also propose a principled way to guide this combination, by using feature selection and model comparison. Experimental results Show that the proposed framework leads to better results than single stage design. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
An all-in-one version of a capacitively coupled contactless conductivity detector is introduced. The absence of moving parts (potentiometers and connectors) makes it compact (6.5 cm(3)) and robust. A local oscillator, working at 1.1 MHz, was optimized to use capillaries of id from 20 to 100 lam. Low noise circuitry and a high-resolution analog-to-digital converter (ADC) (21 bits effective) grant good sensitivities for capillaries and background electrolytes currently used in capillary electrophoresis. The fixed frequency and amplitude of the signal generator is a drawback that is compensated by the steady calibration curves for conductivity. Another advantage is the possibility of determining the inner diameter of a capillary by reading the ADC when air and subsequently water flow through the capillary. The difference of ADC reading may be converted into the inner diameter by a calibration curve. This feature is granted by the 21-bit ADC, which eliminates the necessity of baseline compensation by hardware. In a typical application, the limits of detection based on the 3 sigma criterion (without baseline filtering) were 0.6, 0.4, 0.3, 0.5, 0.6, and 0.8 mu mol/L for K(+), Ba(2+), Ca(2+), Na(+), Mg(2+), and Li(+), respectively, which is comparable to other high-quality implementations of a capacitively coupled contactless conductivity detector.
Resumo:
The enantiomers of sulfoxide proton pump inhibitors - omeprazole, lansoprazole, rabeprazole and Ro 18-5364 - were enantiomerically separated by liquid chromatography at multimilligram scale on a poly saccharide-based chiral stationary phase using normal and polar organic conditions as mobile phase. The values of the recovery and production rate were significant for each enantiomer; better results were achieved using a solid-phase injection system. However, this system was applied just for the enantionteric separation of omeprazole to demonstrate the applicability of this injection mode at milligram scale. The chiroptical characterization of the compounds was performed using a polarimeter and a circular dichroism detector. The higher enantiomeric purity obtained for the isolated enantiomers suggests that the methods here described should be considered as a simple and rapid way to obtain enantiomeric pure standards for analytical purpose. (C) 2007 Elsevier B.V. All rights reserved.
Resumo:
Since last two decades researches have been working on developing systems that can assistsdrivers in the best way possible and make driving safe. Computer vision has played a crucialpart in design of these systems. With the introduction of vision techniques variousautonomous and robust real-time traffic automation systems have been designed such asTraffic monitoring, Traffic related parameter estimation and intelligent vehicles. Among theseautomatic detection and recognition of road signs has became an interesting research topic.The system can assist drivers about signs they don’t recognize before passing them.Aim of this research project is to present an Intelligent Road Sign Recognition System basedon state-of-the-art technique, the Support Vector Machine. The project is an extension to thework done at ITS research Platform at Dalarna University [25]. Focus of this research work ison the recognition of road signs under analysis. When classifying an image its location, sizeand orientation in the image plane are its irrelevant features and one way to get rid of thisambiguity is to extract those features which are invariant under the above mentionedtransformation. These invariant features are then used in Support Vector Machine forclassification. Support Vector Machine is a supervised learning machine that solves problemin higher dimension with the help of Kernel functions and is best know for classificationproblems.
Resumo:
This report presents an algorithm for locating the cut points for and separatingvertically attached traffic signs in Sweden. This algorithm provides severaladvanced digital image processing features: binary image which representsvisual object and its complex rectangle background with number one and zerorespectively, improved cross correlation which shows the similarity of 2Dobjects and filters traffic sign candidates, simplified shape decompositionwhich smoothes contour of visual object iteratively in order to reduce whitenoises, flipping point detection which locates black noises candidates, chasmfilling algorithm which eliminates black noises, determines the final cut pointsand separates originally attached traffic signs into individual ones. At each step,the mediate results as well as the efficiency in practice would be presented toshow the advantages and disadvantages of the developed algorithm. Thisreport concentrates on contour-based recognition of Swedish traffic signs. Thegeneral shapes cover upward triangle, downward triangle, circle, rectangle andoctagon. At last, a demonstration program would be presented to show howthe algorithm works in real-time environment.
Resumo:
http://digitalcommons.colby.edu/atlasofmaine2009/1028/thumbnail.jpg