246 resultados para Binary image


Relevância:

70.00% 70.00%

Publicador:

Resumo:

Template matching is concerned with measuring the similarity between patterns of two objects. This paper proposes a memory-based reasoning approach for pattern recognition of binary images with a large template set. It seems that memory-based reasoning intrinsically requires a large database. Moreover, some binary image recognition problems inherently need large template sets, such as the recognition of Chinese characters which needs thousands of templates. The proposed algorithm is based on the Connection Machine, which is the most massively parallel machine to date, using a multiresolution method to search for the matching template. The approach uses the pyramid data structure for the multiresolution representation of templates and the input image pattern. For a given binary image it scans the template pyramid searching the match. A binary image of N × N pixels can be matched in O(log N) time complexity by our algorithm and is independent of the number of templates. Implementation of the proposed scheme is described in detail.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper presents recursive algorithms for fast computation of Legendre and Zernike moments of a grey-level image intensity distribution. For a binary image, a contour integration method is developed for the evaluation of Legendre moments using only the boundary information. A method for recursive calculation of Zernike polynomial coefficients is also given. A square-to-circular image transformation scheme is introduced to minimize the computation involved in Zernike moment functions. The recursive formulae can also be used in inverse moment transforms to reconstruct the original image from moments. The mathematical framework of the algorithms is given in detail, and illustrated with binary and grey-level images.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Quantization formats of four digital holographic codes (Lohmann,Lee, Burckhardt and Hsueh-Sawchuk) are evaluated. A quantitative assessment is made from errors in both the Fourier transform and image domains. In general, small errors in the Fourier amplitude or phase alone do not guarantee high image fidelity. From quantization considerations, the Lee hologram is shown to be the best choice for randomly phase coded objects. When phase coding is not feasible, the Lohmann hologram is preferable as it is easier to plot.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The low-frequency (5–100 kHz) dielectric constant epsilon (Porson) has been measured in the temperature range 7 × 10−5 < t = (T − Tc)/Tc < 8 × 10−2. Near Tc an exponent ≈0.11 characterizes the power law behaviour of Image consistent with the theoretically predicted t−α singularity. However, over the full range of t an exponent ≈0.35 is obtained.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The signal-to-noise (S/N) ratio in the reconstructed image from a binary hologram has been quantitatively related to the amplitude and phase quantization levels. The S/N ratio increases monotonically with increasing number of quantization levels. This observation is further supported by experimental results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper the approach for automatic road extraction for an urban region using structural, spectral and geometric characteristics of roads has been presented. Roads have been extracted based on two levels: Pre-processing and road extraction methods. Initially, the image is pre-processed to improve the tolerance by reducing the clutter (that mostly represents the buildings, parking lots, vegetation regions and other open spaces). The road segments are then extracted using Texture Progressive Analysis (TPA) and Normalized cut algorithm. The TPA technique uses binary segmentation based on three levels of texture statistical evaluation to extract road segments where as, Normalizedcut method for road extraction is a graph based method that generates optimal partition of road segments. The performance evaluation (quality measures) for road extraction using TPA and normalized cut method is compared. Thus the experimental result show that normalized cut method is efficient in extracting road segments in urban region from high resolution satellite image.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we present a machine learning approach to measure the visual quality of JPEG-coded images. The features for predicting the perceived image quality are extracted by considering key human visual sensitivity (HVS) factors such as edge amplitude, edge length, background activity and background luminance. Image quality assessment involves estimating the functional relationship between HVS features and subjective test scores. The quality of the compressed images are obtained without referring to their original images ('No Reference' metric). Here, the problem of quality estimation is transformed to a classification problem and solved using extreme learning machine (ELM) algorithm. In ELM, the input weights and the bias values are randomly chosen and the output weights are analytically calculated. The generalization performance of the ELM algorithm for classification problems with imbalance in the number of samples per quality class depends critically on the input weights and the bias values. Hence, we propose two schemes, namely the k-fold selection scheme (KS-ELM) and the real-coded genetic algorithm (RCGA-ELM) to select the input weights and the bias values such that the generalization performance of the classifier is a maximum. Results indicate that the proposed schemes significantly improve the performance of ELM classifier under imbalance condition for image quality assessment. The experimental results prove that the estimated visual quality of the proposed RCGA-ELM emulates the mean opinion score very well. The experimental results are compared with the existing JPEG no-reference image quality metric and full-reference structural similarity image quality metric.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

There are essentially two different phenomenological models available to describe the interdiffusion process in binary systems in the olid state. The first of these, which is used more frequently, is based on the theory of flux partitioning. The second model, developed much more recently, uses the theory of dissociation and reaction. Although the theory of flux partitioning has been widely used, we found that this theory does not account for the mobility of both species and therefore is not suitable for use in most interdiffusion systems. We have first modified this theory to take into account the mobility of both species and then further extended it to develop relations or the integrated diffusion coefficient and the ratio of diffusivities of the species. The versatility of these two different models is examined in the Co-Si system with respect to different end-member compositions. From our analysis, we found that the applicability of the theory of flux partitioning is rather limited but the theory of dissociation and reaction can be used in any binary system.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Remote sensing provides a lucid and effective means for crop coverage identification. Crop coverage identification is a very important technique, as it provides vital information on the type and extent of crop cultivated in a particular area. This information has immense potential in the planning for further cultivation activities and for optimal usage of the available fertile land. As the frontiers of space technology advance, the knowledge derived from the satellite data has also grown in sophistication. Further, image classification forms the core of the solution to the crop coverage identification problem. No single classifier can prove to satisfactorily classify all the basic crop cover mapping problems of a cultivated region. We present in this paper the experimental results of multiple classification techniques for the problem of crop cover mapping of a cultivated region. A detailed comparison of the algorithms inspired by social behaviour of insects and conventional statistical method for crop classification is presented in this paper. These include the Maximum Likelihood Classifier (MLC), Particle Swarm Optimisation (PSO) and Ant Colony Optimisation (ACO) techniques. The high resolution satellite image has been used for the experiments.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The LISA Parameter Estimation Taskforce was formed in September 2007 to provide the LISA Project with vetted codes, source distribution models and results related to parameter estimation. The Taskforce's goal is to be able to quickly calculate the impact of any mission design changes on LISA's science capabilities, based on reasonable estimates of the distribution of astrophysical sources in the universe. This paper describes our Taskforce's work on massive black-hole binaries (MBHBs). Given present uncertainties in the formation history of MBHBs, we adopt four different population models, based on (i) whether the initial black-hole seeds are small or large and (ii) whether accretion is efficient or inefficient at spinning up the holes. We compare four largely independent codes for calculating LISA's parameter-estimation capabilities. All codes are based on the Fisher-matrix approximation, but in the past they used somewhat different signal models, source parametrizations and noise curves. We show that once these differences are removed, the four codes give results in extremely close agreement with each other. Using a code that includes both spin precession and higher harmonics in the gravitational-wave signal, we carry out Monte Carlo simulations and determine the number of events that can be detected and accurately localized in our four population models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The presence of folded solution conformations in the peptides Boc-Ala-(Aib-Ala)2-OMe, Boc-Val-(Aib-Val) 2-OMe, Boc-Ala-(Aib-Ala)3-OMe and Boc-Val-(Aib-Val)3-OMe has been established by 270MHz 1H NMR. Intramolecularly H-bonded NH groups have been identified using temperature and solvent dependence of NH chemical shifts and paramagnetic radical induced broadening of NH resonances. Both pentapeptides adopt 310 helical conformations possessing 3 intramolecular H-bonds in CDCl3 and (CD3)2SO. The heptapeptides favour helical structures with 5 H-bonds in CDCl3. In (CD3)2SO only 4 H-bonds are readily detected.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Measurements of the ratio of diffusion coefficient to mobility (D/ mu ) of electrons in SF6-N2 and CCl2F2-N2 mixtures over the range 80

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Experimental results are presented of ionisation (a)a nd electron attachment ( v ) coefficients evaluated from the steady-state Townsend curregnrto wth curves for SFsN2 and CC12FrN2 mixtures over the range 60 S E/P 6 240 (where E is the electric field in V cm" and P is the pressure in Torr reduced to 20'C). In both the mixtures the attachment coefficients (vmu) evaluated were found to follow the relationship; where 7 is the attachment coefficient of pure electronegative gas, F is the fraction of the electronegative gas in the mixture and /3 is a constant. The ionisation coefficients (amlx) generally obeyed the relationship where w2a nd aAa re thei onisation coefficients of nitrogen and the attachinggraess pectively. However, in case of CC12FrN2 mixtures, there were maxima in the a,,,v,a,l ues for CCI2F2 concentrations varying between 10% and 30% at all values of E/P investigated. Effective ionisation coefficients (a - p)/P obtained in these binary mixtures show that the critical E/P (corresponding to (a - q)/P = 0) increases with increase in the concentration of the electronegative gas up to 40%. Further increase in the electronegative gas content does not seem to alter the critical E/P.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Lateral or transaxial truncation of cone-beam data can occur either due to the field of view limitation of the scanning apparatus or iregion-of-interest tomography. In this paper, we Suggest two new methods to handle lateral truncation in helical scan CT. It is seen that reconstruction with laterally truncated projection data, assuming it to be complete, gives severe artifacts which even penetrates into the field of view. A row-by-row data completion approach using linear prediction is introduced for helical scan truncated data. An extension of this technique known as windowed linear prediction approach is introduced. Efficacy of the two techniques are shown using simulation with standard phantoms. A quantitative image quality measure of the resulting reconstructed images are used to evaluate the performance of the proposed methods against an extension of a standard existing technique.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A novel method, designated the holographic spectrum reconstruction (HSR) method, is proposed for achieving simultaneous display of the spectrum and image of an object in a single plane. A study of the scaling behaviour of both the spectrum and the image has been carried out and based on this study, it is demonstrated that a lensless coherent optical processor can be realized.