163 resultados para pre-image attack


Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes a denoising algorithm which performs non-local means bilateral filtering. As existing literature suggests, non-local means (NLM) is one of the widely used denoising techniques, but has a critical drawback of smoothing of edges. In order to improve this, we perform fast and efficient NLM using Approximate Nearest Neighbour Fields and improve the edge content in denoising by formulating a joint-bilateral filter. Using the proposed joint bilateral, we are able to denoise smooth regions using the NLM approach and efficient edge reconstruction is obtained from the bilateral filter. Furthermore, to avoid tedious parameter selection, we carry out a noise estimation before performing joint bilateral filtering. The proposed approach is observed to perform well on high noise images.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Image inpainting is the process of filling the unwanted region in an image marked by the user. It is used for restoring old paintings and photographs, removal of red eyes from pictures, etc. In this paper, we propose an efficient inpainting algorithm which takes care of false edge propagation. We use the classical exemplar based technique to find out the priority term for each patch. To ensure that the edge content of the nearest neighbor patch found by minimizing L-2 distance between patches, we impose an additional constraint that the entropy of the patches be similar. Entropy of the patch acts as a good measure of edge content. Additionally, we fill the image by considering overlapping patches to ensure smoothness in the output. We use structural similarity index as the measure of similarity between ground truth and inpainted image. The results of the proposed approach on a number of examples on real and synthetic images show the effectiveness of our algorithm in removing objects and thin scratches or text written on image. It is also shown that the proposed approach is robust to the shape of the manually selected target. Our results compare favorably to those obtained by existing techniques

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The discovery of microRNAs (miRNAs) has added a new dimension to the gene regulatory networks, making aberrantly expressed miRNAs as therapeutically important targets. Small molecules that can selectively target and modulate miRNA levels can thus serve as lead structures. Cationic cyclic peptides containing sugar amino acids represent a new class of small molecules that can target miRNA selectively. Upon treatment of these small molecules in breast cancer cell line, we profiled 96 therapeutically important miRNAs associated with cancer and observed that these peptides can selectively target paralogous miRNAs of the same seed family. This selective inhibition is of prime significance in cases when miRNAs of the same family have tissue-specific expression and perform different functions. During these conditions, targeting an entire miRNA family could lead to undesired adverse effects. The selective targeting is attributable to the difference in the three-dimensional structures of precursor miRNAs. Hence, the core structure of these peptides can be used as a scaffold for designing more potent inhibitors of miRNA maturation and hence function.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Using in situ, high-speed imaging of a hard wedge sliding against pure aluminum, and image analysis by particle image velocimetry, the deformation field in sliding is mapped at high resolution. This model system is representative of asperity contacts on engineered surfaces and die-workpiece contacts in deformation and machining processes. It is shown that large, uniform plastic strains of 1-5 can be imposed at the Al surface, up to depths of 500 mu m, under suitable sliding conditions. The spatial strain and strain rate distributions are significantly influenced by the initial deformation state of the Al, e.g., extent of work hardening, and sliding incidence angle. Uniform straining occurs only under conditions of steady laminar flow in the metal. Large pre-strains and higher sliding angles promote breakdown in laminar flow due to surface fold formation or flow localization in the form of shear bands, thus imposing limits on uniform straining by sliding. Avoidance of unsteady sliding conditions, and selection of parameters like sliding angle, thus provides a way to control the deformation field. Key characteristics of the sliding deformation such as strain and strain rate, laminar flow, folding and prow formation are well predicted by finite element simulation. The deformation field provides a quantitative basis for interpreting wear particle formation. Implications for engineering functionally graded surfaces, sliding wear and ductile failure in metals are discussed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Purpose: A prior image based temporally constrained reconstruction ( PITCR) algorithm was developed for obtaining accurate temperature maps having better volume coverage, and spatial, and temporal resolution than other algorithms for highly undersampled data in magnetic resonance (MR) thermometry. Methods: The proposed PITCR approach is an algorithm that gives weight to the prior image and performs accurate reconstruction in a dynamic imaging environment. The PITCR method is compared with the temporally constrained reconstruction (TCR) algorithm using pork muscle data. Results: The PITCR method provides superior performance compared to the TCR approach with highly undersampled data. The proposed approach is computationally expensive compared to the TCR approach, but this could be overcome by the advantage of reconstructing with fewer measurements. In the case of reconstruction of temperature maps from 16% of fully sampled data, the PITCR approach was 1.57x slower compared to the TCR approach, while the root mean square error using PITCR is 0.784 compared to 2.815 with the TCR scheme. Conclusions: The PITCR approach is able to perform more accurate reconstructions of temperature maps compared to the TCR approach with highly undersampled data in MR guided high intensity focused ultrasound. (C) 2015 American Association of Physicists in Medicine.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Soil shrinkage curve represents a decrease of total porosity or an increase of bulk density with water loss. However, our knowledge of the dynamics of pores and their geometry during soil shrinkage is scarce, partially due to lack of reliable methods for determining soil pores in relation to change in soil water. This study aimed to investigate the dynamics of macropores (>30 mu m) of paddy soils during shrinkage. Two, paddy soils, which were sampled from one paddy field cultivated for 20 years (YPF) and the other one for over 100 years (OPF), represented difference in crack geometry in the field. Macropore parameters (volume, connectivity, and orientation of pores) and soil shrinkage parameters were determined on the same undisturbed soil cores by X-ray microtomography and shrinkage curve, respectively. Macroporosity was on average four times larger in the YPF than in the OPF whereas the shrinkage capacity was lower in the YPF as compared to the OPF (0.09 vs. 0.15 COLE). Soil shrinkage increased the volume of pores by 3.7% in the YPF and by 1.6% in the OPF as well as their connectivity. The formation of macropores occurred mostly in the proportional shrinkage phase. As a result, the slope of the proportional shrinkage phase was smaller in the YPF (0.65) than in the OPF (0.89). New macropores were cracks and extended pre-existing pores in the range of 225-1215 pm size without any preferential orientation. This work provides image evidences that in paddy soils with high shrinkage capacity more macropores are generated in the soil presenting a smaller proportional shrinkage slope. (C) 2015 Elsevier B.V. All rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents the design and implementation of PolyMage, a domain-specific language and compiler for image processing pipelines. An image processing pipeline can be viewed as a graph of interconnected stages which process images successively. Each stage typically performs one of point-wise, stencil, reduction or data-dependent operations on image pixels. Individual stages in a pipeline typically exhibit abundant data parallelism that can be exploited with relative ease. However, the stages also require high memory bandwidth preventing effective utilization of parallelism available on modern architectures. For applications that demand high performance, the traditional options are to use optimized libraries like OpenCV or to optimize manually. While using libraries precludes optimization across library routines, manual optimization accounting for both parallelism and locality is very tedious. The focus of our system, PolyMage, is on automatically generating high-performance implementations of image processing pipelines expressed in a high-level declarative language. Our optimization approach primarily relies on the transformation and code generation capabilities of the polyhedral compiler framework. To the best of our knowledge, this is the first model-driven compiler for image processing pipelines that performs complex fusion, tiling, and storage optimization automatically. Experimental results on a modern multicore system show that the performance achieved by our automatic approach is up to 1.81x better than that achieved through manual tuning in Halide, a state-of-the-art language and compiler for image processing pipelines. For a camera raw image processing pipeline, our performance is comparable to that of a hand-tuned implementation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Inaccuracies in prediction of circulating viral strain genotypes and the possibility of novel reassortants causing a pandemic outbreak necessitate the development of an anti-influenza vaccine with increased breadth of protection and potential for rapid production and deployment. The hemagglutinin (HA) stem is a promising target for universal influenza vaccine as stem-specific antibodies have the potential to be broadly cross-reactive towards different HA subtypes. Here, we report the design of a bacterially expressed polypeptide that mimics a H5 HA stem by protein minimization to focus the antibody response towards the HA stem. The HA mini-stem folds as a trimer mimicking the HA prefusion conformation. It is resistant to thermal/chemical stress, and it binds to conformation-specific, HA stem-directed broadly neutralizing antibodies with high affinity. Mice vaccinated with the group 1 HA mini-stems are protected from morbidity and mortality against lethal challenge by both group 1 (H5 and H1) and group 2 (H3) influenza viruses, the first report of cross-group protection. Passive transfer of immune serum demonstrates the protection is mediated by stem-specific antibodies. Furthermore, antibodies indudced by these HA stems have broad HA reactivity, yet they do not have antibody-dependent enhancement activity.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Imaging flow cytometry is an emerging technology that combines the statistical power of flow cytometry with spatial and quantitative morphology of digital microscopy. It allows high-throughput imaging of cells with good spatial resolution, while they are in flow. This paper proposes a general framework for the processing/classification of cells imaged using imaging flow cytometer. Each cell is localized by finding an accurate cell contour. Then, features reflecting cell size, circularity and complexity are extracted for the classification using SVM. Unlike the conventional iterative, semi-automatic segmentation algorithms such as active contour, we propose a noniterative, fully automatic graph-based cell localization. In order to evaluate the performance of the proposed framework, we have successfully classified unstained label-free leukaemia cell-lines MOLT, K562 and HL60 from video streams captured using custom fabricated cost-effective microfluidics-based imaging flow cytometer. The proposed system is a significant development in the direction of building a cost-effective cell analysis platform that would facilitate affordable mass screening camps looking cellular morphology for disease diagnosis. Lay description In this article, we propose a novel framework for processing the raw data generated using microfluidics based imaging flow cytometers. Microfluidics microscopy or microfluidics based imaging flow cytometry (mIFC) is a recent microscopy paradigm, that combines the statistical power of flow cytometry with spatial and quantitative morphology of digital microscopy, which allows us imaging cells while they are in flow. In comparison to the conventional slide-based imaging systems, mIFC is a nascent technology enabling high throughput imaging of cells and is yet to take the form of a clinical diagnostic tool. The proposed framework process the raw data generated by the mIFC systems. The framework incorporates several steps: beginning from pre-processing of the raw video frames to enhance the contents of the cell, localising the cell by a novel, fully automatic, non-iterative graph based algorithm, extraction of different quantitative morphological parameters and subsequent classification of cells. In order to evaluate the performance of the proposed framework, we have successfully classified unstained label-free leukaemia cell-lines MOLT, K562 and HL60 from video streams captured using cost-effective microfluidics based imaging flow cytometer. The cell lines of HL60, K562 and MOLT were obtained from ATCC (American Type Culture Collection) and are separately cultured in the lab. Thus, each culture contains cells from its own category alone and thereby provides the ground truth. Each cell is localised by finding a closed cell contour by defining a directed, weighted graph from the Canny edge images of the cell such that the closed contour lies along the shortest weighted path surrounding the centroid of the cell from a starting point on a good curve segment to an immediate endpoint. Once the cell is localised, morphological features reflecting size, shape and complexity of the cells are extracted and used to develop a support vector machine based classification system. We could classify the cell-lines with good accuracy and the results were quite consistent across different cross validation experiments. We hope that imaging flow cytometers equipped with the proposed framework for image processing would enable cost-effective, automated and reliable disease screening in over-loaded facilities, which cannot afford to hire skilled personnel in large numbers. Such platforms would potentially facilitate screening camps in low income group countries; thereby transforming the current health care paradigms by enabling rapid, automated diagnosis for diseases like cancer.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Among the multiple advantages and applications of remote sensing, one of the most important uses is to solve the problem of crop classification, i.e., differentiating between various crop types. Satellite images are a reliable source for investigating the temporal changes in crop cultivated areas. In this letter, we propose a novel bat algorithm (BA)-based clustering approach for solving crop type classification problems using a multispectral satellite image. The proposed partitional clustering algorithm is used to extract information in the form of optimal cluster centers from training samples. The extracted cluster centers are then validated on test samples. A real-time multispectral satellite image and one benchmark data set from the University of California, Irvine (UCI) repository are used to demonstrate the robustness of the proposed algorithm. The performance of the BA is compared with two other nature-inspired metaheuristic techniques, namely, genetic algorithm and particle swarm optimization. The performance is also compared with the existing hybrid approach such as the BA with K-means. From the results obtained, it can be concluded that the BA can be successfully applied to solve crop type classification problems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Fingerprints are used for identification in forensics and are classified into Manual and Automatic. Automatic fingerprint identification system is classified into Latent and Exemplar. A novel Exemplar technique of Fingerprint Image Verification using Dictionary Learning (FIVDL) is proposed to improve the performance of low quality fingerprints, where Dictionary learning method reduces the time complexity by using block processing instead of pixel processing. The dynamic range of an image is adjusted by using Successive Mean Quantization Transform (SMQT) technique and the frequency domain noise is reduced using spectral frequency Histogram Equalization. Then, an adaptive nonlinear dynamic range adjustment technique is utilized to determine the local spectral features on corresponding fingerprint ridge frequency and orientation. The dictionary is constructed using spatial fundamental frequency that is determined from the spectral features. These dictionaries help in removing the spurious noise present in fingerprints and reduce the time complexity by using block processing instead of pixel processing. Further, dictionaries are used to reconstruct the image for matching. The proposed FIVDL is verified on FVC database sets and Experimental result shows an improvement over the state-of-the-art techniques. (C) 2015 The Authors. Published by Elsevier B.V.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The polyhedral model provides an expressive intermediate representation that is convenient for the analysis and subsequent transformation of affine loop nests. Several heuristics exist for achieving complex program transformations in this model. However, there is also considerable scope to utilize this model to tackle the problem of automatic memory footprint optimization. In this paper, we present a new automatic storage optimization technique which can be used to achieve both intra-array as well as inter-array storage reuse with a pre-determined schedule for the computation. Our approach works by finding statement-wise storage partitioning hyper planes that partition a unified global array space so that values with overlapping live ranges are not mapped to the same partition. Our heuristic is driven by a fourfold objective function which not only minimizes the dimensionality and storage requirements of arrays required for each high-level statement, but also maximizes inter statement storage reuse. The storage mappings obtained using our heuristic can be asymptotically better than those obtained by any existing technique. We implement our technique and demonstrate its practical impact by evaluating its effectiveness on several benchmarks chosen from the domains of image processing, stencil computations, and high-performance computing.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We address the problem of denoising images corrupted by multiplicative noise. The noise is assumed to follow a Gamma distribution. Compared with additive noise distortion, the effect of multiplicative noise on the visual quality of images is quite severe. We consider the mean-square error (MSE) cost function and derive an expression for an unbiased estimate of the MSE. The resulting multiplicative noise unbiased risk estimator is referred to as MURE. The denoising operation is performed in the wavelet domain by considering the image-domain MURE. The parameters of the denoising function (typically, a shrinkage of wavelet coefficients) are optimized for by minimizing MURE. We show that MURE is accurate and close to the oracle MSE. This makes MURE-based image denoising reliable and on par with oracle-MSE-based estimates. Analogous to the other popular risk estimation approaches developed for additive, Poisson, and chi-squared noise degradations, the proposed approach does not assume any prior on the underlying noise-free image. We report denoising results for various noise levels and show that the quality of denoising obtained is on par with the oracle result and better than that obtained using some state-of-the-art denoisers.