980 resultados para image coding
Resumo:
Wave-number spectrum technique is proposed to retrieve coastal water depths by means of Synthetic Aperture Radar (SAR) image of waves. Based on the general dispersion relation of ocean waves, the wavelength changes of a surface wave over varying water depths can be derived from SAR. Approaching the analysis of SAR images of waves and using the general dispersion relation of ocean waves, this indirect technique of remote sensing bathymetry has been applied to a coastal region of Xiapu in Fujian Province, China. Results show that this technique is suitable for the coastal waters especially for the near-shore regions with variable water depths.
Resumo:
在JPEG静止图象压缩的基础上,设计了一种扩充的自适应量化器.利用人眼的视觉特征,通过分析MCU块的局部视觉活动性,以MCU活动性函数确定量化因子,并引入亮度掩盖算子调节量化参量.实验结果表明,本文所设计的自适应量化器能减少图象编码主观失真,改善图象质量,获得更好的压缩效果
Resumo:
With the development of oil/gas seismic exploration, seismic survey for fracture/porosity type reservoir is becoming more and more important. As for China, since it has over 60% store of low porosity and low permeability oil/gas reservoir, it’s more urgent to validly describe fracture/porosity type oil/gas trap and proposing the related, developed seismic technique. To achieve mapping fracture/porosity region and its development status, it demands profound understanding of seismic wave propagation discipline in complex fractured/pored media. Meanwhile, it has profound scientific significance and applied worth to study forward modeling of fracture/porosity type media and pre-stacked reverse time migration. Especially, pre-stacked reverse-time migration is the lead edge technique in the field of seismology and seismic exploration. In this paper, the author has summarized the meaning, history and the present state of numerical simulation of seismic propagation in fractured/pored media and seismic exploration of fractured/pored reservoirs. Extensive Dilatancy Anisotropy (EDA) model is selected as media object in this work. As to forward modeling, due to local limitation of solving spatial partial derivative when using finite-difference and finite-element method, the author turns to pseudo-spectral method (PSM), which is based on the global characteristic of Fourier transform to simulate three-component elastic wave-field. Artifact boundary effect reduction and simulation algorithm stability are also discussed in the work. The author has completed successfully forward modeling coding of elastic wave-field and numerical simulation of two-dimensional and three-dimensional EDA models with different symmetric axis. Seismic dynamic and kinematical properties of EDA media are analyzed from time slices and seismic records of wave propagation. As to pre-stacked reverse-time migration for elastic wave-field in fractured/pored media, based on the successful experience in forward modeling results with PSM, the author has studied pre-stacked reverse-time depth-domain migration technique using PSM of elastic wave-field in two dimensional EDA media induced by preferred fracture/pore distribution. At the same time, different image conditions will bring up what kind of migration result is detailed in this paper. The author has worded out software for pre-stacked reverse-time depth-domain migration of elastic wave-field in EDA media. After migration processing of a series of seismic shot gathers, influences to migration from different isotropic and anisotropy models are described in the paper. In summary, following creative research achievements are obtained: Realizing two-dimensional and three-dimensional elastic wave-field modeling for fractured/pored media and related software has been completed. Proposed pre-stacked reverse-time depth-domain migration technique using PSM of elastic wave-field. Through analysis of the seismic dynamic and kinematical properties of EDA media, the author made a conclusion that collection of multi-component seismic data can provide important data basis for locating and describing the fracture/pore regions and their magnitudes and the preferred directions. Pre-stacked reverse-time depth-domain migration technique has the ability to reconstruct complex geological object with steep formations and tilt fracture distribution. Neglecting seismic anisotropy induced by the preferred fracture/pore distribution, will lead to the disastrous imaging results.
Resumo:
In this paper we present some extensions to the k-means algorithm for vector quantization that permit its efficient use in image segmentation and pattern classification tasks. It is shown that by introducing state variables that correspond to certain statistics of the dynamic behavior of the algorithm, it is possible to find the representative centers fo the lower dimensional maniforlds that define the boundaries between classes, for clouds of multi-dimensional, mult-class data; this permits one, for example, to find class boundaries directly from sparse data (e.g., in image segmentation tasks) or to efficiently place centers for pattern classification (e.g., with local Gaussian classifiers). The same state variables can be used to define algorithms for determining adaptively the optimal number of centers for clouds of data with space-varying density. Some examples of the applicatin of these extensions are also given.
Resumo:
This paper consists of two major parts. First, we present the outline of a simple approach to very-low bandwidth video-conferencing system relying on an example-based hierarchical image compression scheme. In particular, we discuss the use of example images as a model, the number of required examples, faces as a class of semi-rigid objects, a hierarchical model based on decomposition into different time-scales, and the decomposition of face images into patches of interest. In the second part, we present several algorithms for image processing and animation as well as experimental evaluations. Among the original contributions of this paper is an automatic algorithm for pose estimation and normalization. We also review and compare different algorithms for finding the nearest neighbors in a database for a new input as well as a generalized algorithm for blending patches of interest in order to synthesize new images. Finally, we outline the possible integration of several algorithms to illustrate a simple model-based video-conference system.
Resumo:
The need to generate new views of a 3D object from a single real image arises in several fields, including graphics and object recognition. While the traditional approach relies on the use of 3D models, we have recently introduced techniques that are applicable under restricted conditions but simpler. The approach exploits image transformations that are specific to the relevant object class and learnable from example views of other "prototypical" objects of the same class. In this paper, we introduce such a new technique by extending the notion of linear class first proposed by Poggio and Vetter. For linear object classes it is shown that linear transformations can be learned exactly from a basis set of 2D prototypical views. We demonstrate the approach on artificial objects and then show preliminary evidence that the technique can effectively "rotate" high- resolution face images from a single 2D view.
Resumo:
This paper describes a machine vision system that classifies reflectance properties of surfaces such as metal, plastic, or paper, under unknown real-world illumination. We demonstrate performance of our algorithm for surfaces of arbitrary geometry. Reflectance estimation under arbitrary omnidirectional illumination proves highly underconstrained. Our reflectance estimation algorithm succeeds by learning relationships between surface reflectance and certain statistics computed from an observed image, which depend on statistical regularities in the spatial structure of real-world illumination. Although the algorithm assumes known geometry, its statistical nature makes it robust to inaccurate geometry estimates.
Resumo:
This is a collection of data on the construction operation and performance of the two image dissector cameras. Some of this data is useful in deciding whether certain shortcomings are significant for a given application and if so how to compensate for them.
Resumo:
We present an algorithm that uses multiple cues to recover shading and reflectance intrinsic images from a single image. Using both color information and a classifier trained to recognize gray-scale patterns, each image derivative is classified as being caused by shading or a change in the surface's reflectance. Generalized Belief Propagation is then used to propagate information from areas where the correct classification is clear to areas where it is ambiguous. We also show results on real images.
Resumo:
The goal of low-level vision is to estimate an underlying scene, given an observed image. Real-world scenes (e.g., albedos or shapes) can be very complex, conventionally requiring high dimensional representations which are hard to estimate and store. We propose a low-dimensional representation, called a scene recipe, that relies on the image itself to describe the complex scene configurations. Shape recipes are an example: these are the regression coefficients that predict the bandpassed shape from bandpassed image data. We describe the benefits of this representation, and show two uses illustrating their properties: (1) we improve stereo shape estimates by learning shape recipes at low resolution and applying them at full resolution; (2) Shape recipes implicitly contain information about lighting and materials and we use them for material segmentation.
Resumo:
Binary image classifiction is a problem that has received much attention in recent years. In this paper we evaluate a selection of popular techniques in an effort to find a feature set/ classifier combination which generalizes well to full resolution image data. We then apply that system to images at one-half through one-sixteenth resolution, and consider the corresponding error rates. In addition, we further observe generalization performance as it depends on the number of training images, and lastly, compare the system's best error rates to that of a human performing an identical classification task given teh same set of test images.
Resumo:
We present an image-based approach to infer 3D structure parameters using a probabilistic "shape+structure'' model. The 3D shape of a class of objects may be represented by sets of contours from silhouette views simultaneously observed from multiple calibrated cameras. Bayesian reconstructions of new shapes can then be estimated using a prior density constructed with a mixture model and probabilistic principal components analysis. We augment the shape model to incorporate structural features of interest; novel examples with missing structure parameters may then be reconstructed to obtain estimates of these parameters. Model matching and parameter inference are done entirely in the image domain and require no explicit 3D construction. Our shape model enables accurate estimation of structure despite segmentation errors or missing views in the input silhouettes, and works even with only a single input view. Using a dataset of thousands of pedestrian images generated from a synthetic model, we can perform accurate inference of the 3D locations of 19 joints on the body based on observed silhouette contours from real images.
Resumo:
We formulate and interpret several multi-modal registration methods in the context of a unified statistical and information theoretic framework. A unified interpretation clarifies the implicit assumptions of each method yielding a better understanding of their relative strengths and weaknesses. Additionally, we discuss a generative statistical model from which we derive a novel analysis tool, the "auto-information function", as a means of assessing and exploiting the common spatial dependencies inherent in multi-modal imagery. We analytically derive useful properties of the "auto-information" as well as verify them empirically on multi-modal imagery. Among the useful aspects of the "auto-information function" is that it can be computed from imaging modalities independently and it allows one to decompose the search space of registration problems.
Resumo:
This thesis addresses the problem of recognizing solid objects in the three-dimensional world, using two-dimensional shape information extracted from a single image. Objects can be partly occluded and can occur in cluttered scenes. A model based approach is taken, where stored models are matched to an image. The matching problem is separated into two stages, which employ different representations of objects. The first stage uses the smallest possible number of local features to find transformations from a model to an image. This minimizes the amount of search required in recognition. The second stage uses the entire edge contour of an object to verify each transformation. This reduces the chance of finding false matches.
Resumo:
Rapid judgments about the properties and spatial relations of objects are the crux of visually guided interaction with the world. Vision begins, however, with essentially pointwise representations of the scene, such as arrays of pixels or small edge fragments. For adequate time-performance in recognition, manipulation, navigation, and reasoning, the processes that extract meaningful entities from the pointwise representations must exploit parallelism. This report develops a framework for the fast extraction of scene entities, based on a simple, local model of parallel computation.sAn image chunk is a subset of an image that can act as a unit in the course of spatial analysis. A parallel preprocessing stage constructs a variety of simple chunks uniformly over the visual array. On the basis of these chunks, subsequent serial processes locate relevant scene components and assemble detailed descriptions of them rapidly. This thesis defines image chunks that facilitate the most potentially time-consuming operations of spatial analysis---boundary tracing, area coloring, and the selection of locations at which to apply detailed analysis. Fast parallel processes for computing these chunks from images, and chunk-based formulations of indexing, tracing, and coloring, are presented. These processes have been simulated and evaluated on the lisp machine and the connection machine.