399 resultados para Quantization
Resumo:
This paper describes an MPEG (moving pictures expert group) audio layer II - LFE (lower frequency extension) bit-stream processor targeting DAB (digital audio broadcasting) receivers that will handle the decoding of the frames in a computationally efficient manner to provide a synthesis sub-band filter with the reconstructed sub-band samples. Focus is given to the frequency sample reconstruction part, which handles the re-quantization and re-scaling of the samples once the necessary information is extracted from the frame. The comparison to a direct implementation of the frequency sample reconstruction block is carried out to prove increased computational efficiency.
Resumo:
Le présent mémoire comprend un survol des principales méthodes de rendu en demi-tons, de l’analog screening à la recherche binaire directe en passant par l’ordered dither, avec une attention particulière pour la diffusion d’erreur. Ces méthodes seront comparées dans la perspective moderne de la sensibilité à la structure. Une nouvelle méthode de rendu en demi-tons par diffusion d’erreur est présentée et soumise à diverses évaluations. La méthode proposée se veut originale, simple, autant à même de préserver le caractère structurel des images que la méthode à l’état de l’art, et plus rapide que cette dernière par deux à trois ordres de magnitude. D’abord, l’image est décomposée en fréquences locales caractéristiques. Puis, le comportement de base de la méthode proposée est donné. Ensuite, un ensemble minutieusement choisi de paramètres permet de modifier ce comportement de façon à épouser les différents caractères fréquentiels locaux. Finalement, une calibration détermine les bons paramètres à associer à chaque fréquence possible. Une fois l’algorithme assemblé, toute image peut être traitée très rapidement : chaque pixel est attaché à une fréquence propre, cette fréquence sert d’indice pour la table de calibration, les paramètres de diffusion appropriés sont récupérés, et la couleur de sortie déterminée pour le pixel contribue en espérance à souligner la structure dont il fait partie.
Resumo:
On révise les prérequis de géométrie différentielle nécessaires à une première approche de la théorie de la quantification géométrique, c'est-à-dire des notions de base en géométrie symplectique, des notions de groupes et d'algèbres de Lie, d'action d'un groupe de Lie, de G-fibré principal, de connexion, de fibré associé et de structure presque-complexe. Ceci mène à une étude plus approfondie des fibrés en droites hermitiens, dont une condition d'existence de fibré préquantique sur une variété symplectique. Avec ces outils en main, nous commençons ensuite l'étude de la quantification géométrique, étape par étape. Nous introduisons la théorie de la préquantification, i.e. la construction des opérateurs associés à des observables classiques et la construction d'un espace de Hilbert. Des problèmes majeurs font surface lors de l'application concrète de la préquantification : les opérateurs ne sont pas ceux attendus par la première quantification et l'espace de Hilbert formé est trop gros. Une première correction, la polarisation, élimine quelques problèmes, mais limite grandement l'ensemble des observables classiques que l'on peut quantifier. Ce mémoire n'est pas un survol complet de la quantification géométrique, et cela n'est pas son but. Il ne couvre ni la correction métaplectique, ni le noyau BKS. Il est un à-côté de lecture pour ceux qui s'introduisent à la quantification géométrique. D'une part, il introduit des concepts de géométrie différentielle pris pour acquis dans (Woodhouse [21]) et (Sniatycki [18]), i.e. G-fibrés principaux et fibrés associés. Enfin, il rajoute des détails à quelques preuves rapides données dans ces deux dernières références.
Resumo:
This thesis deals with some aspects of the Physics of the early universe, like phase transitions, bubble nucleations and premodial density perturbations which lead to the formation structures in the universe. Quantum aspects of the gravitational interaction play an essential role in retical high-energy physics. The questions of the quantum gravity are naturally connected with early universe and Grand Unification Theories. In spite of numerous efforts, the various problems of quantum gravity remain still unsolved. In this condition, the consideration of different quantum gravity models is an inevitable stage to study the quantum aspects of gravitational interaction. The important role of gravitationally coupled scalar field in the physics of the early universe is discussed in this thesis. The study shows that the scalar-gravitational coupling and the scalar curvature did play a crucial role in determining the nature of phase transitions that took place in the early universe. The key idea in studying the formation structure in the universe is that of gravitational instability.
Resumo:
The thesis introduced the octree and addressed the complete nature of problems encountered, while building and imaging system based on octrees. An efficient Bottom-up recursive algorithm and its iterative counterpart for the raster to octree conversion of CAT scan slices, to improve the speed of generating the octree from the slices, the possibility of utilizing the inherent parallesism in the conversion programme is explored in this thesis. The octree node, which stores the volume information in cube often stores the average density information could lead to “patchy”distribution of density during the image reconstruction. In an attempt to alleviate this problem and explored the possibility of using VQ to represent the imformation contained within a cube. Considering the ease of accommodating the process of compressing the information during the generation of octrees from CAT scan slices, proposed use of wavelet transforms to generate the compressed information in a cube. The modified algorithm for generating octrees from the slices is shown to accommodate the eavelet compression easily. Rendering the stored information in the form of octree is a complex task, necessarily because of the requirement to display the volumetric information. The reys traced from each cube in the octree, sum up the density en-route, accounting for the opacities and transparencies produced due to variations in density.
Resumo:
This thesis Entitled Studies on Quasinormal modes and Late-time tails black hole spacetimes. In this thesis, the signature of these new theories are probed on the evolution of field perturbations on the black hole spacetimes in the theory. Chapter 1 gives a general introduction to black holes and its perturbation formalism. Various concepts in the area covered by the thesis are also elucidated in this chapter. Chapter 2 describes the evolution of massive, charged scalar field perturbations around a Reissner-Nordstrom black hole surrounded by a static and spherically symmetric quintessence. Chapter 3 comprises the evolution of massless scalar, electromagnetic and gravitational fields around spherically symmetric black hole whose asymptotes are defined by the quintessence, with special interest on the late-time behavior. Chapter 4 examines the evolution of Dirac field around a Schwarzschild black hole surrounded by quintessence. Detailed numerical simulations are done to analyze the nature of field on different surfaces of constant radius . Chapter 5is dedicated to the study of the evolution of massless fields around the black hole geometry in the HL gravity.
Resumo:
In 1931 Dirac studied the motion of an electron in the field of a magnetic monopole and found that the quantization of electric charge can be explained by postulating the mere existence of a magnetic monopole. Since 1974 there has been a resurgence of interest in magnetic monopole due to the work of ‘t’ Hooft and Polyakov who independently observed that monopoles can exist as finite energy topologically stable solutions to certain spontaneously broken gauge theories. The thesis, “Studies on Magnetic Monopole Solutions of Non-abelian Gauge Theories and Related Problems”, reports a systematic investigation of classical solutions of non-abelian gauge theories with special emphasis on magnetic monopoles and dyons which possess both electric and magnetic charges. The formation of bound states of a dyon with fermions and bosons is also studied in detail. The thesis opens with an account of a new derivation of a relationship between the magnetic charge of a dyon and the topology of the gauge fields associated with it. Although this formula has been reported earlier in the literature, the present method has two distinct advantages. In the first place, it does not depend either on the mechanism of symmetry breaking or on the nature of the residual symmetry group. Secondly, the results can be generalized to finite temperature monopoles.
Resumo:
Any automatically measurable, robust and distinctive physical characteristic or personal trait that can be used to identify an individual or verify the claimed identity of an individual, referred to as biometrics, has gained significant interest in the wake of heightened concerns about security and rapid advancements in networking, communication and mobility. Multimodal biometrics is expected to be ultra-secure and reliable, due to the presence of multiple and independent—verification clues. In this study, a multimodal biometric system utilising audio and facial signatures has been implemented and error analysis has been carried out. A total of one thousand face images and 250 sound tracks of 50 users are used for training the proposed system. To account for the attempts of the unregistered signatures data of 25 new users are tested. The short term spectral features were extracted from the sound data and Vector Quantization was done using K-means algorithm. Face images are identified based on Eigen face approach using Principal Component Analysis. The success rate of multimodal system using speech and face is higher when compared to individual unimodal recognition systems
Resumo:
The thesis explores the area of still image compression. The image compression techniques can be broadly classified into lossless and lossy compression. The most common lossy compression techniques are based on Transform coding, Vector Quantization and Fractals. Transform coding is the simplest of the above and generally employs reversible transforms like, DCT, DWT, etc. Mapped Real Transform (MRT) is an evolving integer transform, based on real additions alone. The present research work aims at developing new image compression techniques based on MRT. Most of the transform coding techniques employ fixed block size image segmentation, usually 8×8. Hence, a fixed block size transform coding is implemented using MRT and the merits and demerits are analyzed for both 8×8 and 4×4 blocks. The N2 unique MRT coefficients, for each block, are computed using templates. Considering the merits and demerits of fixed block size transform coding techniques, a hybrid form of these techniques is implemented to improve the performance of compression. The performance of the hybrid coder is found to be better compared to the fixed block size coders. Thus, if the block size is made adaptive, the performance can be further improved. In adaptive block size coding, the block size may vary from the size of the image to 2×2. Hence, the computation of MRT using templates is impractical due to memory requirements. So, an adaptive transform coder based on Unique MRT (UMRT), a compact form of MRT, is implemented to get better performance in terms of PSNR and HVS The suitability of MRT in vector quantization of images is then experimented. The UMRT based Classified Vector Quantization (CVQ) is implemented subsequently. The edges in the images are identified and classified by employing a UMRT based criteria. Based on the above experiments, a new technique named “MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ)”is developed. Its performance is evaluated and compared against existing techniques. A comparison with standard JPEG & the well-known Shapiro’s Embedded Zero-tree Wavelet (EZW) is done and found that the proposed technique gives better performance for majority of images
Resumo:
The image comparison operation ??sessing how well one image matches another ??rms a critical component of many image analysis systems and models of human visual processing. Two norms used commonly for this purpose are L1 and L2, which are specific instances of the Minkowski metric. However, there is often not a principled reason for selecting one norm over the other. One way to address this problem is by examining whether one metric better captures the perceptual notion of image similarity than the other. With this goal, we examined perceptual preferences for images retrieved on the basis of the L1 versus the L2 norm. These images were either small fragments without recognizable content, or larger patterns with recognizable content created via vector quantization. In both conditions the subjects showed a consistent preference for images matched using the L1 metric. These results suggest that, in the domain of natural images of the kind we have used, the L1 metric may better capture human notions of image similarity.
Resumo:
In this paper, an information theoretic framework for image segmentation is presented. This approach is based on the information channel that goes from the image intensity histogram to the regions of the partitioned image. It allows us to define a new family of segmentation methods which maximize the mutual information of the channel. Firstly, a greedy top-down algorithm which partitions an image into homogeneous regions is introduced. Secondly, a histogram quantization algorithm which clusters color bins in a greedy bottom-up way is defined. Finally, the resulting regions in the partitioning algorithm can optionally be merged using the quantized histogram
Resumo:
The principles of operation of an experimental prototype instrument known as J-SCAN are described along with the derivation of formulae for the rapid calculation of normalized impedances; the structure of the instrument; relevant probe design parameters; digital quantization errors; and approaches for the optimization of single frequency operation. An eddy current probe is used As the inductance element of a passive tuned-circuit which is repeatedly excited with short impulses. Each impulse excites an oscillation which is subject to decay dependent upon the values of the tuned-circuit components: resistance, inductance and capacitance. Changing conditions under the probe that affect the resistance and inductance of this circuit will thus be detected through changes in the transient response. These changes in transient response, oscillation frequency and rate of decay, are digitized, and then normalized values for probe resistance and inductance changes are calculated immediately in a micro processor. This approach coupled with a minimum analogue processing and maximum of digital processing has advantages compared with the conventional approaches to eddy current instruments. In particular there are: the absence of an out of balance condition and the flexibility and stability of digital data processing.
Resumo:
In the past decade, the amount of data in biological field has become larger and larger; Bio-techniques for analysis of biological data have been developed and new tools have been introduced. Several computational methods are based on unsupervised neural network algorithms that are widely used for multiple purposes including clustering and visualization, i.e. the Self Organizing Maps (SOM). Unfortunately, even though this method is unsupervised, the performances in terms of quality of result and learning speed are strongly dependent from the neuron weights initialization. In this paper we present a new initialization technique based on a totally connected undirected graph, that report relations among some intersting features of data input. Result of experimental tests, where the proposed algorithm is compared to the original initialization techniques, shows that our technique assures faster learning and better performance in terms of quantization error.
Resumo:
The precision of quasioptical null-balanced bridge instruments for transmission and reflection coefficient measurements at millimeter and submillimeter wavelengths is analyzed. A Jones matrix analysis is used to describe the amount of power reaching the detector as a function of grid angle orientation, sample transmittance/reflectance and phase delay. An analysis is performed of the errors involved in determining the complex transmission and reflection coefficient after taking into account the quantization error in the grid angle and micrometer readings, the transmission or reflection coefficient of the sample, the noise equivalent power of the detector, the source power and the post-detection bandwidth. For a system fitted with a rotating grid with resolution of 0.017 rad and a micrometer quantization error of 1 μm, a 1 mW source, and a detector with a noise equivalent power 5×10−9 W Hz−1/2, the maximum errors at an amplitude transmission or reflection coefficient of 0.5 are below ±0.025.