92 resultados para Lempel-Ziv (LZ)
Resumo:
Lossless compression algorithms of the Lempel-Ziv (LZ) family are widely used nowadays. Regarding time and memory requirements, LZ encoding is much more demanding than decoding. In order to speed up the encoding process, efficient data structures, like suffix trees, have been used. In this paper, we explore the use of suffix arrays to hold the dictionary of the LZ encoder, and propose an algorithm to search over it. We show that the resulting encoder attains roughly the same compression ratios as those based on suffix trees. However, the amount of memory required by the suffix array is fixed, and much lower than the variable amount of memory used by encoders based on suffix trees (which depends on the text to encode). We conclude that suffix arrays, when compared to suffix trees in terms of the trade-off among time, memory, and compression ratio, may be preferable in scenarios (e.g., embedded systems) where memory is at a premium and high speed is not critical.
Resumo:
This paper demonstrates the oscillatory characteristics of electrical signals acquired from two ornamental plant types (Epipremnum pinnatum and Philodendron scandens - Family Araceae), using a noninvasive acquisition system. The electrical signal was recorded using Ag/AgCl superficial electrodes inside a Faraday cage. The presence of the oscillatory electric generator was shown using a classical power spectral density. The Lempel and Ziv complexity measurement showed that the plant signal was not noise despite its nonlinear behavior. The oscillatory characteristics of the signal were explained using a simulated electrical model that establishes that for a frequency range from 5 to 15 Hz, the oscillatory characteristic is higher than for other frequency ranges. All results show that non-invasive electrical plant signals can be acquired with improvement of signal-to-noise ratio using a Faraday cage, and a simple electrical model is able to explain the electrical signal being generated. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
Molti metodi di compressione lossless si basano sulle idee che nel 1977 i ricercatori israeliani Abraham Lempel e Jacob Ziv hanno presentato nell’articolo “A universal Algorithm for sequential Data Compression”. In questa tesi viene descritto il metodo di fattorizzazione LZ77, illustrato appunto da Lempel e Ziv, e vengono esposte le strutture dati fondamentali per la sua realizzazione. Sono inoltre descritti due algoritmi CPS1 e CPS2 che realizzano LZ77. Infine, sfruttando i dati raccolti sperimentalmente da Al-Haffedh et al. in “A Comparison of Index-Based Lempel-Ziv LZ77 Factorization Algorithms” [2012], gli algoritmi descritti vengono confrontati in termini di spazio e tempo.
Resumo:
Objective The neurodevelopmental–neurodegenerative debate is a basic issue in the field of the neuropathological basis of schizophrenia (SCH). Neurophysiological techniques have been scarcely involved in such debate, but nonlinear analysis methods may contribute to it. Methods Fifteen patients (age range 23–42 years) matching DSM IV-TR criteria for SCH, and 15 sex- and age-matched control subjects (age range 23–42 years) underwent a resting-state magnetoencephalographic evaluation and Lempel–Ziv complexity (LZC) scores were calculated. Results Regression analyses indicated that LZC values were strongly dependent on age. Complexity scores increased as a function of age in controls, while SCH patients exhibited a progressive reduction of LZC values. A logistic model including LZC scores, age and the interaction of both variables allowed the classification of patients and controls with high sensitivity and specificity. Conclusions Results demonstrated that SCH patients failed to follow the “normal” process of complexity increase as a function of age. In addition, SCH patients exhibited a significant reduction of complexity scores as a function of age, thus paralleling the pattern observed in neurodegenerative diseases. Significance Our results support the notion of a progressive defect in SCH, which does not contradict the existence of a basic neurodevelopmental alteration. Highlights ► Schizophrenic patients show higher complexity values as compared to controls. ► Schizophrenic patients showed a tendency to reduced complexity values as a function of age while controls showed the opposite tendency. ► The tendency observed in schizophrenic patients parallels the tendency observed in Alzheimer disease patients.
Resumo:
Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disease characterized by motor neurons degeneration, which reduces muscular force, being very difficult to diagnose. Mathematical methods are used in order to analyze the surface electromiographic signal’s dynamic behavior (Fractal Dimension (FD) and Multiscale Entropy (MSE)), evaluate different muscle group’s synchronization (Coherence and Phase Locking Factor (PLF)) and to evaluate the signal’s complexity (Lempel-Ziv (LZ) techniques and Detrended Fluctuation Analysis (DFA)). Surface electromiographic signal acquisitions were performed in upper limb muscles, being the analysis executed for instants of contraction for ipsilateral acquisitions for patients and control groups. Results from LZ, DFA and MSE analysis present capability to distinguish between the patient group and the control group, whereas coherence, PLF and FD algorithms present results very similar for both groups. LZ, DFA and MSE algorithms appear then to be a good measure of corticospinal pathways integrity. A classification algorithm was applied to the results in combination with extracted features from the surface electromiographic signal, with an accuracy percentage higher than 70% for 118 combinations for at least one classifier. The classification results demonstrate capability to distinguish members between patients and control groups. These results can demonstrate a major importance in the disease diagnose, once surface electromyography (sEMG) may be used as an auxiliary diagnose method.
Resumo:
BACKGROUND: Modern theories define chronic pain as a multidimensional experience - the result of complex interplay between physiological and psychological factors with significant impact on patients' physical, emotional and social functioning. The development of reliable assessment tools capable of capturing the multidimensional impact of chronic pain has challenged the medical community for decades. A number of validated tools are currently used in clinical practice however they all rely on self-reporting and are therefore inherently subjective. In this study we show that a comprehensive analysis of physical activity (PA) under real life conditions may capture behavioral aspects that may reflect physical and emotional functioning.¦METHODOLOGY: PA was monitored during five consecutive days in 60 chronic pain patients and 15 pain-free healthy subjects. To analyze the various aspects of pain-related activity behaviors we defined the concept of PA 'barcoding'. The main idea was to combine different features of PA (type, intensity, duration) to define various PA states. The temporal sequence of different states was visualized as a 'barcode' which indicated that significant information about daily activity can be contained in the amount and variety of PA states, and in the temporal structure of sequence. This information was quantified using complementary measures such as structural complexity metrics (information and sample entropy, Lempel-Ziv complexity), time spent in PA states, and two composite scores, which integrate all measures. The reliability of these measures to characterize chronic pain conditions was assessed by comparing groups of subjects with clinically different pain intensity.¦CONCLUSION: The defined measures of PA showed good discriminative features. The results suggest that significant information about pain-related functional limitations is captured by the structural complexity of PA barcodes, which decreases when the intensity of pain increases. We conclude that a comprehensive analysis of daily-life PA can provide an objective appraisal of the intensity of pain.
Resumo:
In this study we simulate numerically the Reynolds' experiment for the transition from laminar to turbulent flow in a pipe. We present a discussion of the results from a dynamical systems perspective when a control parameter, the Reynolds number, is increased. The Landau scenario, where the transition is described by the excitation of infinite oscillatory modes within the fluid, is not observed. Instead what happens is best explained by the Ruelle-Takens scenario in terms of strange attractors. The Lyapunov exponent and fractal dimension for the attractor are calculated together with a measure of complex behaviour called the Lempel-Ziv complexity. (C) 2001 Elsevier B.V. B.V. All rights reserved.
Resumo:
The dynamics of dissipative and coherent N-body systems, such as a Bose-Einstein condensate, which can be described by an extended Gross-Pitaevskii formalism, is investigated. In order to analyze chaotic and unstable regimes, two approaches are considered: a metric one, based on calculations of Lyapunov exponents, and an algorithmic one, based on the Lempel-Ziv criterion. The consistency of both approaches is established, with the Lempel-Ziv algorithmic found as an efficient complementary approach to the metric one for the fast characterization of dynamical behaviors obtained from finite sequences. © 2013 Elsevier B.V. All rights reserved.
Resumo:
Pós-graduação em Matemática Universitária - IGCE
Resumo:
The Wyner-Ziv video coding (WZVC) rate distortion performance is highly dependent on the quality of the side information, an estimation of the original frame, created at the decoder. This paper, characterizes the WZVC efficiency when motion compensated frame interpolation (MCFI) techniques are used to generate the side information, a difficult problem in WZVC especially because the decoder only has available some reference decoded frames. The proposed WZVC compression efficiency rate model relates the power spectral of the estimation error to the accuracy of the MCFI motion field. Then, some interesting conclusions may be derived related to the impact of the motion field smoothness and the correlation to the true motion trajectories on the compression performance.
Resumo:
Wyner-Ziv (WZ) video coding is a particular case of distributed video coding, the recent video coding paradigm based on the Slepian-Wolf and Wyner-Ziv theorems that exploits the source correlation at the decoder and not at the encoder as in predictive video coding. Although many improvements have been done over the last years, the performance of the state-of-the-art WZ video codecs still did not reach the performance of state-of-the-art predictive video codecs, especially for high and complex motion video content. This is also true in terms of subjective image quality mainly because of a considerable amount of blocking artefacts present in the decoded WZ video frames. This paper proposes an adaptive deblocking filter to improve both the subjective and objective qualities of the WZ frames in a transform domain WZ video codec. The proposed filter is an adaptation of the advanced deblocking filter defined in the H.264/AVC (advanced video coding) standard to a WZ video codec. The results obtained confirm the subjective quality improvement and objective quality gains that can go up to 0.63 dB in the overall for sequences with high motion content when large group of pictures are used.
Resumo:
Wyner - Ziv (WZ) video coding is a particular case of distributed video coding (DVC), the recent video coding paradigm based on the Slepian - Wolf and Wyner - Ziv theorems which exploits the source temporal correlation at the decoder and not at the encoder as in predictive video coding. Although some progress has been made in the last years, WZ video coding is still far from the compression performance of predictive video coding, especially for high and complex motion contents. The WZ video codec adopted in this study is based on a transform domain WZ video coding architecture with feedback channel-driven rate control, whose modules have been improved with some recent coding tools. This study proposes a novel motion learning approach to successively improve the rate-distortion (RD) performance of the WZ video codec as the decoding proceeds, making use of the already decoded transform bands to improve the decoding process for the remaining transform bands. The results obtained reveal gains up to 2.3 dB in the RD curves against the performance for the same codec without the proposed motion learning approach for high motion sequences and long group of pictures (GOP) sizes.
Resumo:
Recently, several distributed video coding (DVC) solutions based on the distributed source coding (DSC) paradigm have appeared in the literature. Wyner-Ziv (WZ) video coding, a particular case of DVC where side information is made available at the decoder, enable to achieve a flexible distribution of the computational complexity between the encoder and decoder, promising to fulfill novel requirements from applications such as video surveillance, sensor networks and mobile camera phones. The quality of the side information at the decoder has a critical role in determining the WZ video coding rate-distortion (RD) performance, notably to raise it to a level as close as possible to the RD performance of standard predictive video coding schemes. Towards this target, efficient motion search algorithms for powerful frame interpolation are much needed at the decoder. In this paper, the RD performance of a Wyner-Ziv video codec is improved by using novel, advanced motion compensated frame interpolation techniques to generate the side information. The development of these type of side information estimators is a difficult problem in WZ video coding, especially because the decoder only has available some reference, decoded frames. Based on the regularization of the motion field, novel side information creation techniques are proposed in this paper along with a new frame interpolation framework able to generate higher quality side information at the decoder. To illustrate the RD performance improvements, this novel side information creation framework has been integrated in a transform domain turbo coding based Wyner-Ziv video codec. Experimental results show that the novel side information creation solution leads to better RD performance than available state-of-the-art side information estimators, with improvements up to 2 dB: moreover, it allows outperforming H.264/AVC Intra by up to 3 dB with a lower encoding complexity.
Resumo:
Video coding technologies have played a major role in the explosion of large market digital video applications and services. In this context, the very popular MPEG-x and H-26x video coding standards adopted a predictive coding paradigm, where complex encoders exploit the data redundancy and irrelevancy to 'control' much simpler decoders. This codec paradigm fits well applications and services such as digital television and video storage where the decoder complexity is critical, but does not match well the requirements of emerging applications such as visual sensor networks where the encoder complexity is more critical. The Slepian Wolf and Wyner-Ziv theorems brought the possibility to develop the so-called Wyner-Ziv video codecs, following a different coding paradigm where it is the task of the decoder, and not anymore of the encoder, to (fully or partly) exploit the video redundancy. Theoretically, Wyner-Ziv video coding does not incur in any compression performance penalty regarding the more traditional predictive coding paradigm (at least for certain conditions). In the context of Wyner-Ziv video codecs, the so-called side information, which is a decoder estimate of the original frame to code, plays a critical role in the overall compression performance. For this reason, much research effort has been invested in the past decade to develop increasingly more efficient side information creation methods. This paper has the main objective to review and evaluate the available side information methods after proposing a classification taxonomy to guide this review, allowing to achieve more solid conclusions and better identify the next relevant research challenges. After classifying the side information creation methods into four classes, notably guess, try, hint and learn, the review of the most important techniques in each class and the evaluation of some of them leads to the important conclusion that the side information creation methods provide better rate-distortion (RD) performance depending on the amount of temporal correlation in each video sequence. It became also clear that the best available Wyner-Ziv video coding solutions are almost systematically based on the learn approach. The best solutions are already able to systematically outperform the H.264/AVC Intra, and also the H.264/AVC zero-motion standard solutions for specific types of content. (C) 2013 Elsevier B.V. All rights reserved.