15 resultados para Compression efficiency
em Repositório Científico do Instituto Politécnico de Lisboa - Portugal
Resumo:
The Wyner-Ziv video coding (WZVC) rate distortion performance is highly dependent on the quality of the side information, an estimation of the original frame, created at the decoder. This paper, characterizes the WZVC efficiency when motion compensated frame interpolation (MCFI) techniques are used to generate the side information, a difficult problem in WZVC especially because the decoder only has available some reference decoded frames. The proposed WZVC compression efficiency rate model relates the power spectral of the estimation error to the accuracy of the MCFI motion field. Then, some interesting conclusions may be derived related to the impact of the motion field smoothness and the correlation to the true motion trajectories on the compression performance.
Resumo:
Video coding technologies have played a major role in the explosion of large market digital video applications and services. In this context, the very popular MPEG-x and H-26x video coding standards adopted a predictive coding paradigm, where complex encoders exploit the data redundancy and irrelevancy to 'control' much simpler decoders. This codec paradigm fits well applications and services such as digital television and video storage where the decoder complexity is critical, but does not match well the requirements of emerging applications such as visual sensor networks where the encoder complexity is more critical. The Slepian Wolf and Wyner-Ziv theorems brought the possibility to develop the so-called Wyner-Ziv video codecs, following a different coding paradigm where it is the task of the decoder, and not anymore of the encoder, to (fully or partly) exploit the video redundancy. Theoretically, Wyner-Ziv video coding does not incur in any compression performance penalty regarding the more traditional predictive coding paradigm (at least for certain conditions). In the context of Wyner-Ziv video codecs, the so-called side information, which is a decoder estimate of the original frame to code, plays a critical role in the overall compression performance. For this reason, much research effort has been invested in the past decade to develop increasingly more efficient side information creation methods. This paper has the main objective to review and evaluate the available side information methods after proposing a classification taxonomy to guide this review, allowing to achieve more solid conclusions and better identify the next relevant research challenges. After classifying the side information creation methods into four classes, notably guess, try, hint and learn, the review of the most important techniques in each class and the evaluation of some of them leads to the important conclusion that the side information creation methods provide better rate-distortion (RD) performance depending on the amount of temporal correlation in each video sequence. It became also clear that the best available Wyner-Ziv video coding solutions are almost systematically based on the learn approach. The best solutions are already able to systematically outperform the H.264/AVC Intra, and also the H.264/AVC zero-motion standard solutions for specific types of content. (C) 2013 Elsevier B.V. All rights reserved.
Resumo:
The growing heterogeneity of networks, devices and consumption conditions asks for flexible and adaptive video coding solutions. The compression power of the HEVC standard and the benefits of the distributed video coding paradigm allow designing novel scalable coding solutions with improved error robustness and low encoding complexity while still achieving competitive compression efficiency. In this context, this paper proposes a novel scalable video coding scheme using a HEVC Intra compliant base layer and a distributed coding approach in the enhancement layers (EL). This design inherits the HEVC compression efficiency while providing low encoding complexity at the enhancement layers. The temporal correlation is exploited at the decoder to create the EL side information (SI) residue, an estimation of the original residue. The EL encoder sends only the data that cannot be inferred at the decoder, thus exploiting the correlation between the original and SI residues; however, this correlation must be characterized with an accurate correlation model to obtain coding efficiency improvements. Therefore, this paper proposes a correlation modeling solution to be used at both encoder and decoder, without requiring a feedback channel. Experiments results confirm that the proposed scalable coding scheme has lower encoding complexity and provides BD-Rate savings up to 3.43% in comparison with the HEVC Intra scalable extension under development. © 2014 IEEE.
Resumo:
As high dynamic range video is gaining popularity, video coding solutions able to efficiently provide both low and high dynamic range video, notably with a single bitstream, are increasingly important. While simulcasting can provide both dynamic range videos at the cost of some compression efficiency penalty, bit-depth scalable video coding can provide a better trade-off between compression efficiency, adaptation flexibility and computational complexity. Considering the widespread use of H.264/AVC video, this paper proposes a H.264/AVC backward compatible bit-depth scalable video coding solution offering a low dynamic range base layer and two high dynamic range enhancement layers with different qualities, at low complexity. Experimental results show that the proposed solution has an acceptable rate-distortion performance penalty regarding the HDR H.264/AVC single-layer coding solution.
Resumo:
Lossless compression algorithms of the Lempel-Ziv (LZ) family are widely used nowadays. Regarding time and memory requirements, LZ encoding is much more demanding than decoding. In order to speed up the encoding process, efficient data structures, like suffix trees, have been used. In this paper, we explore the use of suffix arrays to hold the dictionary of the LZ encoder, and propose an algorithm to search over it. We show that the resulting encoder attains roughly the same compression ratios as those based on suffix trees. However, the amount of memory required by the suffix array is fixed, and much lower than the variable amount of memory used by encoders based on suffix trees (which depends on the text to encode). We conclude that suffix arrays, when compared to suffix trees in terms of the trade-off among time, memory, and compression ratio, may be preferable in scenarios (e.g., embedded systems) where memory is at a premium and high speed is not critical.
Resumo:
We use the first and second laws of thermodynamics to analyze the behavior of an ideal jet engine. Simple analytical expressions for the thermal efficiency, the overall efficiency, and the reduced thrust are derived. We show that the thermal efficiency depends only on the compression ratio r and on the velocity of the aircraft. The other two performance measures depend also on the ratio of the temperature at the turbine to the inlet temperature in the engine, T-3/T-i. An analysis of these expressions shows that it is not possible to choose an optimal set of values of r and T-3/T-i that maximize both the overall efficiency and thrust. We study how irreversibilities in the compressor and the turbine decrease the overall efficiency of jet engines and show that this effect is more pronounced for smaller T-3/T-i.
Resumo:
A compactação dos equipamentos de tecnologia de informação e os aumentos simultâneos no consumo de energia dos processadores levam a que seja assegurada a distribuição adequada de ar frio, a remoção do ar quente, a capacidade adequada de arrefecimento e uma diminuição do consumo de energia. Considerando-se a cogeração como uma alternativa energeticamente eficiente em relação a outros métodos de produção de energia, com este trabalho faz-se a análise à rentabilidade de uma eventual integração de um sistema de cogeração num centro informático.
Resumo:
The effects of the Miocene through Present compression in the Tagus Abyssal Plain are mapped using the most up to date available to scientific community multi-channel seismic reflection and refraction data. Correlation of the rift basin fault pattern with the deep crustal structure is presented along seismic line IAM-5. Four structural domains were recognized. In the oceanic realm mild deformation concentrates in Domain I adjacent to the Tore-Madeira Rise. Domain 2 is characterized by the absence of shortening structures, except near the ocean-continent transition (OCT), implying that Miocene deformation did not propagate into the Abyssal Plain, In Domain 3 we distinguish three sub-domains: Sub-domain 3A which coincides with the OCT, Sub-domain 3B which is a highly deformed adjacent continental segment, and Sub-domain 3C. The Miocene tectonic inversion is mainly accommodated in Domain 3 by oceanwards directed thrusting at the ocean-continent transition and continentwards on the continental slope. Domain 4 corresponds to the non-rifted continental margin where only minor extensional and shortening deformation structures are observed. Finite element numerical models address the response of the various domains to the Miocene compression, emphasizing the long-wavelength differential vertical movements and the role of possible rheologic contrasts. The concentration of the Miocene deformation in the transitional zone (TC), which is the addition of Sub-domain 3A and part of 3B, is a result of two main factors: (1) focusing of compression in an already stressed region due to plate curvature and sediment loading; and (2) theological weakening. We estimate that the frictional strength in the TC is reduced in 30% relative to the surrounding regions. A model of compressive deformation propagation by means of horizontal impingement of the middle continental crust rift wedge and horizontal shearing on serpentinized mantle in the oceanic realm is presented. This model is consistent with both the geological interpretation of seismic data and the results of numerical modelling.
Resumo:
We have studied, in particular under normality of the implied random variables, the connections between different measures of risk such as the standard deviation, the W-ruin probability and the p-V@R. We discuss conditions granting the equivalence of these measures with respect to risk preference relations and the equivalence of dominance and efficiency of risk-reward criteria involving these measures. Then more specifically we applied these concepts to rigorously face the problem of finding the efficient set of de Finetti’s variable quota share proportional reinsurance.
Resumo:
Based on our recent discovery of closed form formulae of efficient Mean Variance retentions in variable quota-share proportional reinsurance under group correlation, we analyzed the influence of different combination of correlation and safety loading levels on the efficient frontier, both in a single period stylized problem and in a multiperiod one.
Resumo:
Since industrialization and the formation of larger urban centers in the nineteenth century, pollution of the environment was always present in daily life in various ways, namely in the form of light. Light pollution can cause various consequences, both for humans and for their ecosystem, producing effects on environmental, social, economic and scientific level. In Portugal, the lighting is responsible for 3% of total electricity consumption, energy costs are in some cases more than 50% towards the costs incurred by municipalities with energy, checking-in recent years a trend similar to that improvement of illumination levels in the region (about 4 to 5% per year). Proper use of lighting brings many benefits both to the citizen and environment, since greater energy efficiency can contribute to reducing CO2 emissions, energy costs, as well as to decrease the use of resources not-renewable and/or contamination of renewable resources, which can occurs in the process of obtaining electricity. The present study has a main goal to analyze the illuminance levels associated to the public lighting of the village of Vialonga, Vila Franca de Xira (Portugal), to verify if it is efficient. The aim is also to relate the efficiency of street lighting with the existence of light pollution.
Resumo:
Purpose: Evaluate the type of breast compression (gradual or no gradual) that provides less discomfort to the patient. Methods and Materials: The standard projections were simulated [craniocaudal/(CC) and mediolateral-oblique/(MLO)] with the two breast compressions in 90 volunteers women aged between 19 and 86. The women were organised in groups according to the breast density. The intensity of discomfort was evaluated using the scale that have represented several faces (0-10) proposed by Wong Baker in the end of each simulation. It was also applied an interview using focus group to debate the score that were attributed during pain evaluation and to identify the criteria that were considered to do the classification. Results: The women aged between 19-29y (with higher breast density) classified the pain during no gradual compression as 4 and the gradual compression as 2 for both projections. The MLO projection was considered the most uncomfortable. During the focus group interview applied to this group was highlighted that compression did not promoted pain but discomfort. They considered that the high expectations of pain did not correspond to the discomfort that they felt. Similar results were identified for the older women (30-50y; > 50y). Conclusion: The radiographers should considerer the technique for breast compression. The gradual compression was considered for the majority of the women as the most comfortable regardless of breast density. The MLO projection was considered as uncomfortable due to the positioning (axila and inclusion of pectoral muscle) and due to the higher breast compression compared to the CC projection.
Resumo:
Urinary tract infection (UTI) is one of the most prevalent pathologies in developed countries, particularly in women, characterized by the presence of bacterial growth in any part of the urinary system. Currently, urine culture is considered the gold standard method for the diagnosis of UTI. However, this method has several disadvantages including the time necessary for obtaining the results and the associated high costs. Therefore, it is important to evaluate new efficient and valuable methods for the diagnosis of these infections. Objectives: Presently, dipsticks are considered a possible valuable alternative to urine culture. This method has very low costs associated and the results can be obtained in few minutes. Here we aim to compare the sensibility, specificity, predictive value of a positive test and a negative test of both methods in order to determine the efficiency of the test strips method and also to characterize the microorganism more frequently isolated.
Resumo:
Reporter genes are routinely used in every laboratory for molecular and cellular biology for studying heterologous gene expression and general cellular biological mechanisms, such as transfection processes. Although well characterized and broadly implemented, reporter genes present serious limitations, either by involving time-consuming procedures or by presenting possible side effects on the expression of the heterologous gene or even in the general cellular metabolism. Fourier transform mid-infrared (FT-MIR) spectroscopy was evaluated to simultaneously analyze in a rapid (minutes) and high-throughput mode (using 96-wells microplates), the transfection efficiency, and the effect of the transfection process on the host cell biochemical composition and metabolism. Semi-adherent HEK and adherent AGS cell lines, transfected with the plasmid pVAX-GFP using Lipofectamine, were used as model systems. Good partial least squares (PLS) models were built to estimate the transfection efficiency, either considering each cell line independently (R 2 ≥ 0.92; RMSECV ≤ 2 %) or simultaneously considering both cell lines (R 2 = 0.90; RMSECV = 2 %). Additionally, the effect of the transfection process on the HEK cell biochemical and metabolic features could be evaluated directly from the FT-IR spectra. Due to the high sensitivity of the technique, it was also possible to discriminate the effect of the transfection process from the transfection reagent on KEK cells, e.g., by the analysis of spectral biomarkers and biochemical and metabolic features. The present results are far beyond what any reporter gene assay or other specific probe can offer for these purposes.
Resumo:
Arguably, the most difficult task in text classification is to choose an appropriate set of features that allows machine learning algorithms to provide accurate classification. Most state-of-the-art techniques for this task involve careful feature engineering and a pre-processing stage, which may be too expensive in the emerging context of massive collections of electronic texts. In this paper, we propose efficient methods for text classification based on information-theoretic dissimilarity measures, which are used to define dissimilarity-based representations. These methods dispense with any feature design or engineering, by mapping texts into a feature space using universal dissimilarity measures; in this space, classical classifiers (e.g. nearest neighbor or support vector machines) can then be used. The reported experimental evaluation of the proposed methods, on sentiment polarity analysis and authorship attribution problems, reveals that it approximates, sometimes even outperforms previous state-of-the-art techniques, despite being much simpler, in the sense that they do not require any text pre-processing or feature engineering.