11 resultados para compression ratio
em Repositório Científico do Instituto Politécnico de Lisboa - Portugal
Resumo:
Lossless compression algorithms of the Lempel-Ziv (LZ) family are widely used nowadays. Regarding time and memory requirements, LZ encoding is much more demanding than decoding. In order to speed up the encoding process, efficient data structures, like suffix trees, have been used. In this paper, we explore the use of suffix arrays to hold the dictionary of the LZ encoder, and propose an algorithm to search over it. We show that the resulting encoder attains roughly the same compression ratios as those based on suffix trees. However, the amount of memory required by the suffix array is fixed, and much lower than the variable amount of memory used by encoders based on suffix trees (which depends on the text to encode). We conclude that suffix arrays, when compared to suffix trees in terms of the trade-off among time, memory, and compression ratio, may be preferable in scenarios (e.g., embedded systems) where memory is at a premium and high speed is not critical.
Resumo:
We use the first and second laws of thermodynamics to analyze the behavior of an ideal jet engine. Simple analytical expressions for the thermal efficiency, the overall efficiency, and the reduced thrust are derived. We show that the thermal efficiency depends only on the compression ratio r and on the velocity of the aircraft. The other two performance measures depend also on the ratio of the temperature at the turbine to the inlet temperature in the engine, T-3/T-i. An analysis of these expressions shows that it is not possible to choose an optimal set of values of r and T-3/T-i that maximize both the overall efficiency and thrust. We study how irreversibilities in the compressor and the turbine decrease the overall efficiency of jet engines and show that this effect is more pronounced for smaller T-3/T-i.
Resumo:
Low-density parity-check (LDPC) codes are nowadays one of the hottest topics in coding theory, notably due to their advantages in terms of bit error rate performance and low complexity. In order to exploit the potential of the Wyner-Ziv coding paradigm, practical distributed video coding (DVC) schemes should use powerful error correcting codes with near-capacity performance. In this paper, new ways to design LDPC codes for the DVC paradigm are proposed and studied. The new LDPC solutions rely on merging parity-check nodes, which corresponds to reduce the number of rows in the parity-check matrix. This allows to change gracefully the compression ratio of the source (DCT coefficient bitplane) according to the correlation between the original and the side information. The proposed LDPC codes reach a good performance for a wide range of source correlations and achieve a better RD performance when compared to the popular turbo codes.
Resumo:
The Wyner-Ziv video coding (WZVC) rate distortion performance is highly dependent on the quality of the side information, an estimation of the original frame, created at the decoder. This paper, characterizes the WZVC efficiency when motion compensated frame interpolation (MCFI) techniques are used to generate the side information, a difficult problem in WZVC especially because the decoder only has available some reference decoded frames. The proposed WZVC compression efficiency rate model relates the power spectral of the estimation error to the accuracy of the MCFI motion field. Then, some interesting conclusions may be derived related to the impact of the motion field smoothness and the correlation to the true motion trajectories on the compression performance.
Resumo:
O principal objectivo desta tese é obter uma relação directa entre a composição dos gases liquefeitos de petróleo (GLP), propano, n-butano e isobutano, usados como aerossóis propulsores numa lata de poliuretano de um componente, com as propriedades das espumas produzidas por spray. As espumas obtidas, terão de ter como requisito principal, um bom desempenho a temperaturas baixas, -10ºC, sendo por isso designadas por espumas de Inverno. Uma espuma é considerada como tendo um bom desempenho se não apresentar a -10/-10ºC (temperatura lata/ spray) glass bubbles, base holes e cell collapse. As espumas deverão ainda ter densidades do spray no molde a +23/+23ºC abaixo dos 30 g/L, um rendimento superior a 30 L, boa estabilidade dimensional e um caudal de espuma a +5/+5ºC superior a 5 g/s. Os ensaios experimentais foram realizados a +23/+23ºC, +5/+5ºC e a -10/-10ºC. A cada temperatura, as espumas desenvolvidas, foram submetidas a testes que permitiram determinar a sua qualidade. Testes esses que incluem os designados por Quick Tests (QT): o spray no papel e no molde das espumas nas referidas temperaturas. As amostras do papel e no molde são especialmente analisadas, quanto, às glass bubbles, cell collapse, base holes, cell structur e, cutting shrinkage, para além de outras propriedades. Os QT também incluem a análise da densidade no molde (ODM) e o estudo do caudal de espumas. Além dos QT foram realizados os testes da estabilidade dimensional das espumas, testes físicos de compressão e adesão, testes de expansão das espumas após spray e do rendimento por lata de espuma. Em todos os ensaios foi utilizado um tubo adaptador colocado na válvula da lata como método de spray e ainda mantida constante a proporção das matérias-primas (excepto os gases, em estudo). As experiências iniciaram-se com o estudo de GLPs presentes no mercado de aerossóis. Estes resultaram que o GLP: propano/ n-butano/ isobutano: (30/ 0/ 70 w/w%), produz as melhores espumas de inverno a -10/-10ºC, reduzindo desta forma as glass bubbles, base holes e o cell collapse produzido pelos restantes GLP usados como aerossóis nas latas de poliuretano. Testes posteriores tiveram como objectivo estudar a influência directa de cada gás, propano, n-butano e isobutano nas espumas. Para tal, foram usadas duas referências do estudo com GLP comercializáveis, 7396 (30 /0 /70 w/w %) e 7442 (0/ 0/ 100 w/w %). Com estes resultados concluí-se que o n-butano produz más propriedades nas espumas a -10/- 10ºC, formando grandes quantidades de glass bubbles, base holes e cell collapse. Contudo, o uso de propano reduz essas glass bubbles, mas em contrapartida, forma cell collapse.Isobutano, porém diminui o cell collapse mas não as glass bubbles. Dos resultados experimentais podemos constatar que o caudal a +5/+5ºC e densidade das espumas a +23/+23ºC, são influenciados pela composição do GLP. O propano e n-butano aumentam o caudal de espuma das latas e a sua densidade, ao contrário com o que acontece com o isobutano. Todavia, pelos resultados obtidos, o isobutano proporciona os melhores rendimentos de espumas por lata. Podemos concluir que os GLPs que contivessem cerca de 30 w/w % de propano (bons caudais a +5/+5ºC e menos glass bubbles a -10/-10ºC), e cerca 70 w/w % de isobutano (bons rendimentos de espumas, bem como menos cell collapse a -10/-10ºC) produziam as melhores espumas. Também foram desenvolvidos testes sobre a influência da quantidade de gás GLP presente numa lata. A análise do volume de GLP usado, foi realizada com base na melhor espuma obtida nos estudos anteriores, 7396, com um GLP (30 / 0/ 70 w/w%), e foram feitas alterações ao seu volume gás GLP presente no pré-polímero. O estudo concluiu, que o aumento do volume pode diminuir a densidade das espumas, e o seu decréscimo, um aumento da densidade. Também indico u que um mau ajuste do volume poderá causar más propriedades nas espumas. A análise económica, concluiu que o custo das espumas com mais GLP nas suas formulações, reduz-se em cerca de 3%, a quando de um aumento do volume de GLP no pré-polímero de cerca de 8 %. Esta diminuição de custos deveu-se ao facto, de um aumento de volume de gás, implicar uma diminuição na quantidade das restantes matérias-primas, com custos superiores, já que o volume útil total da lata terá de ser sempre mantido nos 750 mL. Com o objectivo de melhorar a qualidade da espuma 7396 (30/0/70 w/w %) obtida nos ensaios anteriores adicionou-se à formulação 7396 o HFC-152a (1,1-di fluoroetano). Os resultados demonstram que se formam espumas com más propriedades, especialmente a -10/-10ºC, contudo proporcionou excelentes shaking rate da lata. Através de uma pequena análise de custos não é aconselhável o seu uso pelos resultados obtidos, não proporcionando um balanço custo/benefício favorável. As três melhores espumas obtidas de todos os estudos foram comparadas com uma espuma de inverno presente no mercado. 7396 e 7638 com um volume de 27 % no prépolímero e uma composição de GLP (30/ 0 / 70 w/w%) e (13,7/ 0/ 86,3 w/w%), respectivamente, e 7690, com 37 % de volume no pré-polímero e GLP (30/ 0 / 70 w/w%), apresentaram em geral melhores resultados, comparando com a espuma benchmark . Contudo, os seus shaking rate a -10/-10ºC, de cada espuma, apresentaram valores bastante inferiores à composição benchmarking.
Resumo:
The effects of the Miocene through Present compression in the Tagus Abyssal Plain are mapped using the most up to date available to scientific community multi-channel seismic reflection and refraction data. Correlation of the rift basin fault pattern with the deep crustal structure is presented along seismic line IAM-5. Four structural domains were recognized. In the oceanic realm mild deformation concentrates in Domain I adjacent to the Tore-Madeira Rise. Domain 2 is characterized by the absence of shortening structures, except near the ocean-continent transition (OCT), implying that Miocene deformation did not propagate into the Abyssal Plain, In Domain 3 we distinguish three sub-domains: Sub-domain 3A which coincides with the OCT, Sub-domain 3B which is a highly deformed adjacent continental segment, and Sub-domain 3C. The Miocene tectonic inversion is mainly accommodated in Domain 3 by oceanwards directed thrusting at the ocean-continent transition and continentwards on the continental slope. Domain 4 corresponds to the non-rifted continental margin where only minor extensional and shortening deformation structures are observed. Finite element numerical models address the response of the various domains to the Miocene compression, emphasizing the long-wavelength differential vertical movements and the role of possible rheologic contrasts. The concentration of the Miocene deformation in the transitional zone (TC), which is the addition of Sub-domain 3A and part of 3B, is a result of two main factors: (1) focusing of compression in an already stressed region due to plate curvature and sediment loading; and (2) theological weakening. We estimate that the frictional strength in the TC is reduced in 30% relative to the surrounding regions. A model of compressive deformation propagation by means of horizontal impingement of the middle continental crust rift wedge and horizontal shearing on serpentinized mantle in the oceanic realm is presented. This model is consistent with both the geological interpretation of seismic data and the results of numerical modelling.
Resumo:
In-plane deformation of foams was studied experimentally by subjecting bidisperse foams to cycles of traction and compression at a prescribed rate. Each foam contained bubbles of two sizes with given area ratio and one of three initial arrangements: sorted perpendicular to the axis of deformation (iso-strain), sorted parallel to the axis of deformation (iso-stress), or randomly mixed. Image analysis was used to measure the characteristics of the foams, including the number of edges separating small from large bubbles N-sl, the perimeter (surface energy), the distribution of the number of sides of the bubbles, and the topological disorder mu(2)(N). Foams that were initially mixed were found to remain mixed after the deformation. The response of sorted foams, however, depended on the initial geometry, including the area fraction of small bubbles and the total number of bubbles. For a given experiment we found that (i) the perimeter of a sorted foam varied little; (ii) each foam tended towards a mixed state, measured through the saturation of N-sl; and (iii) the topological disorder mu(2)(N) increased up to an "equilibrium" value. The results of different experiments showed that (i) the change in disorder, Delta mu(2)(N), decreased with the area fraction of small bubbles under iso-strain, but was independent of it under iso-stress; and (ii) Delta mu(2)(N) increased with Delta N-sl under iso-strain, but was again independent of it under iso-stress. We offer explanations for these effects in terms of elementary topological processes induced by the deformations that occur at the bubble scale.
Resumo:
This study explores a large set of OC and EC measurements in PM(10) and PM(2.5) aerosol samples, undertaken with a long term constant analytical methodology, to evaluate the capability of the OC/EC minimum ratio to represent the ratio between the OC and EC aerosol components resulting from fossil fuel combustion (OC(ff)/EC(ff)). The data set covers a wide geographical area in Europe, but with a particular focus upon Portugal, Spain and the United Kingdom, and includes a great variety of sites: urban (background, kerbside and tunnel), industrial, rural and remote. The highest minimum ratios were found in samples from remote and rural sites. Urban background sites have shown spatially and temporally consistent minimum ratios, of around 1.0 for PM(10) and 0.7 for PM(2.5).The consistency of results has suggested that the method could be used as a tool to derive the ratio between OC and EC from fossil fuel combustion and consequently to differentiate OC from primary and secondary sources. To explore this capability, OC and EC measurements were performed in a busy roadway tunnel in central Lisbon. The OC/EC ratio, which reflected the composition of vehicle combustion emissions, was in the range of 03-0.4. Ratios of OC/EC in roadside increment air (roadside minus urban background) in Birmingham, UK also lie within the range 03-0.4. Additional measurements were performed under heavy traffic conditions at two double kerbside sites located in the centre of Lisbon and Madrid. The OC/EC minimum ratios observed at both sites were found to be between those of the tunnel and those of urban background air, suggesting that minimum values commonly obtained for this parameter in open urban atmospheres over-predict the direct emissions of OC(ff) from road transport. Possible reasons for this discrepancy are explored. (C) 2011 Elsevier Ltd. All rights reserved.
Resumo:
The fatty acid profile of erythrocyte membranes has been considered a good biomarker for several pathologic situations. Dietary intake, digestion, absorption, metabolism, storage and exchange amongst compartments, greatly influence the fatty acids composition of different cells and tissues. Lipoprotein and hepatic lipases were also involved in fatty acid availability. In the present work we examined the correlations between fatty acid in Red Blood Cells (RBCs) membranes, the fatty acid desaturase and elongase activities, glycaemia, blood lipids, lipoproteins and apoproteins, and the endothelial lipase (EL) mass in plasma. Twenty one individuals were considered in the present study, with age >18 y. RBCs membranes were obtained and analysed for fatty acid composition by gas chromatography. The amount of fatty acids (as percentage) were analysed, and the ratios between fatty acid 16:1/16:0; 18:1/18:0; 18:0/16:0; 22:6 n-3/20:5 n-3 and 20:4 n-6/18:2 n-6 were calculated. Bivariate analysis (rs) and partial correlations were determined. SCD16 estimation activity correlated positively with BMI (rs=0.466, p=0.043) and triacylglycerols (TAG) (rs=0.483, p=0.026), and negatively with the ratio ApoA1/ApoB (rs=-0.566, p=0.007). Endothelial lipase (EL) correlated positively with the EPA/AA ratio in RBCs membranes (rs=0.524, p=0.045). After multi-adjustment for BMI, age, hs-CRP and dietary n3/n6 ratio, the correlations remained significant between EL and EPA/AA ratio. At the best of our knowledge this is the first report that correlated EL with the fatty acid profile of RBCs plasma membranes. The association found here can suggest that the enzyme may be involved in the bioavailability and distribution of n-3/n-6 fatty acids, suggesting a major role for EL in the pathophysiological mechanisms involving biomembranes’ fatty acids, such as in inflammatory response and eicosanoids metabolites pathways.
Resumo:
Purpose: Evaluate the type of breast compression (gradual or no gradual) that provides less discomfort to the patient. Methods and Materials: The standard projections were simulated [craniocaudal/(CC) and mediolateral-oblique/(MLO)] with the two breast compressions in 90 volunteers women aged between 19 and 86. The women were organised in groups according to the breast density. The intensity of discomfort was evaluated using the scale that have represented several faces (0-10) proposed by Wong Baker in the end of each simulation. It was also applied an interview using focus group to debate the score that were attributed during pain evaluation and to identify the criteria that were considered to do the classification. Results: The women aged between 19-29y (with higher breast density) classified the pain during no gradual compression as 4 and the gradual compression as 2 for both projections. The MLO projection was considered the most uncomfortable. During the focus group interview applied to this group was highlighted that compression did not promoted pain but discomfort. They considered that the high expectations of pain did not correspond to the discomfort that they felt. Similar results were identified for the older women (30-50y; > 50y). Conclusion: The radiographers should considerer the technique for breast compression. The gradual compression was considered for the majority of the women as the most comfortable regardless of breast density. The MLO projection was considered as uncomfortable due to the positioning (axila and inclusion of pectoral muscle) and due to the higher breast compression compared to the CC projection.
Resumo:
Arguably, the most difficult task in text classification is to choose an appropriate set of features that allows machine learning algorithms to provide accurate classification. Most state-of-the-art techniques for this task involve careful feature engineering and a pre-processing stage, which may be too expensive in the emerging context of massive collections of electronic texts. In this paper, we propose efficient methods for text classification based on information-theoretic dissimilarity measures, which are used to define dissimilarity-based representations. These methods dispense with any feature design or engineering, by mapping texts into a feature space using universal dissimilarity measures; in this space, classical classifiers (e.g. nearest neighbor or support vector machines) can then be used. The reported experimental evaluation of the proposed methods, on sentiment polarity analysis and authorship attribution problems, reveals that it approximates, sometimes even outperforms previous state-of-the-art techniques, despite being much simpler, in the sense that they do not require any text pre-processing or feature engineering.