945 resultados para tablet compression
Resumo:
Purpose: To study the mechanical and dynamic swelling properties of grewia gum, evaluate its compression behaviour and determine the effect of drying methods on its properties. Methods: Compacts (500 mg) of both freeze-dried and air-dried grewia gum were separately prepared by compression on a potassium bromide (KBr) press at different pressures and subjected to Heckel analysis. Swelling studies were performed using 200 mg compacts of the gum (freeze-dried or air-dried) compressed on a KBr press. The mechanical properties of the films of the gum prepared by casting 1 % dispersions of the gum were evaluated using Hounsfield tensiometer. The mechanical properties of grewia gum films were compared with films of pullulan and guar gum which were similarly prepared. The effect of temperature on the water uptake of the compacts was studied and the data subjected to Schott's analysis. Results: Drying conditions had no effect on the yield pressure of the gum compacts as both air-dried and freeze-dried fractions had a yield pressure of 322.6 MPa. The plots based on Schott's equation for the grewia gum samples showed that both samples (freeze-dried and air-dried) exhibited long swelling times. Grewia gum film had a tensile strength of 19.22±3.61 MPa which was similar to that of pullulan films (p > 0.05). It had an elastic modulus of 2.13±0.12 N/mm2 which was significantly lower (p < 0.05) than those of pullulan and guar gum with elastic moduli of 3.33±0.00 and 2.86±0.00 N/mm2, respectively. Conclusion: The type of drying method used does not have any effect on the degree of plasticity of grewia gum compacts. Grewia gum obtained by either drying method exhibited extended swelling duration. Matrix tablet formulations of the gum will likely swell slowly and promote sustained release of drug. © Pharmacotherapy Group, Faculty of Pharmacy, University of Benin, Benin City.
Resumo:
We present comprehensive design rules to optimize the process of spectral compression arising from nonlinear pulse propagation in an optical fiber. Extensive numerical simulations are used to predict the performance characteristics of the process as well as to identify the optimal operational conditions within the space of system parameters. It is shown that the group velocity dispersion of the fiber is not detrimental and, in fact, helps achieve optimum compression. We also demonstrate that near-transform-limited rectangular and parabolic pulses can be generated in the region of optimum compression.
Resumo:
En exploitant une modulation de phase sinusoïdale temporelle additionnelle, nous montrons qu’il est possible d’améliorer significativement les performances d’une compression spectrale réalisée en régime de propagation hautement non-linéaire. Les simulations numériques indiquent ainsi une amélioration des facteurs de compression ainsi que du rapport de Strehl.
Resumo:
The contributions of this dissertation are in the development of two new interrelated approaches to video data compression: (1) A level-refined motion estimation and subband compensation method for the effective motion estimation and motion compensation. (2) A shift-invariant sub-decimation decomposition method in order to overcome the deficiency of the decimation process in estimating motion due to its shift-invariant property of wavelet transform. ^ The enormous data generated by digital videos call for an intense need of efficient video compression techniques to conserve storage space and minimize bandwidth utilization. The main idea of video compression is to reduce the interpixel redundancies inside and between the video frames by applying motion estimation and motion compensation (MEMO) in combination with spatial transform coding. To locate the global minimum of the matching criterion function reasonably, hierarchical motion estimation by coarse to fine resolution refinements using discrete wavelet transform is applied due to its intrinsic multiresolution and scalability natures. ^ Due to the fact that most of the energies are concentrated in the low resolution subbands while decreased in the high resolution subbands, a new approach called level-refined motion estimation and subband compensation (LRSC) method is proposed. It realizes the possible intrablocks in the subbands for lower entropy coding while keeping the low computational loads of motion estimation as the level-refined method, thus to achieve both temporal compression quality and computational simplicity. ^ Since circular convolution is applied in wavelet transform to obtain the decomposed subframes without coefficient expansion, symmetric-extended wavelet transform is designed on the finite length frame signals for more accurate motion estimation without discontinuous boundary distortions. ^ Although wavelet transformed coefficients still contain spatial domain information, motion estimation in wavelet domain is not as straightforward as in spatial domain due to the shift variance property of the decimation process of the wavelet transform. A new approach called sub-decimation decomposition method is proposed, which maintains the motion consistency between the original frame and the decomposed subframes, improving as a consequence the wavelet domain video compressions by shift invariant motion estimation and compensation. ^
Resumo:
The focus of this thesis is placed on text data compression based on the fundamental coding scheme referred to as the American Standard Code for Information Interchange or ASCII. The research objective is the development of software algorithms that result in significant compression of text data. Past and current compression techniques have been thoroughly reviewed to ensure proper contrast between the compression results of the proposed technique with those of existing ones. The research problem is based on the need to achieve higher compression of text files in order to save valuable memory space and increase the transmission rate of these text files. It was deemed necessary that the compression algorithm to be developed would have to be effective even for small files and be able to contend with uncommon words as they are dynamically included in the dictionary once they are encountered. A critical design aspect of this compression technique is its compatibility to existing compression techniques. In other words, the developed algorithm can be used in conjunction with existing techniques to yield even higher compression ratios. This thesis demonstrates such capabilities and such outcomes, and the research objective of achieving higher compression ratio is attained.
Resumo:
This research aims to systematize a proposal of developing a mobile tablet application in order to help implementing the Semantic Differential technique – SD, under the approach of Participatory Design. In 1975, Osgood et al. created the Semantic Differential technique. Since then, many experiments use it to measure the affective perception of individuals concerning objects and concepts by means of compounded scales of bipolar adjectives, based on the theoretical models that support the technique: the conductible, spatial and metric models. During the application of the technique with potential users, the researcher must simultaneously manage several contexts, that is, audio recorder, when authorized, and observe and record spontaneous reports of the respondent. It is noticeable that often occurs a cognitive overload during this event. Thus, the use of a single application whose interface is assigned to its users and respondents could assist researchers in applying the SD technique. This research aimed to understand the processes inherent to the task of implementing the Semantic Differential technique and obeyed the following steps: a) training of users, b) background questionnaire c) interview with Focus Group, and d) cooperative evaluation. Besides these procedures, one can also observe the degrees of facilitation or difficulty concerning the use of the conventional model, which is the development and application of scales with the aid of printed material, pencil, pens, clipboards, and recorder software for editing the document and data analysis. This paper comprises reactions and impressions from the experiences of users of SD technique. Considering the data recollected from the user’s observation, the hypothesis of the experiment proved to be right. It means that the development of the application for mobile tablet employing the technique of Semantic Differential is viable, since it assembles all the steps in one only tool, increases the accomplishment of the task between user/researcher and user/respondent resulting in their mutual satisfaction.
Resumo:
Medical imaging technologies are experiencing a growth in terms of usage and image resolution, namely in diagnostics systems that require a large set of images, like CT or MRI. Furthermore, legal restrictions impose that these scans must be archived for several years. These facts led to the increase of storage costs in medical image databases and institutions. Thus, a demand for more efficient compression tools, used for archiving and communication, is arising. Currently, the DICOM standard, that makes recommendations for medical communications and imaging compression, recommends lossless encoders such as JPEG, RLE, JPEG-LS and JPEG2000. However, none of these encoders include inter-slice prediction in their algorithms. This dissertation presents the research work on medical image compression, using the MRP encoder. MRP is one of the most efficient lossless image compression algorithm. Several processing techniques are proposed to adapt the input medical images to the encoder characteristics. Two of these techniques, namely changing the alignment of slices for compression and a pixel-wise difference predictor, increased the compression efficiency of MRP, by up to 27.9%. Inter-slice prediction support was also added to MRP, using uni and bi-directional techniques. Also, the pixel-wise difference predictor was added to the algorithm. Overall, the compression efficiency of MRP was improved by 46.1%. Thus, these techniques allow for compression ratio savings of 57.1%, compared to DICOM encoders, and 33.2%, compared to HEVC RExt Random Access. This makes MRP the most efficient of the encoders under study.
Resumo:
Medical imaging technology and applications are continuously evolving, dealing with images of increasing spatial and temporal resolutions, which allow easier and more accurate medical diagnosis. However, this increase in resolution demands a growing amount of data to be stored and transmitted. Despite the high coding efficiency achieved by the most recent image and video coding standards in lossy compression, they are not well suited for quality-critical medical image compression where either near-lossless or lossless coding is required. In this dissertation, two different approaches to improve lossless coding of volumetric medical images, such as Magnetic Resonance and Computed Tomography, were studied and implemented using the latest standard High Efficiency Video Encoder (HEVC). In a first approach, the use of geometric transformations to perform inter-slice prediction was investigated. For the second approach, a pixel-wise prediction technique, based on Least-Squares prediction, that exploits inter-slice redundancy was proposed to extend the current HEVC lossless tools. Experimental results show a bitrate reduction between 45% and 49%, when compared with DICOM recommended encoders, and 13.7% when compared with standard HEVC.
Resumo:
We propose two new approaches to enhance the spectral compression process arising from nonlinear pulse propagation in an optical fibre. We numerically show that an additional sinusoidal temporal phase modulation of the pulse enables efficient reduction of the intensity level of side lobes in the spectrum. Another strategy is to select a regime of propagation in which normal group-velocity dispersion reshapes the initial stretched pulse to a near-Fourier-transform-limited rectangular waveform.
Resumo:
Atomisation of an aqueous solution for tablet film coating is a complex process with multiple factors determining droplet formation and properties. The importance of droplet size for an efficient process and a high quality final product has been noted in the literature, with smaller droplets reported to produce smoother, more homogenous coatings whilst simultaneously avoiding the risk of damage through over-wetting of the tablet core. In this work the effect of droplet size on tablet film coat characteristics was investigated using X-ray microcomputed tomography (XμCT) and confocal laser scanning microscopy (CLSM). A quality by design approach utilising design of experiments (DOE) was used to optimise the conditions necessary for production of droplets at a small (20 μm) and large (70 μm) droplet size. Droplet size distribution was measured using real-time laser diffraction and the volume median diameter taken as a response. DOE yielded information on the relationship three critical process parameters: pump rate, atomisation pressure and coating-polymer concentration, had upon droplet size. The model generated was robust, scoring highly for model fit (R2 = 0.977), predictability (Q2 = 0.837), validity and reproducibility. Modelling confirmed that all parameters had either a linear or quadratic effect on droplet size and revealed an interaction between pump rate and atomisation pressure. Fluidised bed coating of tablet cores was performed with either small or large droplets followed by CLSM and XμCT imaging. Addition of commonly used contrast materials to the coating solution improved visualisation of the coating by XμCT, showing the coat as a discrete section of the overall tablet. Imaging provided qualitative and quantitative evidence revealing that smaller droplets formed thinner, more uniform and less porous film coats.
Resumo:
We propose a new, simple approach to enhance the spectral compression process arising from nonlinear pulse propagation in an optical fiber. We numerically show that an additional sinusoidal temporal phase modulation of the pulse enables efficient reduction of the intensity level of the side lobes in the spectrum that are produced by the mismatch between the initial linear negative chirp of the pulse and the self-phase modulation-induced nonlinear positive chirp. Remarkable increase of both the extent of spectrum narrowing and the quality of the compressed spectrum is afforded by the proposed approach across a wide range of experimentally accessible parameters.
Resumo:
We propose a simple approach to enhance the spectral compression arising from nonlinear pulse propagation in a Kerr medium. We numerically show that an additional sinusoidal temporal phase modulation enables efficient reduction of the intensity level of spectral side lobes.
Resumo:
A substantial amount of information on the Internet is present in the form of text. The value of this semi-structured and unstructured data has been widely acknowledged, with consequent scientific and commercial exploitation. The ever-increasing data production, however, pushes data analytic platforms to their limit. This thesis proposes techniques for more efficient textual big data analysis suitable for the Hadoop analytic platform. This research explores the direct processing of compressed textual data. The focus is on developing novel compression methods with a number of desirable properties to support text-based big data analysis in distributed environments. The novel contributions of this work include the following. Firstly, a Content-aware Partial Compression (CaPC) scheme is developed. CaPC makes a distinction between informational and functional content in which only the informational content is compressed. Thus, the compressed data is made transparent to existing software libraries which often rely on functional content to work. Secondly, a context-free bit-oriented compression scheme (Approximated Huffman Compression) based on the Huffman algorithm is developed. This uses a hybrid data structure that allows pattern searching in compressed data in linear time. Thirdly, several modern compression schemes have been extended so that the compressed data can be safely split with respect to logical data records in distributed file systems. Furthermore, an innovative two layer compression architecture is used, in which each compression layer is appropriate for the corresponding stage of data processing. Peripheral libraries are developed that seamlessly link the proposed compression schemes to existing analytic platforms and computational frameworks, and also make the use of the compressed data transparent to developers. The compression schemes have been evaluated for a number of standard MapReduce analysis tasks using a collection of real-world datasets. In comparison with existing solutions, they have shown substantial improvement in performance and significant reduction in system resource requirements.