76 resultados para soft-commutation technique
Resumo:
This paper presents an approach for automatic classification of pulsed Terahertz (THz), or T-ray, signals highlighting their potential in biomedical, pharmaceutical and security applications. T-ray classification systems supply a wealth of information about test samples and make possible the discrimination of heterogeneous layers within an object. In this paper, a novel technique involving the use of Auto Regressive (AR) and Auto Regressive Moving Average (ARMA) models on the wavelet transforms of measured T-ray pulse data is presented. Two example applications are examined - the classi. cation of normal human bone (NHB) osteoblasts against human osteosarcoma (HOS) cells and the identification of six different powder samples. A variety of model types and orders are used to generate descriptive features for subsequent classification. Wavelet-based de-noising with soft threshold shrinkage is applied to the measured T-ray signals prior to modeling. For classi. cation, a simple Mahalanobis distance classi. er is used. After feature extraction, classi. cation accuracy for cancerous and normal cell types is 93%, whereas for powders, it is 98%.
Resumo:
We describe a high-level design method to synthesize multi-phase regular arrays. The method is based on deriving component designs using classical regular (or systolic) array synthesis techniques and composing these separately evolved component design into a unified global design. Similarity transformations ar e applied to component designs in the composition stage in order to align data ow between the phases of the computations. Three transformations are considered: rotation, re ection and translation. The technique is aimed at the design of hardware components for high-throughput embedded systems applications and we demonstrate this by deriving a multi-phase regular array for the 2-D DCT algorithm which is widely used in many vide ocommunications applications.
Resumo:
A sparse kernel density estimator is derived based on the zero-norm constraint, in which the zero-norm of the kernel weights is incorporated to enhance model sparsity. The classical Parzen window estimate is adopted as the desired response for density estimation, and an approximate function of the zero-norm is used for achieving mathemtical tractability and algorithmic efficiency. Under the mild condition of the positive definite design matrix, the kernel weights of the proposed density estimator based on the zero-norm approximation can be obtained using the multiplicative nonnegative quadratic programming algorithm. Using the -optimality based selection algorithm as the preprocessing to select a small significant subset design matrix, the proposed zero-norm based approach offers an effective means for constructing very sparse kernel density estimates with excellent generalisation performance.
Resumo:
This preface outlines the reasons for undertaking the special issue, comments on the review process that was used, provides a brief summary of the papers included and discusses some of the implications of these papers on in vitro gas technique methodology and its applications.
Resumo:
Genetic algorithms (GAs) have been introduced into site layout planning as reported in a number of studies. In these studies, the objective functions were defined so as to employ the GAs in searching for the optimal site layout. However, few studies have been carried out to investigate the actual closeness of relationships between site facilities; it is these relationships that ultimately govern the site layout. This study has determined that the underlying factors of site layout planning for medium-size projects include work flow, personnel flow, safety and environment, and personal preferences. By finding the weightings on these factors and the corresponding closeness indices between each facility, a closeness relationship has been deduced. Two contemporary mathematical approaches - fuzzy logic theory and an entropy measure - were adopted in finding these results in order to minimize the uncertainty and vagueness of the collected data and improve the quality of the information. GAs were then applied to searching for the optimal site layout in a medium-size government project using the GeneHunter software. The objective function involved minimizing the total travel distance. An optimal layout was obtained within a short time. This reveals that the application of GA to site layout planning is highly promising and efficient.
Resumo:
This paper presents the evaluation in power consumption of a clocking technique for pipelined designs. The technique shows a dynamic power consumption saving of around 30% over a conventional global clocking mechanism. The results were obtained from a series of experiments of a systolic circuit implemented in Virtex-II devices. The conversion from a global-clocked pipelined design to the proposed technique is straightforward, preserving the original datapath design. The savings can be used immediately either as a power reduction benefit or to increase the frequency of operation of a design for the same power consumption.
Resumo:
A new technique is reported for micro-machining millimetre-wave rectangular waveguide components. S-parameter measurements on these structures show that they achieve lower loss than those produced using any other on-chip fabrication technique, have highly accurate dimensions, are physically robust, and are cheap and easy to manufacture.
Resumo:
This study proposes a conceptual model for customer experience quality and its impact on customer relationship outcomes. Customer experience is conceptualized as the customer’s subjective response to the holistic direct and indirect encounter with the firm, and customer experience quality as its perceived excellence or superiority. Using the repertory grid technique in 40 interviews in B2B and B2C contexts, the authors find that customer experience quality is judged with respect to its contribution to value-in-use, and hence propose that value-in-use mediates between experience quality and relationship outcomes. Experience quality includes evaluations not just of the firm’s products and services but also of peer-to-peer and complementary supplier encounters. In assessing experience quality in B2B contexts, customers place a greater emphasis on firm practices that focus on understanding and delivering value-in-use than is generally the case in B2C contexts. Implications for practitioners’ customer insight processes and future research directions are suggested.
Resumo:
The simulation and development work that has been undertaken to produce a signal equaliser used to improve the data rates from oil well logging instruments is presented. The instruments are lowered into the drill bore hole suspended by a cable which has poor electrical characteristics. The equaliser described in the paper corrects for the distortions that occur from the cable (dispersion and attenuation) with the result that the instrument can send data at 100 K.bits/second down its own suspension cable of 12 Km in length. The use of simulation techniques and tools were invaluable in generating a model for the distortions and proved to be a useful tool when site testing was not available.
Resumo:
The real time hardware architecture of a deterministic video echo canceller (deghoster) system is presented. The deghoster is capable of calculating all the multipath channel distortion characteristics from terrestrial and cable television in one single pass while performing real time video in-line ghost cancellation. The results from the actual system are also presented in this paper.
Resumo:
Water-soluble polymers are often capable of forming interpolymer complexes in solutions and at interfaces, which offers an excellent opportunity for surface modification. The complex formation may be driven by H-bonding between poly(carboxylic acids) and non-ionic polymers or by electrostatic attraction between oppositely-charged polyelectrolytes. In the present communication the following applications of interpolymer complexation in coating technologies will be considered: (1) Complexation between poly(acrylic acid) and non-ionic polymers via H-bonding was used to coat glass surfaces. It was realised using layer-by-layer deposition of IPC on glass surfaces with subsequent cross-linking of dry multilayers by thermal treatment. Depending on the glass surface functionality this complexation resulted in detachable and non-detachable hydrogel films; (2) Electrostatic layer-by-layer self-assembly between glycol chitosan and bovine serum albumin (BSA) was used to coat magnetic nanoparticles. It was demonstrated that the native structure of BSA remains unaffected by the self-assembling process.