993 resultados para Decoding algorithm
Resumo:
Genetic algorithms (GAs) have been used to tackle non-linear multi-objective optimization (MOO) problems successfully, but their success is governed by key parameters which have been shown to be sensitive to the nature of the particular problem, incorporating concerns such as the numbers of objectives and variables, and the size and topology of the search space, making it hard to determine the best settings in advance. This work describes a real-encoded multi-objective optimizing GA (MOGA) that uses self-adaptive mutation and crossover, and which is applied to optimization of an airfoil, for minimization of drag and maximization of lift coefficients. The MOGA is integrated with a Free-Form Deformation tool to manage the section geometry, and XFoil which evaluates each airfoil in terms of its aerodynamic efficiency. The performance is compared with those of the heuristic MOO algorithms, the Multi-Objective Tabu Search (MOTS) and NSGA-II, showing that this GA achieves better convergence.
Resumo:
An achievable rate is given for discrete memoryless channels with a given (possibly suboptimal) decoding rule. The result is obtained using a refinement of the superposition coding ensemble. The rate is tight with respect to the ensemble average, and can be weakened to the LM rate of Hui and Csiszár-Körner, and to Lapidoth's rate based on parallel codebooks. © 2013 IEEE.
Resumo:
State-of-the-art speech recognisers are usually based on hidden Markov models (HMMs). They model a hidden symbol sequence with a Markov process, with the observations independent given that sequence. These assumptions yield efficient algorithms, but limit the power of the model. An alternative model that allows a wide range of features, including word- and phone-level features, is a log-linear model. To handle, for example, word-level variable-length features, the original feature vectors must be segmented into words. Thus, decoding must find the optimal combination of segmentation of the utterance into words and word sequence. Features must therefore be extracted for each possible segment of audio. For many types of features, this becomes slow. In this paper, long-span features are derived from the likelihoods of word HMMs. Derivatives of the log-likelihoods, which break the Markov assumption, are appended. Previously, decoding with this model took cubic time in the length of the sequence, and longer for higher-order derivatives. This paper shows how to decode in quadratic time. © 2013 IEEE.
Resumo:
Flow measurement data at the district meter area (DMA) level has the potential for burst detection in the water distribution systems. This work investigates using a polynomial function fitted to the historic flow measurements based on a weighted least-squares method for automatic burst detection in the U.K. water distribution networks. This approach, when used in conjunction with an expectationmaximization (EM) algorithm, can automatically select useful data from the historic flow measurements, which may contain normal and abnormal operating conditions in the distribution network, e.g., water burst. Thus, the model can estimate the normal water flow (nonburst condition), and hence the burst size on the water distribution system can be calculated from the difference between the measured flow and the estimated flow. The distinguishing feature of this method is that the burst detection is fully unsupervised, and the burst events that have occurred in the historic data do not affect the procedure and bias the burst detection algorithm. Experimental validation of the method has been carried out using a series of flushing events that simulate burst conditions to confirm that the simulated burst sizes are capable of being estimated correctly. This method was also applied to eight DMAs with known real burst events, and the results of burst detections are shown to relate to the water company's records of pipeline reparation work. © 2014 American Society of Civil Engineers.
Resumo:
The alternate combinational approach of genetic algorithm and neural network (AGANN) has been presented to correct the systematic error of the density functional theory (DFT) calculation. It treats the DFT as a black box and models the error through external statistical information. As a demonstration, the AGANN method has been applied in the correction of the lattice energies from the DFT calculation for 72 metal halides and hydrides. Through the AGANN correction, the mean absolute value of the relative errors of the calculated lattice energies to the experimental values decreases from 4.93% to 1.20% in the testing set. For comparison, the neural network approach reduces the mean value to 2.56%. And for the common combinational approach of genetic algorithm and neural network, the value drops to 2.15%. The multiple linear regression method almost has no correction effect here.
Resumo:
A novel approach for multi-dimension signals processing, that is multi-weight neural network based on high dimensional geometry theory, is proposed. With this theory, the geometry algorithm for building the multi-weight neuron is mentioned. To illustrate the advantage of the novel approach, a Chinese speech emotion recognition experiment has been done. From this experiment, the human emotions are classified into 6 archetypal classes: fear, anger, happiness, sadness, surprise and disgust. And the amplitude, pitch frequency and formant are used as the feature parameters for speech emotion recognition. Compared with traditional GSVM model, the new method has its superiority. It is noted that this method has significant values for researches and applications henceforth.
Resumo:
In this paper, we constructed a Iris recognition algorithm based on point covering of high-dimensional space and Multi-weighted neuron of point covering of high-dimensional space, and proposed a new method for iris recognition based on point covering theory of high-dimensional space. In this method, irises are trained as "cognition" one class by one class, and it doesn't influence the original recognition knowledge for samples of the new added class. The results of experiments show the rejection rate is 98.9%, the correct cognition rate and the error rate are 95.71% and 3.5% respectively. The experimental results demonstrate that the rejection rate of test samples excluded in the training samples class is very high. It proves the proposed method for iris recognition is effective.
Resumo:
A design algorithm of an associative memory neural network is proposed. The benefit of this design algorithm is to make the designed associative memory model can implement the hoped situation. On the one hand, the designed model has realized the nonlinear association of infinite value pattern from n dimension space to m dimension space. The result has improved the ones of some old associative memory neural network. On the other hand, the memory samples are in the centers of the fault-tolerant. In average significance the radius of the memory sample fault-tolerant field is maximum.
Resumo:
In this paper, a novel algorithm for removing facial makeup disturbances as a face detection preprocess based on high dimensional imaginal geometry is proposed. After simulation and practical application experiments, the algorithm is theoretically analyzed. Its apparent effect of removing facial makeup and the advantages of face detection with this pre-process over face detection without it are discussed. Furthermore, in our experiments with color images, the proposed algorithm even gives some surprises.
Resumo:
The quantum coherence control of a solid-state charge qubit is studied by using a suboptimal continuous feedback algorithm within the Bayesian feedback scheme. For the coherent Rabi oscillation, the present algorithm suggests a simple bang-bang control protocol, in which the control parameter is modulated between two values. For the coherence protection of the idle state, the present approach is applicable to arbitrary states, including those lying on the equator of the Bloch sphere which are out of control in the previous Markovian feedback scheme.
Resumo:
A novel image restoration approach based on high-dimensional space geometry is proposed, which is quite different from the existing traditional image restoration techniques. It is based on the homeomorphisms and "Principle of Homology Continuity" (PHC), an image is mapped to a point in high-dimensional space. Begin with the original blurred image, we get two further blurred images, then the restored image can be obtained through the regressive curve derived from the three points which are mapped form the images. Experiments have proved the availability of this "blurred-blurred-restored" algorithm, and the comparison with the classical Wiener Filter approach is presented in final.