632 resultados para Decoding


Relevância:

20.00% 20.00%

Publicador:

Resumo:

We study the information rates of non-coherent, stationary, Gaussian, multiple-input multiple-output (MIMO) flat-fading channels that are achievable with nearest neighbour decoding and pilot-aided channel estimation. In particular, we analyse the behaviour of these achievable rates in the limit as the signal-to-noise ratio (SNR) tends to infinity. We demonstrate that nearest neighbour decoding and pilot-aided channel estimation achieves the capacity pre-logwhich is defined as the limiting ratio of the capacity to the logarithm of SNR as the SNR tends to infinityof non-coherent multiple-input single-output (MISO) flat-fading channels, and it achieves the best so far known lower bound on the capacity pre-log of non-coherent MIMO flat-fading channels. © 2011 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper studies a noncoherent multiple-input multiple-output (MIMO) fading multiple-access channel (MAC). The rate region that is achievable with nearest neighbour decoding and pilot-assisted channel estimation is analysed and the corresponding pre-log region, defined as the limiting ratio of the rate region to the logarithm of the signal-to-noise ratio (SNR) as the SNR tends to infinity, is determined. © 2011 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An achievable rate is given for discrete memoryless channels with a given (possibly suboptimal) decoding rule. The result is obtained using a refinement of the superposition coding ensemble. The rate is tight with respect to the ensemble average, and can be weakened to the LM rate of Hui and Csiszár-Körner, and to Lapidoth's rate based on parallel codebooks. © 2013 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

State-of-the-art speech recognisers are usually based on hidden Markov models (HMMs). They model a hidden symbol sequence with a Markov process, with the observations independent given that sequence. These assumptions yield efficient algorithms, but limit the power of the model. An alternative model that allows a wide range of features, including word- and phone-level features, is a log-linear model. To handle, for example, word-level variable-length features, the original feature vectors must be segmented into words. Thus, decoding must find the optimal combination of segmentation of the utterance into words and word sequence. Features must therefore be extracted for each possible segment of audio. For many types of features, this becomes slow. In this paper, long-span features are derived from the likelihoods of word HMMs. Derivatives of the log-likelihoods, which break the Markov assumption, are appended. Previously, decoding with this model took cubic time in the length of the sequence, and longer for higher-order derivatives. This paper shows how to decode in quadratic time. © 2013 IEEE.