4 resultados para Time delay
em Universidad de Alicante
Resumo:
Different non-Fourier models of heat conduction, that incorporate time lags in the heat flux and/or the temperature gradient, have been increasingly considered in the last years to model microscale heat transfer problems in engineering. Numerical schemes to obtain approximate solutions of constant coefficients lagging models of heat conduction have already been proposed. In this work, an explicit finite difference scheme for a model with coefficients variable in time is developed, and their properties of convergence and stability are studied. Numerical computations showing examples of applications of the scheme are presented.
Resumo:
Paper submitted to the XVIII Conference on Design of Circuits and Integrated Systems (DCIS), Ciudad Real, España, 2003.
Resumo:
Paper submitted to 10th IEEE International Conference on Electronics, Circuits and Systems (ICECS), Sharjah, Emiratos Árabes, 2003.
Resumo:
This paper deals with the estimation of a time-invariant channel spectrum from its own nonuniform samples, assuming there is a bound on the channel’s delay spread. Except for this last assumption, this is the basic estimation problem in systems providing channel spectral samples. However, as shown in the paper, the delay spread bound leads us to view the spectrum as a band-limited signal, rather than the Fourier transform of a tapped delay line (TDL). Using this alternative model, a linear estimator is presented that approximately minimizes the expected root-mean-square (RMS) error for a deterministic channel. Its main advantage over the TDL is that it takes into account the spectrum’s smoothness (time width), thus providing a performance improvement. The proposed estimator is compared numerically with the maximum likelihood (ML) estimator based on a TDL model in pilot-assisted channel estimation (PACE) for OFDM.