Main Content

# Time-Delay Channel Estimation Through Adaptive Filtering

This example shows how to adaptively estimate the time delay for a noisy input signal using the LMS adaptive FIR algorithm.

Assume a signal where is a white Gaussian process and is deterministic. The signal is measured with an echo of samples and attenuation (both are unknown), resulting in the overall measurement:

The goal is to estimate the delay and the echo attenuation . One can determine these parameters by solving the filter identification problem for , combined with the prior . Provided that the filter can be identified from the measurements signal and the original signal , one can derive and

Such a filter identification problem can be posed in terms of adaptive LTI filtering. The reference signal is , the input feed is , and the adaptive filter is . Clearly, if the adaptation process concludes with then the error signal vanishes.

There are numerous adaptive filtering algorithms. For this paricular problem setup and signal model, the normalized LMS algorithm is suitable, and is available in the LMS Filter block.

Run the simulation. The peaks in the filter taps vector indicates the time-delay estimate. In this case and .

For details, see S. Haykin, Adaptive Filter Theory, 3rd Ed., Prentice-Hall 1996.

Get trial now