Abstract
Adaptive identification in a time-varying context is studied when controlled by the LMS algorithms with constant gain µ, under the assumption of correlated successive input vectors. It is well-known by experience that the tracking mean square error (MSE) ε(µ) results from the trade-off between the gradient part which is µ-increasing and the lag contribution which is µ-decreasing. In this paper we clarify the relative roles of the gradient and lag errors by proving their decoupled character. This property relies upon independence between the additive noise at the output of the plant to be identified and the information vector at the plant input. Convergence of the MSE is established rather than assumed. Quantitative evaluations of upper and lower bounds allow an approximate optimization of the gain. In two important cases the optimum is exact. One of these cases is "slow-variations". It is defined in a quantitative manner thanks to the ratio of the "variation"-noise to the output additive noise.