Performance Analysis of the LMS Algorithm with a Tapped Delay Line (Two-Dimensional Case)



Similar documents
Analysis of Mean-Square Error and Transient Speed of the LMS Adaptive Algorithm

Background 2. Lecture 2 1. The Least Mean Square (LMS) algorithm 4. The Least Mean Square (LMS) algorithm 3. br(n) = u(n)u H (n) bp(n) = u(n)d (n)

4F7 Adaptive Filters (and Spectrum Estimation) Least Mean Square (LMS) Algorithm Sumeetpal Singh Engineering Department sss40@eng.cam.ac.

Adaptive Equalization of binary encoded signals Using LMS Algorithm

ADAPTIVE ALGORITHMS FOR ACOUSTIC ECHO CANCELLATION IN SPEECH PROCESSING

1 Short Introduction to Time Series

The Filtered-x LMS Algorithm

Kristine L. Bell and Harry L. Van Trees. Center of Excellence in C 3 I George Mason University Fairfax, VA , USA kbell@gmu.edu, hlv@gmu.

Scott C. Douglas, et. Al. Convergence Issues in the LMS Adaptive Filter CRC Press LLC. <

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.

Analysis of Load Frequency Control Performance Assessment Criteria

By choosing to view this document, you agree to all provisions of the copyright laws protecting it.

Final Year Project Progress Report. Frequency-Domain Adaptive Filtering. Myles Friel. Supervisor: Dr.Edward Jones

Adaptive Variable Step Size in LMS Algorithm Using Evolutionary Programming: VSSLMSEV

Log-Likelihood Ratio-based Relay Selection Algorithm in Wireless Network

Advanced Signal Processing and Digital Noise Reduction

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS

Information Theory and Stock Market

Analysis of Filter Coefficient Precision on LMS Algorithm Performance for G.165/G.168 Echo Cancellation

Solutions to Exam in Speech Signal Processing EN2300

A Regime-Switching Model for Electricity Spot Prices. Gero Schindlmayr EnBW Trading GmbH

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Relations Between Time Domain and Frequency Domain Prediction Error Methods - Tomas McKelvey

Probability and Random Variables. Generation of random variables (r.v.)

CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York

Characterization Of Polynomials Using Reflection Coefficients

Computer exercise 2: Least Mean Square (LMS)

State Space Time Series Analysis

Automatic Detection of Emergency Vehicles for Hearing Impaired Drivers

SYSTEMS OF REGRESSION EQUATIONS

Understanding and Applying Kalman Filtering

NUMERICAL ANALYSIS OF THE EFFECTS OF WIND ON BUILDING STRUCTURES

SECRET sharing schemes were introduced by Blakley [5]

5 Numerical Differentiation

Loudspeaker Equalization with Post-Processing

1646 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 6, NO. 12, DECEMBER 1997

Statistical Machine Learning

Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Ericsson T18s Voice Dialing Simulator

System Identification for Acoustic Comms.:

Sound Power Measurement

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

Load Balancing and Switch Scheduling

Formulations of Model Predictive Control. Dipartimento di Elettronica e Informazione

> plot(exp.btgpllm, main = "treed GP LLM,", proj = c(1)) > plot(exp.btgpllm, main = "treed GP LLM,", proj = c(2)) quantile diff (error)

3.3. Solving Polynomial Equations. Introduction. Prerequisites. Learning Outcomes

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

LMS is a simple but powerful algorithm and can be implemented to take advantage of the Lattice FPGA architecture.

Adaptive Notch Filter for EEG Signals Based on the LMS Algorithm with Variable Step-Size Parameter

OPTIMAl PREMIUM CONTROl IN A NON-liFE INSURANCE BUSINESS

Metric Spaces. Chapter Metrics

Hydrodynamic Limits of Randomized Load Balancing Networks

Maximum Likelihood Estimation of ADC Parameters from Sine Wave Test Data. László Balogh, Balázs Fodor, Attila Sárhegyi, and István Kollár

Cross Validation. Dr. Thomas Jensen Expedia.com

Nonlinear Model Predictive Control of Hammerstein and Wiener Models Using Genetic Algorithms

LINEAR ALGEBRA W W L CHEN

How to Win the Stock Market Game

METHODOLOGICAL CONSIDERATIONS OF DRIVE SYSTEM SIMULATION, WHEN COUPLING FINITE ELEMENT MACHINE MODELS WITH THE CIRCUIT SIMULATOR MODELS OF CONVERTERS.

Adaptive Sampling Rate Correction for Acoustic Echo Control in Voice-Over-IP Matthias Pawig, Gerald Enzner, Member, IEEE, and Peter Vary, Fellow, IEEE

Characterization and Modeling of Packet Loss of a VoIP Communication

System Aware Cyber Security

IMPROVED NETWORK PARAMETER ERROR IDENTIFICATION USING MULTIPLE MEASUREMENT SCANS

Kalman and Extended Kalman Filters: Concept, Derivation and Properties

SIMPLIFIED PERFORMANCE MODEL FOR HYBRID WIND DIESEL SYSTEMS. J. F. MANWELL, J. G. McGOWAN and U. ABDULWAHID

Introduction to Kalman Filtering

Hardware Implementation of Probabilistic State Machine for Word Recognition

Lecture 5: Variants of the LMS algorithm

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

On the Interaction and Competition among Internet Service Providers

Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data

IN current film media, the increase in areal density has

A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

Multivariate Normal Distribution

Financial TIme Series Analysis: Part II

(Quasi-)Newton methods

On the Traffic Capacity of Cellular Data Networks. 1 Introduction. T. Bonald 1,2, A. Proutière 1,2

Least Squares Approach for Initial Data Recovery in Dynamic

SLOT FRINGING EFFECT ON THE MAGNETIC CHARACTERISTICS OF ELECTRICAL MACHINES

International Journal of Information Technology, Modeling and Computing (IJITMC) Vol.1, No.3,August 2013

Roots of Polynomials

Should we Really Care about Building Business. Cycle Coincident Indexes!

Mixed Method of Model Reduction for Uncertain Systems

(Refer Slide Time: 01:11-01:27)

Time series analysis as a framework for the characterization of waterborne disease outbreaks

Breaking Generalized Diffie-Hellman Modulo a Composite is no Easier than Factoring

Modeling and Verification of Sampled-Data Hybrid Systems

Machine Learning and Pattern Recognition Logistic Regression

Optimum Resolution Bandwidth for Spectral Analysis of Stationary Random Vibration Data

CS 688 Pattern Recognition Lecture 4. Linear Models for Classification

An Introduction to Machine Learning

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Luigi Piroddi Active Noise Control course notes (January 2015)

Transcription:

1542 IEEE TRANSACTIONS ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL. ASSP-34, NO, 6, DECEMBER 1986 Performance Analysis of the LMS Algorithm with a Tapped Delay Line (Two-Dimensional Case) SHAUL FLORIAN AND ARIE FEUER Abstract-This paper presents exact analytical results of the LMS performances on correlated data vectors. The case addressed here is of data vectors constructed from a tapped delay line. While the results are for the two-dimensional case, they do provide indications of the effect this structure of data vectors, so common to LMS applications, has on LMS performance. They show conclusively the possibility of bias in weight vector mean value its source. There is also a clear increase in MSE values over the case with uncorrelated data a decrease of the convergence region of adaptation coefficient. The results presented here strongly suggest that the correlation inherent to the above-mentioned data vectors could be a factor in some applications. I. INTRODUCTION HE least mean square (LMS) is one of the most com- T monly used algorithms in adaptive signal processing. As a result, many researchers have investigated its convergence performance properties. However, to date, the analysis is still behind, the algorithm works for cases which are not supported by analytical results. The basic problem for which the LMS is designed can be described as follows. Sequences of n-dimensional vectors (X@)) scalars {d(k)) are measured. It is desired to find a weight vector W so that y(k) = W X(k) 0) is as close to d(k) as possible in the mean square error (MSE) sense. Namely, a W that will minimize ~(k) = E{e(k)2) = E([d(k) - WTX(k)12}. (2) Assuming that the measured data are stationary, the solution is well known of the form W* = R-IP (3) where R = E{X(k) X(k)T} (4) P = E{d(k) X(k)). (5) Manuscript received January 17, 1986; revised May 7, 1986. S. Florian is with RAFAEL, P.O. Box 2250, Haifa, Israel. A. Feuer is with the Electrical Engineering, Technion-Israel Institute of Technology, Haifa 32000, Israel. IEEE Log Number 8610385. E{ *] denotes the expected value of the term inside the braces. 0096-3518/86/1200-1542$01.00 O 1986 IEEE This optimal solution can also be obtained recursively through the steepest descent (SD) algorithm, provided R P are available. The LMS is applicable for the more common situation when R P are not available. It can be viewed as the simplest approximation of the SD algorithm has the form W(k + 1) = (I - 2pX(k) X(k)*) W(k) + 2pd(k) X(k) (6) where W(k) is a vector of weights updated every sampling interval p > 0 is a constant parameter which determines the step size. To be able to carry out a complete analysis, resulting in convergence conditions for both the E{ W(k)) the MSE as well as quantitative performance investigation, the sequence (X(k)) had to be assumed statistically independent in time (see, e.g., [1]-[4]). Unfortunately, in most applications of this algorithm this assumption does not hold, then the results available are considerably more limited. References [5]-[8] provide a sample of the existing results when the above assumption does not hold. Generally, these results either depend on assuming a small step size p or are in the form of bounds on the algorithm performance. We refer the reader to [8] for a more detailed overview of previous results. In this paper, we try to isolate one main source of dependence in the sequence {X(k)} in many applications. We consider the case where X(k) is constructed from a tapped delay line (TDL), namely, X(k) = [x@), x(k - 1), * * *, x(k - n + l)lt (7) where {x(k)) is a scalar i.i.d. sequence, Gaussian, with zero mean variance 02. We note that even with this independence assumption on (x(k)) we have 0 0 *. * 0 0 I.. 0 0 E{X(k) X(k - 1) ) = :2 0: # 0... 0 0... o2 Hence, there is a clear correlation in {X(k)) which is inherent in the structure (7), the effects of this correlation are the subject of our investigation. The analysis is still very difficult, even the conver-

FLORIAN AND FEUER: PERFORMANCE ANALYSIS OF LMS ALGORITHM 1543 gence of E( W(k)} to the optimal value W* is not guaranteed, as will be shown in the sequel. However, we limit our discussion to the two-dimensional case, through which we introduce an approach that may be applicable to higher dimensions. For this case, we derive the exact equations governing the behavior of the E( W(k)} the MSE. From these equations, we observe that while the transient behavior of E{ W(k)} is similar to the case when {X(k)} are independent, there is here a possibility for bias in the steady state. The behavior of the MSE is different both in its transient in its steady-state values. Some simulation experiments were performed to verify the analytical results, a sample of their results is presented here. 11. CONVERGENCE OF WEIGHT MEAN VALUES Recall that we are dealing with a two-dimensional case, namely, Denoting ~,(k) = ~ ( k -) wf, i = I, 2 (9) substituting both (8) (9) in (6), we get V&k + 1) = (1-2px(k)2) V,(k) - 2px(k) x(k - 1) V2(k) + 2pe*(k) x(k) (10) V2(k + 1) = -2px(k) x(k - 1) V,(k) + (1-2px(k - 1)2) V&) + 2pe*(k) x(k - 1) (1 1) where e*(k) = d(k) - x(~)~w*. (12) Since {x@)} is i.i.d., we observe that both Vl(k) V2(k) are independent of x(@. It can also be readily verified that E{" * ( k) X@)} = 0; (13) hence, taking the expected value of both sides of (IO) (1 1) will result in E{ Vl(k + I)} = (1-2pa2) E{ Vl(k)} (14) E{ V2(k + I)} = E{ V2(k)} - 2pE{x(k - 1)2 V2(k)}. Let us now define then from (1 1) we get (15) V,(k) = x(k - 1)2 V&); (16) v5(k + 1) = 2px(k)3 x(k - 1) ~,(k) + (1-2px(k - V,(k)! + 2pe*(k) x(k - I), (17) or by taking the expected value on both sides, we have where A = [; ;2 + 2p r" 0 LE{e*(k) 1-2pa2 0 0-2p j* -2p.a x(k This equation will converge if all eigenvalues of A are inside the unit circle. The eigenvalues of A are 0, 1-2p.a2, 1-2p.a2, so the convergence is guaranteed if 2 y = po < 1. (20) This condition is identical to the condition derived when {X@)} are assumed independent. However, the steadystate values here will be E{Vl(+ = 0 2P E{V2(oo)) = -7 E{e*(k) x(k - I)}, (21) 0 so whether E{W2(k)} will converge to W: depends, on whether E{e*(k) ~ ( k x(k ) ~ - 1)) is equal to zero or not. In many cases, it is equal to zero [e.g., (d(k), x(k)) jointly Gaussian], but we will show later a case where it is not zero. III. CONVERGENCE OF WEIGHT VARIANCES AND MSE As was argued in [2] [4], convergence of the mean does not guarantee finite variances, hence, is not sufficient to guarantee algorithm convergence. To investigate the behavior of weight variances, we square both sides of (10) (1 I), take expected values, make use of the independence between x(k) V,(k), V2(k) to get E{Vl(k + l)'} = (1-4pa2 + 12p2.a4) E{ Vl(k)2} + 4p202E{x(k - 1)2 V2(k)2} + 4p E{e*(k)2 + 4pE{(1-2pm2> e*"() V,(k)} - 8 ~~E{x(k)~ - 1) e*(k) V2(k)} (22)

1544 IEEE TRANSACTIONS ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL. ASSP-34, NO. 6, DECEMBER 1986 E{V2(k + l) } = 4p2a2E{x(k - 1) Vl(k) > V,(k) = ~ (- k 1) Vl(k) + E{V (k) } - 4pE{x(k - 1) VZ(k) } V,(k) = x(k - 1) V (k) use (10) (11) to get + 4p2E{x(k - 1)4 V (k) } + 4p2E{x(k - 1) e*(@ } - 8p2E{x(k) x(k - 1) e*(k) Vl(k)> + 4pE{(1-2/4k - 1) ) * x(k - 1) e*@) V2(k)3. (23) Comparing this to similar developments in [2] or [4], we note a considerable complication due to the correlations V3(k + 1) = x(k)(l - 2px(k) > V,(k) - 2px(k) V,(k) + 2pe*(k) x(k) (24) V4(k + 1) = -2px(k) V3(k) + ~ (k) (1-2px(k - 1) ) - V2(k) + 2pe*(k) x(k) x(k - 1). (25) Now, squaring (17), (24), (25), taking the expected values, using the independence of x&), we get, together with (22) (23), present here. To proceed, we define U(k + 1) = FU(k) + 4p2g + 4pb(k) - 8p2c(k) (26) where F = a (1-12p.a + 60p2a4) I [k) }, E{ V5(k > >I 4p2a2 1 0-4P I 4P 12P2041 0-4p 1 4p2a2-12p.$ 1 12p2a4 g = b(k) = c(k) = E(X(~)~ e*(k) VI@)} + E {x(~)~ x(k - 1) e* (k) V2(k)} E{x(k) x(k - 1) e*(k) Vl(k)> + E{x(k - 1)3 e*&) V2(k)) E{x(~)~ e*@) Vl(k)) + E {x(~)~ x(k - 1) e* (k) V2(k)} E{x(Q3 x(k - 1) e*@) V,(k)} + E{x(k) x(k - 1)3 e*(k) V2(k)>.E{.x(~)~ x(k - 1) e*@) VI@)} + E{x(k)4 x(k - 1)3 e*@) V2(k)]

FLORIAN AND FEUER: PERFORMANCE ANALYSIS OF LMS ALGORITHM 1545 Assuming that g, b(k), c(k) are bounded, the boundedness of U(k) will depend on the eigenvalues of F; they must dl be inside the unit circle to guarantee stability. The characteristic equation of F is det (a1 - F) = a[a4-2(1-4y + 12y2) a3 + (1-4y)(1-47 + 24y2)a2 + 192y4(1-12y2) 01-64y4(l - 24y2 + 144y4)1 = 0 2 where y = pa. Using root locus2 methods (with y as a parameter), the condition for all roots to be inside the unit circle was found td be o < < 0.20i16. (31) quence with zero mean variance of, independent of (x@)} [recall that we assume x@) to be Gaussian (0, 02)]. With the above, clearly w* = E;] e*(k) = n(k). This (1 1) imply that e*(k) V2(k) are independent. Using all this (28)-(30), we get Then from (33) we get where from (26) loj 2 g=oe* b(k) = c(k) = 0. lo2 I L3a4 1 E(=)) = E* + a2u1(03) + U3(03) (36) U(03) = 4p2(1 - F)-'g. (37) Substituting (27), (35), (37) in (36) will result in (1-2y)2LT; (a) == 1-6y + 8y2-4874 - 57v' (38) w'ni'le tor the independent {Xk) one would get (1-2y>;a; :(a) =. 1-4y. (39) It can be observed that (39) approximates (38) for sufficiently small y [neglecting the y4 y6 terms in (38) will bring (382 to (39)l. In Fig. 1, we compare the results through the commonly used measure IV. STEADY-STATE MSE FOR Two SPECIAL CASES Case I: An Estimation Problem Let d(m = a&) + a2x(k - 1) + rzw (34) where al, a2 are fixed scalars {n(k)} is an i.i.d. se- 'One could also use Jury's stability test to get the same result. M, = (a) - E* E* ' often referred to as misadjustment. In addition, we picked an approximation presented in [7], which for the case we deal with wili result in Z(03) = (1 + 2y)oi, (40) the corresponding misadjustment is presented in Fig. 1. We observe that all three are close for small y. The approximation (40) becomes quite poor for large y. The comparison of (38) (39) clearly indicates that the inherent dependence of X(k) in (7) results, for a fixed y, in 4Etle:riorated skeady-slate performance (larger MSE in steady state).

IEEE TRANSACTIONS ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL. ASSP-34, NO. 6, DECEMBER 1986 E{ Vdk)} ~(k) = 2$ 6a2E{V5(k)} [{V6(k)8 6fJ2 E{ ] V6(k) } - -------- where we made use of the independence between x(k) Vl(k), V2(k) x(k - l), V5(k) is defined in (16) V6(k) = x(k - Vz(k). (47) -.I-J.-J -.w.02.oi.ffi.ob.io.i2,i4.i6.18.20.^2.)4.1g To find the steady-state MSE, we use (33), because r= pa2 of (45), (36) also applies here. However, to find U(m), we need to find E{V5(m)} E{V6(m)}. From (19), it Fig. 1. Comparison of misadjustment for Case 1. can easily be verified that Case 2: d(kj = x(kj x(k - 1) This case is constructed only for the purpose of demonstrating a bias in the convergence of the mean. Here E{V&q} = 0. (48) For V6(k), we use (1 1) (42) to get V6(k + 1) = -2p.~(k)~ V,(k) + ~ (k)~ [V,(k) - 2pV5(k)] + 2p [~(k)~ - x(~)~u ] x(k - 1) ; (41) taking expected values on both sides comparing to (18),we observe that Then Hence, from (48) we have [see (21)) E{x(k) e*@) x(k - l)} = 2u6 # 0 E(1/6(03)) = 12pcaS (49) b(m) = 0 Ll so the mean does not converge to the optimal solution. For the MSE, we get [see (28)-(30)] 1 g = 22 39u2 (44) Now, since from (26) b(m) = 0 L11704) U(m) = 4p2(Z - F)- [g - 2c(oo)],

~~ ~ FLORIAN AND FEUER: PERFORMANCE ANALYSIS OF LMS ALGORITHM 1547.- we get eventually $8 yo: Fig. 2. Comparison of misadjustment for Case 2 or 1-2y + 48y2-102y3 + 360y4,- 2167 M, = 1-67 + 8y2-48y4-576~~ 87. Again, for comparison, if we work under the independence assumption, we get Equations (51)-(53) are presented in Fig. 2. We can - 87 observe here considerably larger differences between the M, = ~ 1-4y (52) three curves at smaller y. which can be reached from (51) when, for small y, one V. SIMULATION RESULTS neglects all powers of y higher than 2. In our simulation experiments, we aimed to verify The approximation in [7] for this case will be support the analytical results conclusions of previous QS = 8y. (53) sections. The various experiments are described in the following table. Case 1 Case 2 TDL-LMS Unc.-LMS TDL-MOD2-LMS d(k) = x(k) + x(k - 1) + n(k) d(k) = x(k)2 x(k - 1) X(k) = pk)] Ir,(k)J x,(k), x,(k) uncorrelated x,(k), x2(k) uncorrelated d(k) = x,(k) + x2(k) + n(k) 4k) = XI(k) x2w Same as TDL-LMS except the weights are updated only every two samples. X(k) = p(k)] Lr,(k)J

1548 IEEE TRANSACTIONS ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL. ASSP-34, NO. 6, DECEMBER 1986 6 09 -.S.E 1.16 1.0.o '89- t 2.0 1.5 1.u - 0. 16. 20. 30. 40. 60. 60. 70. 80. 90. 100. 1111. ( C ) TDL-MOD2-LMS Fig. 3. MSE behavior for the three processings in Case 1. k Fig. 5. MSE behavior for the three processings in Case 2, p = 0.1. 4.0 - I I I I I I k M.S E 5.0 ( o 'DL-LMS p i. 0.05 rtx = 1.0 E{W2(k)? I (a) ua.05, 02=1.0 1.2 i Unc. -LMS M. S E TDL-LMS Fig. 4 7.n - I.R - I.O.B - - 1, I I I I I I I I I Ik,I. IO. PO. 31,. 40. 1.0. 611. 'X,. I///. lll. 11111. ( C 1 TDL-MODZ-LMS MSE behavior for the three processings in Case 2, p = 0.05. 0 10 20 30 40 50 60 70 (b)!~=0.1, 02=1.0 Fig. 6. E{ W,(k)) behavior for the three processings in Case 2.

FLORIAN AND FEUER: PERFORMANCE ANALYSIS OF LMS ALGORITHM 1549-50. M.S.E E exact results for it in the literature. For the two-dimen- 40. - sional case, we have shown that the correlation not does 35. - affect the rate of weight mean convergence, but may cause 90. - I I I 2s. - (maybe for pathological cases only) bias in steady- the 20. - state values. For the MSE, the we cor- have 15. - relation in the data results in higher MSE in the steady state, while, again, the transient behavior is hardly af- 0. fected. 0. IO. 20. 30. 40. 50. M. IO. 60. In conclusion, analyzing the LMS with correlated data ( a ) IDL-LMS M.S.E under the assumption that this correlation does not exist, lo* as was done in a number of results in the literature, may give a misleadingly optimistic performance picture that 0: = 1.0 Q = 1.0 6. does not fit the reality. 3.62 I. - 2. - M.S.E 1 1. 6. A ( b ) Unc.-LMS ::r, 0. -1-0. IO. 20. 90. 40. MI. 60. M. 00. ( c ) TDL-MODZ-LMS Fig. 7. Verification of stability bound (Case 1). The experiments on Case 1 showed identical weight mean behavior in the TDL-LMS the Unc.-LMS twice slower convergence with the TDL-MOD2-LMS. In Fig. 3 we present the MSE behavior for this case. We observe that, as expected, there is a difference in steadystate values for the TDL-LMS the Unc.-LMS, with the first being higher. The steady-state value for the TDL- MOD2-LMS is equal to the one for the Unc.-LMS, but its convergence is slower. The same behavior of the MSE can be observed in the experiments on Case 2 for different choices of p in Figs. 4 5. But here we can observe the bias in the weight mean as shown in Fig. 6. However, one should note that if the weight update is done every two samples, namely, in the TDL-MOD2-LMS, the bias disappears (with the tradeoff in slower convergence). Finally, in Fig. 7 we demonstrate that a choice y = 0.21 causes the TDL-LMS to diverge, while the other two still converge. VI. CONCLUSION In this paper, we have presented exact analytical results for the LMS algorithm applied to special types of correlated data vectors-constructed from a TDL. This is a very common case in practical applications, but there are no REFERENCES [l] B. Widrow, J. R. Glover, J. M. McCool, J. Kaunitz, C. S. Williams, R. H. Hearn, J. R. Zeidler, E. Dong, R. C. Goodin, Adaptive noise cancellation: Principles applications, Proc. IEEE, vol. 63, pp. 1692-1716, 1975. [21 L. L. Horowitz K. S. Senne, Performance advantage of complex LMS for controlling narrow-b adaptive arrays, IEEE Trans. Acoust., Speech, Signal Processing, vol. ASSP-29, pp. 722-736, June 1981. [3] W. A. Gardner, Learning characteristics of stochastic-gradient-descent algorithm: A general study, analysis critique, Signal Processing, No. 6. Amsterdam, The Netherls: North-Holl, 1984, pp. 113-133. [4] A. Feuer E. Weinstein, Convergence analysis of LMS filters with uncorrelated Gaussian data, IEEE Trans. Acoust., Speech, Signal Processing, vol. ASSP-33, pp. 222-230, Feb. 1985. [5] T. P. Daniell, Adaptive estimation with mutually collated training sequences, IEEE Trans. Syst. Sci. Cybern., vol. SSC-6, pp. 12-19, Jan. 1970. [6] D. C. Farden, Tracking properties of adaptive signal processing algorithms, IEEE Trans. Acoust., Speech, Signal Processing, vol. ASSP-29, pp. 439-446, June 1981. 171 S. K. Jones, R. K. Cavin, W. M. Reed, Analysis of error-gradient adaptive linear estimators for a class of stationary dependent pro- cesses, IEEE Trans. Inform. Theory, vol. IT-28, pp. 318-329, Mar. 1982. [8] 0. M. Macchi E. Eweda, Second-order convergence analysis of stochastic adaptive linear filtering, IEEE Trans. Automat. Contr., vol. AC-28, pp. 76-85, Jan. 1983. Shaul Florian received the B.Sc. M.Sc. degrees from the Technion-Israel Institute of Technology, Haifa, Israel, in 1977 1986, respectively. Since 1977 he has been employed by RAFAEL (Armament Development Authority) doing research on optimal waveform design implementation for EW systems. His current research interests are in adaptive algorithms optimal signal processing. Arie Feuer received the B.Sc. M.Sc. degrees from the Technion-Israel Institute of Technology, Haifa, Israel, in 1967 1973, respectively, the Ph.D. degree from Yale University, New Haven, CT, in 1978. From 1967 to 1970 he was employed by Technomatic-Israel, working on factory automation. From 1978 through 1983 he was with Bell Laboratories, Holmdel, NJ, studying telephone network performance. Since 1983 he has been on the Faculty of the Department of Electrical Engineering, Technion-Israel Institute of Technology. His research interests are in adaptive control adaptive signal processing.