new adaptive filtering part iice.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4....

48
Adaptive Filtering Part II

Upload: others

Post on 13-Sep-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering

Part II

Page 2: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

2 Adaptive Filtering, Part2

In previous Lecture we saw that:

Setting the gradient of cost function equal to zero, we obtain the optimum values of filter coefficients:

(Wiener-Hopf equation)

Page 3: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Method of Steepest Descent

• As shown in Figure the MSE is a quadratic function of the weights that can be pictured as a positive-concave hyper-parabolic surface.

• Adjusting the weights to minimize the error involves descending along this surface until reaching the 'bottom of the bowl.‘

• Various gradient-based algorithms are available. These algorithms are based on making local estimates of the gradient and moving downward toward the bottom of the bowl.

• The selection of an algorithm is usually decided by the speed of convergence, steady-state performance, and the computational complexity.

3 Adaptive Filtering, Part2

Page 4: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• The steepest-descent method reaches the minimum by following the direction in which the performance surface has the greatest rate of decrease.

• The steepest-descent method is an iterative (recursive) technique that starts from some initial (arbitrary) weight vector.

• Let ξ(0)represent the value of the MSE at time n =0 with an arbitrary choice of the weight vector w(0).

• The steepest-descent technique enables us to descend to the bottom of the bowl, wo, in a systematic way.

• The idea is to move on the error surface in the direction of the tangent at that point.

• The weights of the filter are updated at each iteration in the direction of the negative gradient of the error surface

4 Adaptive Filtering, Part2

Page 5: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Each selection of a filter weight vector w(n) corresponds to only one point on the MSE surface, [w(n), ξ(n)].

• Suppose that an initial filter setting w(0) on the MSE surface, [w(0),ξ(0)] is arbitrarily chosen.

• The gradient of the error surface \(n) is defined as the vector of these directional derivatives.

• The concept of steepest descent can be implemented as:

5 Adaptive Filtering, Part2

Page 6: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• where μ is a convergence factor (or step size) that controls stability and the rate of descent to the bottom of the bowl.

• The larger the value of μ, the faster the speed of descent.

• The vector denotes the gradient of the error function with respect to w(n), and the negative sign increments the adaptive weight vector in the negative gradient direction.

• The successive corrections to the weight vector in the direction of the steepest descent of the performance surface should eventually lead to the minimum.

Adaptive Filtering, Part2 6

Page 7: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

The LMS Algorithm

• The increment from w(n) to w(n +1) is in the negative gradient direction, so the weight tracking will closely follow the steepest descent path on the performance surface.

• However, in many practical applications the statistics of d(n) and x(n) are unknown.

• So, the method of steepest descent cannot be used directly, since it assumes exact knowledge of the gradient vector at each iteration.

Adaptive Filtering, Part2 7

Page 8: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Widrow used the instantaneous squared error, e 2(n), to estimate the MSE. That is:

• Therefore the gradient estimate used by the LMS algorithm is:

Adaptive Filtering, Part2 8

Page 9: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Since e(n)= d(n)-wT(n)x(n), so and the gradient estimate becomes

• Therefore the gradient estimate used by the LMS algorithm is :

• This is the well-known LMS algorithm, or stochastic gradient algorithm.

Adaptive Filtering, Part2 9

Page 10: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Summery of LMS 1. Determine L, μ, and w(0), where L is the

order of the filter, μ is the step size, and w(0) is the initial weight vector at time n =0.

2. Compute the adaptive filter output :

3. Compute the error signal:

e(n)=d(n)-y(n)

Adaptive Filtering, Part2 10

Page 11: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

4. Update the adaptive weight vector from by using the LMS algorithm:

Adaptive Filtering, Part2 11

Page 12: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Performance Analysis

• In this section, we present some important properties of the LMS algorithm such as

• stability,

• convergence rate,

• and the excess mean-square error due to gradient estimation error.

Adaptive Filtering, Part2 12

Page 13: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Stability Constraint • The LMS algorithm involves the presence of

feedback. Thus the algorithm is subject to the possibility of becoming unstable.

• μ controls the size of the incremental correction applied to the weight vector as we adapt from one iteration to the next.

• The mean weight-convergence of the LMS algorithm from initial condition w(0) to the optimum filter wo must satisfy:

Adaptive Filtering, Part2 13

Page 14: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• where λmax is the largest eigenvalue of the autocorrelation matrix R defined

• The computation of λmax is difficult, when L is large.

• In practical applications, it is desirable to estimate λmax using a simple method:

• where tr[R] denotes the trace of matrix R.

Adaptive Filtering, Part2 14

Page 15: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• It follows that:

• Where

• denotes the power of x(n). Therefore setting

• Assures the convergence of LMS

Adaptive Filtering, Part2 15

Page 16: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• This equation provides some important information on how to select μ:

1. Since the upper bound on μ is inversely proportional to L, a small μ is used for large-order filters.

2. Since μ is made inversely proportional to the input signal power, weaker signals use a larger μ and stronger signals use a smaller μ.

3. One useful approach is to normalize μ with respect to the input signal power Px. (normalized LMS).

Adaptive Filtering, Part2 16

Page 17: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Convergence of the weight vector w(n) from w(0) to wo

corresponds to the convergence of the MSE from ξ(0)to ξmin.

• Therefore convergence of the MSE toward its minimum value is a commonly performance-measurement in adaptive systems because of its simplicity.

• During adaptation, the squared error e2(n) is non-stationary as the weight vector w(n) adapts toward wo. The corresponding MSE can thus be defined only based on ensemble averages.

• A plot of the MSE versus time n is referred to as the learning curve for a given adaptive algorithm.

• Since the MSE is the performance criterion of LMS algorithms, the learning curve is a natural way to describe the transient behavior.

Adaptive Filtering, Part2 17

Page 18: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Each adaptive mode has its own time constant, which is determined by the overall adaptation constant μ and the eigenvalue λl associated with that mode.

• Overall convergence is clearly limited by the slowest mode. Thus the overall MSE time constant can be approximated as:

Adaptive Filtering, Part2 18

Page 19: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• a small λmin can result in a large time constant (i.e., a slow convergence rate).

• Unfortunately, if λmax is also very large, the selection of μ will be limited by such that only a small μ can satisfy the stability constraint.

• Therefore if λmax is very large and λmin is very small, the time constant can be very large.

• But the fastest convergence of the dominant mode occurs for μ =1/λmax; so

Adaptive Filtering, Part2 19

Page 20: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• the eigenvalues λmin and λmax are very difficult to compute. However, there is an efficient way to estimate the eigenvalue spread from the spectral dynamic range:

• RESULT: input signals with a flat (white) spectrum have the fastest convergence speed.

Adaptive Filtering, Part2 20

Page 21: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Excess Mean-Square Error • The steepest-descent requires knowledge of which

must be estimated at each iteration. • The estimated gradient produces the gradient

estimation+noise. After the algorithm converges, i.e., w(n) is close to wo, the true gradient is almost zero. However, the gradient estimator is not equal to zero.

• Thus the gradient estimation noise prevents w(n + 1) from staying at wo in steady state.

• The result is that it causes ξ(n) to be larger than its minimum value, thus producing excess noise at the filter output.

• The excess MSE, which is caused by random noise in the weight vector after con­vergence, can be approximated as:

Adaptive Filtering, Part2 21

Page 22: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 22

Page 23: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• the excess MSE is directly proportional to μ. The larger the value of μ, the worse the steady-state performance after convergence.

• However a larger μ results in faster convergence. • So, There is a design trade-off between the excess

MSE and the speed of convergence. • It is also proportional to the filter order L, which

means that a larger L results in larger algorithm noise.

• But a larger L implies a smaller μ, resulting in slower convergence.

• On the other hand, a large L also implies better filter characteristics such as sharp cutoff.

Adaptive Filtering, Part2 23

Page 24: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Normalized LMS Algorithm • One important technique to optimize the

speed of convergence while maintaining the desired steady-state performance, independent of the reference signal power, is known as the normalized LMS algorithm (NLMS). The NLMS algorithm is expressed as

• where μ(n) is an adaptive step size that is computed as

Adaptive Filtering, Part2 24

Page 25: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• where is an estimate of the power of x(n) at time n, and α is a normalized step size that satisfies the criterion:

Adaptive Filtering, Part2 25

Page 26: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive System Identification

Adaptive Filtering, Part2 26

Page 27: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Linear Prediction

Adaptive Filtering, Part2 27

Page 28: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Applications: Speech-Coding, separation of noise from signal

• The coefficients are updated as:

Adaptive Filtering, Part2 28

Page 29: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• Proper selection of the prediction delay allows improved frequency estimation performance

• In many digital communications and signal detection applications, the desired broad­band (spread-spectrum) signal is corrupted by an additive narrowband interference.

• narrowband characteristics of the interference allow W(z) to estimate and extract it from past samples of input.

Adaptive Filtering, Part2 29

Page 30: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Channel Equalization

Adaptive Filtering, Part2 30

Page 31: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• In theory, the delayed version of the transmitted signal, x(n -Δ), is the desired response for the adaptive equalizer W(z).

• However, is not available at the receiver. • During the training stage, the adaptive equalizer

coefficients are adjusted by a short training sequence.

• This known transmitted sequence is also generated in the receiver and is used as the desired signal.

• A widely used training signal consists of pseudo-random noise with a broad and flat power spectrum

Adaptive Filtering, Part2 31

Page 32: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

• The transmission of high-speed data through a channel is limited by inter-symbol interference (ISI) caused by distortion in the transmission channel.

• High-speed data transmission through channels with severe distortion can be achieved by an equalizer in the receiver that counteracts the channel distortion.

• Theoretically, the equalizer W(z), is the inverse of the channel transfer function:

Adaptive Filtering, Part2 32

Page 33: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Noise Cancellation

Adaptive Filtering, Part2 33

Page 34: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 34

Page 35: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 35

Page 36: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 36

Page 37: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 37

Page 38: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 38

Page 39: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 39

Page 40: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 40

Page 41: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 41

Page 42: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 42

Page 43: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 43

Page 44: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 44

Page 45: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 45

Page 46: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 46

Page 47: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 47

Page 48: New Adaptive Filtering Part IIce.sharif.edu/courses/93-94/1/ce763-2/resources/root... · 2015. 4. 6. · Adaptive Filtering, Part2 2 In previous Lecture we saw that: Setting the gradient

Adaptive Filtering, Part2 48