V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline

Similar documents
Adaptive Filtering Part II

III.C - Linear Transformations: Optimal Filtering

EE482: Digital Signal Processing Applications

Adaptive Beamforming Algorithms

Adaptive Filter Theory

PMR5406 Redes Neurais e Lógica Fuzzy Aula 3 Single Layer Percetron

Machine Learning and Adaptive Systems. Lectures 3 & 4

Ch4: Method of Steepest Descent

26. Filtering. ECE 830, Spring 2014

EE482: Digital Signal Processing Applications

2.6 The optimum filtering solution is defined by the Wiener-Hopf equation

In the Name of God. Lecture 11: Single Layer Perceptrons

Linear Optimum Filtering: Statement

ADAPTIVE FILTER THEORY

Computer exercise 1: Steepest descent

Revision of Lecture 4

Optimal and Adaptive Filtering

Microphone-Array Signal Processing

Adaptive Systems Homework Assignment 1

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.

INTRODUCTION Noise is present in many situations of daily life for ex: Microphones will record noise and speech. Goal: Reconstruct original signal Wie

SIMON FRASER UNIVERSITY School of Engineering Science

Single layer NN. Neuron Model

Least Mean Square Filtering

Least Mean Squares Regression. Machine Learning Fall 2018

An Adaptive Sensor Array Using an Affine Combination of Two Filters

SNR lidar signal improovement by adaptive tecniques

Generalization. A cat that once sat on a hot stove will never again sit on a hot stove or on a cold one either. Mark Twain

Ch5: Least Mean-Square Adaptive Filtering

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)

IS NEGATIVE STEP SIZE LMS ALGORITHM STABLE OPERATION POSSIBLE?

3.4 Linear Least-Squares Filter

Linear Neural Networks

The Conjugate Gradient Method

Adaptive Filters. un [ ] yn [ ] w. yn n wun k. - Adaptive filter (FIR): yn n n w nun k. (1) Identification. Unknown System + (2) Inverse modeling

Statistical and Adaptive Signal Processing

ECE580 Partial Solution to Problem Set 3

Special Topics: Data Science

MMSE System Identification, Gradient Descent, and the Least Mean Squares Algorithm

Chapter 2 Wiener Filtering

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz

Performance Comparison of Two Implementations of the Leaky. LMS Adaptive Filter. Scott C. Douglas. University of Utah. Salt Lake City, Utah 84112

Machine Learning and Adaptive Systems. Lectures 5 & 6

ADAPTIVE FILTER THEORY

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece.

ESE 531: Digital Signal Processing

EEL 6502: Adaptive Signal Processing Homework #4 (LMS)

A Derivation of the Steady-State MSE of RLS: Stationary and Nonstationary Cases

LMS and eigenvalue spread 2. Lecture 3 1. LMS and eigenvalue spread 3. LMS and eigenvalue spread 4. χ(r) = λ max λ min. » 1 a. » b0 +b. b 0 a+b 1.

Adaptive SP & Machine Intelligence Linear Adaptive Filters and Applications

Recursive Least Squares for an Entropy Regularized MSE Cost Function

Adaptive MMSE Equalizer with Optimum Tap-length and Decision Delay

Artificial Neural Network

In this section again we shall assume that the matrix A is m m, real and symmetric.

Advanced Signal Processing Adaptive Estimation and Filtering

1. A discrete-time recurrent network is described by the following equation: y(n + 1) = A y(n) + B x(n)

Numerical Optimization: Basic Concepts and Algorithms

A FAST AND ACCURATE ADAPTIVE NOTCH FILTER USING A MONOTONICALLY INCREASING GRADIENT. Yosuke SUGIURA

ELEG-636: Statistical Signal Processing

Lecture 3: Linear FIR Adaptive Filtering Gradient based adaptation: Steepest Descent Method

MITIGATING UNCORRELATED PERIODIC DISTURBANCE IN NARROWBAND ACTIVE NOISE CONTROL SYSTEMS

ELEG-636: Statistical Signal Processing

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

COMP-4360 Machine Learning Neural Networks

CS545 Contents XVI. Adaptive Control. Reading Assignment for Next Class. u Model Reference Adaptive Control. u Self-Tuning Regulators

AdaptiveFilters. GJRE-F Classification : FOR Code:

Gradient Descent. Dr. Xiaowei Huang

Ch6-Normalized Least Mean-Square Adaptive Filtering

CS545 Contents XVI. l Adaptive Control. l Reading Assignment for Next Class

Neural networks III: The delta learning rule with semilinear activation function

Optimal and Adaptive Filtering

inear Adaptive Inverse Control

System Identification and Adaptive Filtering in the Short-Time Fourier Transform Domain

Massoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39

Least Mean Squares Regression

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

Variable Learning Rate LMS Based Linear Adaptive Inverse Control *

A METHOD OF ADAPTATION BETWEEN STEEPEST- DESCENT AND NEWTON S ALGORITHM FOR MULTI- CHANNEL ACTIVE CONTROL OF TONAL NOISE AND VIBRATION

ECE 680 Modern Automatic Control. Gradient and Newton s Methods A Review

Adap>ve Filters Part 2 (LMS variants and analysis) ECE 5/639 Sta>s>cal Signal Processing II: Linear Es>ma>on

HST.582J/6.555J/16.456J

Dominant Pole Localization of FxLMS Adaptation Process in Active Noise Control

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

Neuro-Fuzzy Comp. Ch. 4 March 24, R p

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group

Distributed MAP probability estimation of dynamic systems with wireless sensor networks

Comparative Performance Analysis of Three Algorithms for Principal Component Analysis

Adaptive Noise Cancellation

SGN Advanced Signal Processing: Lecture 4 Gradient based adaptation: Steepest Descent Method

The goal of the Wiener filter is to filter out noise that has corrupted a signal. It is based on a statistical approach.

Lecture 3: QR-Factorization

2262 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 47, NO. 8, AUGUST A General Class of Nonlinear Normalized Adaptive Filtering Algorithms

Lecture Notes in Adaptive Filters

Sparse Least Mean Square Algorithm for Estimation of Truncated Volterra Kernels

Conjugate Gradient Method

New Recursive-Least-Squares Algorithms for Nonlinear Active Control of Sound and Vibration Using Neural Networks

A Low-Distortion Noise Canceller and Its Learning Algorithm in Presence of Crosstalk

17 Solution of Nonlinear Systems

CS540 Machine learning Lecture 5

CHAPTER 4 ADAPTIVE FILTERS: LMS, NLMS AND RLS. 4.1 Adaptive Filter

Transcription:

V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline Goals Introduce Wiener-Hopf (WH) equations Introduce application of the steepest descent method to the WH problem Approximation to the Least Mean Square (LMS) algorithm Adaline Description (adaptive linear neuron) Applications to adaptive noise cancellation (ANC) References: [Hagan], [Haykin] EC4460.SuFy06/MPF 1

1) Wiener-Hopf Equations p 0 p 1 p w 0 w 1 w Assume 9 sensors on an array y + Σ + d desired response p q w q Output: ( ) y n = T w p T = 0,, q p p p T = 0,, q w w w LMS criterion: find weights ω i which minimize the Mean Squared (MS) error: σ d y e e = E = E σ e is minimum for ω so that: σ e ω k = 0, k E e = 0, k ω k Assume the filter is real E e = ω k EC4460.SuFy06/MPF

[ ] k σ e = Ee0pk = 0 k= 0,, q Orthogonality Principle: Let the error e = d y. The weights ω k minimize the MSE if ω is chosen so that σ e [ ] = 0 = 0,, E e p k q k (the error e is orthogonal to the input data p k ) Consequences: Ex: The estimate of the desired response to the filter y 0 (n) and the error e 0 (n) are Why? E[ y0e 0] = p 3 p p 1 EC4460.SuFy06/MPF 3

Optimum Weight Computation: Use orthogonality principle [ ] Vk σ e = 0 V k σ e = E e pi = 0 i = 0,, q ( ) 0 E d y p i = i = 0,, q EC4460.SuFy06/MPF 4

EC4460.SuFy06/MPF 5

Summary: u k + p k + filter y k + target output desired response d k noise n k y k W-H equations lead to designing a filter with: yk T = ω So that the MSE e E ( dk yk) p { } σ = is minimum Operations W-H optimum solution: d d d k k k = = = p p p k k l k+ l ω opt = R r 1 u * du Minimum MSE (MMSE) given as: σ e d T ( ) ω opt = R 0 rdu EC4460.SuFy06/MPF 6

Example Assume that you are given: (1) A noisy signal x(n)=s(n)+w(n) () access to w (n) noise correlated to w(n) (3) s(n) wideband (4) w(n) narrowband Goal: minimize the noise effects EC4460.SuFy06/MPF 7

EC4460.SuFy06/MPF 8

) Method of Steepest Descent (SD) W-H equations require to compute the inverse of a matrix. Solution may need to be recomputed if the signal input changes its behavior. Addressed via the method of SD by solving W-H equations iteratively (leading to an adaptive scheme). Filter progressively learns the correlation and cross-correlation, and the filter coefficients converge to the optimum values given by the W-H equations. i.e., finds the minimum values of Descent Directions σ e iteratively. take steps that lead downhill for the MSE function: ( ) = E{ e } k F w T = E dk ω p EC4460.SuFy06/MPF 9

F(ω) F min = MMSE ω 0 ω updated weight 1 ω ( ( )) n+ 1 = ωn + γ Vω F ω n γ > 0 EC4460.SuFy06/MPF 10

Stability of Steepest Descent Iterative solution: General solution: Note n ωn+ 1 = I γr p ωn + γr pd n ( I Rp ) ( ) ω = ω + γ ω ω SD algorithm CV opt 0 opt n ( I Rp ) λ( I γrp ) γλ ( Rp ) ωn ωopt iff γ 0 iff < 1 iff 1 < 1 ( Rp ) λ( Rp ) iff 1 γλ < 1 0< γ < SD algorithm CV when ( R ) 0< γ < λmax p In practice use the fact that (for a filter of length Q): λ max Q 1 ( ) ( ) R trace R R i, i QR 0 p p p p i= 0 ( ) 0< γ < QR p 0 Q filter length EC4460.SuFy06/MPF 11

Geometrical Significance of Eigenvectors and Eigenvalues 1 7 R p ω R p 1 9 1 u = ; =, = u, d = 4 i σ find ω and MMSE EC4460.SuFy06/MPF 1

Error surface shape and eigenvalue ratios a 1 = 0.195 a = 0.95 λ 1 = 1.1 λ = 0.9 χ = 1. a 1 = 0.975 a = 0.95 λ 1 = 1.5 λ = 0.5 χ = 3 EC4460.SuFy06/MPF 13

λ 1 = 1.81 λ = 0.18 χ = 10 λ 1 = 1.957 λ = 0.0198 χ = 100 EC4460.SuFy06/MPF 14

Effects of varying the step size on the iterative scheme behavior EC4460.SuFy06/MPF 15

Adaline Network (Textbook notations) Input R a = Goal: find w and b which minimize the error between target and network outputs. ( ) ( ) ( T F x = E e E t a E t x z) = = ; w p x= z = b 1 p R x 1 1 W S x R b S x 1 Linear Neuron n S S x 1 a = purelin (Wp + b) overall all feature vectors and target outputs a S x 1 ( ) ( ) ( T F x = E t a = E t x z) T T T T = E t x z z x+ x zz x EC4460.SuFy06/MPF 16

When does F(x) have a minimum? In practice, how do we get to the optimum solution? use one of the iterative optimization techniques seen before ( ) x = x αv F x k+ 1 k+ 1 x= x k How to multiply notations and computations? (LMS algorithm) EC4460.SuFy06/MPF 17

Gradient Approximation Consequence in Convergence Behavior LMS trajectory for α =0.1 1 0-1 - - -1 0 1 EC4460.SuFy06/MPF 18

Bound on the Learning Rate α Recall that steepest descent algorithm does: 1 T T F( x) = x Ax+ d + c x k+ 1 = xk + Here we have: T ( ) F x = x Rx x h+ c x + = k 1 T Simplification for an actually usable bound on α T R = E zz When the feature vector magnitude changes dramatically needs to normalize impact of update xk+ 1 = xk + αezk / z + eps EC4460.SuFy06/MPF 19

Application to Adaptive Noise Cancellation (with reference) Assume you receive a noisy signal x(n) + w(n) desired speech-only signal Goal: extract x(n) to make signal easier to understand cockpit noise Ex: pilot with mask case noisy voice: received inside the mask: x(n) + w(n) noise reference: collected outside the mask: w (n) Adaptive filter set up: w (n) filter d(n) + + e(n) d(n) EC4460.SuFy06/MPF 0

Application to Adaptive Noise Cancellation (without reference) Assume you receive a noisy signal x(n) + w(n) Goal: extract x(n) speech-only signal cockpit noise EC4460.SuFy06/MPF 1

EC4460.SuFy06/MPF