Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion
|
|
- Clinton Nash
- 5 years ago
- Views:
Transcription
1 Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE slides were adapted from ENEE630 slides developed by Profs. K.J. Ray Liu and Min Wu at the University of Maryland, College Park. Contact: Updated: February 14, ECE Lecture Part I 1 / 20
2 Complexity in Solving Linear Prediction (Refs: Hayes 5.2; Haykin 4th Ed. 3.3) Recall Augmented Normal Equation for linear prediction: [ ] [ PM 0 FLP R M+1 a M = BLP R 0 M+1 a B M = P M ] As R M+1 is usually non-singular, a M may be obtained by inverting R M+1, or Gaussian elimination for solving equation array: Computational complexity O(M 3 ). ECE Lecture Part I 2 / 20
3 Exploiting Structures in Matrix and LP Problem Complexity in solving a general linear equation array: Method-1: invert the matrix, e.g. compute determinant of R M+1 matrix and the adjacency matrices matrix inversion has O(M 3 ) complexity Method-2: use Gaussian elimination approximately M 3 /3 multiplication and division By exploring the Toeplitz structure of the matrix, Levinson-Durbin recursion can reduce complexity to O(M 2 ) M steps of order recursion, each step has a linear complexity w.r.t. intermediate order Memory use: Gaussian elimination O(M 2 ) for the matrix, vs. Levinson-Durbin O(M) for the autocorrelation vector and model parameter vector. ECE Lecture Part I 3 / 20
4 Levinson-Durbin Recursion The Levinson-Durbin recursion is an order-recursion to efficiently solve linear systems with Toeplitz matrices, e.g., Augmented N.E. M steps of order recursion, each step has a linear complexity w.r.t. intermediate order The recursion can be stated in two ways for the linear prediction problem: 1 Forward prediction point of view 2 Backward prediction point of view ECE Lecture Part I 4 / 20
5 Two Points of View of LD Recursion Denote a m C (m+1) 1 as the tap weight vector of a forward-prediction-error filter of order m = 0,..., M. a m 1,0 = 1, a m 1,m 0, a m,m = Γ m (a constant reflection coefficient ) Forward prediction point of view a m,k = a m 1,k + Γ m am 1,m k, k = 0, 1,..., m [ ] [ ] am 1 0 In vector form: a m = + Γ 0 m a B m 1 ( ) Backward prediction point of view am,m k = a m 1,m k + Γ m a m 1,k, k = 0, 1,..., m [ ] [ ] 0 In vector form: a B m = + Γ a am 1 B m (can be obtained by m 1 0 reordering and conjugating ( )) ECE Lecture Part I 5 / 20
6 Recall: Forward and Backward Prediction Errors f m [n] = u[n] û[n] = a H m u[n] }{{} (m+1) 1 b m [n] = u[n m] û[n m] = a B,T m u[n] ECE Lecture Part I 6 / 20
7 (3) Verify the Update Equations of the LD Recursion Left multiply both sides of ( ) by R m+1 : [ ] Pm LHS: R m+1 a m = (by augmented N.E.) 0 m [ ] [ ] [ ] am 1 Rm r RHS (1): R m+1 = B m am 1 0 r BT m r(0) 0 [ ] P Rm a m 1 = m 1 r BT m a = 0 m 1 where m 1 r BT m a m 1 m 1 RHS (2): R m+1 [ = 0 a B m 1 [ r H a B m 1 R m a B m 1 m 1 ] [ ] [ r(0) r H = r R m ] m 1 = 0 m 1 P m 1 0 a B m 1 ] ECE Lecture Part I 7 / 20
8 Computing Γ m Put together LHS and RHS: for the order update recursion ( ) to hold, we should have [ ] P m 1 Pm m 1 = 0 0 m 1 + Γ m 0 m 1 m m 1 P m 1 { P m = P m 1 + Γ m m 1 0 = m 1 + Γ m P m 1 a m,m = Γ m = m 1 P m 1 P m = P m 1 ( 1 Γm 2) Caution: Do not confuse the power term P m and the ratio term Γ m. ECE Lecture Part I 8 / 20
9 (4) Reflection Coefficients Γ m To ensure the prediction MSE P m 0 and P m non-increasing as we increase the order of the predictor (i.e., 0 P m P m 1 ), we require Γ m 2 1, m > 0. Let P 0 = r(0) as the initial estimation error has power equal to the signal power (i.e., no regression is applied), we have P M = P 0 M m=1 (1 Γ m 2 ) Question: Under what situation Γ m = 0? i.e., increasing order won t reduce error. Consider a process with Markovian-like property in 2nd order statistic sense (e.g. AR process) s.t. info of further past is contained in k recent samples ECE Lecture Part I 9 / 20
10 (5) About m One can show that the cross-correlation of BLP error and FLP error E [ b m 1 [n 1]f m 1 [n]] is equal to m 1. (Derive from the definition m 1 r BT m a m 1, and use definitions of b m 1[n 1], fm 1[n] and orthogonality principle.) Thus the reflection coefficient can be written as Γ m = m 1 = E [ bm 1 [n 1]fm 1 [n]] P m 1 E [ f m 1 [n] 2 ] which is also the negative partial correlation coefficient. Note: for the 0th order predictor, use the mean value, i.e., zero, as the estimate, s.t. f 0 [n] = u[n] = b 0 [n], 0 = E [b 0 [n 1]f 0 [n]] = E [u[n 1]u [n]] = r( 1) = r (1) ECE Lecture Part I 10 / 20
11 Preview: Relations of w.s.s and LP Parameters For any w.s.s. process {u[n]}: ECE Lecture Part I 11 / 20
12 (6) Computing a M and P M by Forward Recursion Case-1 : If we know the autocorrelation function r( ): # of iterations = M m=1 m = M(M+1) 2, comp. complexity is O(M 2 ) r(k) may be estimated from time average of one realization of {u[n]}: ˆr(k) = 1 N k N n=k+1 u[n]u [n k], k = 0, 1,..., M (recall correlation ergodicity) ECE Lecture Part I 12 / 20
13 (6) Computing a M and P M by Forward Recursion Case-2 : If we know Γ 1, Γ 2,..., Γ M and P 0 = r(0), we can carry out the recursion for m = 1, 2,..., M: { a m,k = a m 1,k + Γ m a m 1,m k, k = 1,..., m P m = P m 1 ( 1 Γm 2) Note: a m,m = a m 1,m + Γ m a m 1,0 = 0 + Γ m 1 = Γ m ECE Lecture Part I 13 / 20
14 (7) Inverse Form of Levinson-Durbin Recursion Given the tap-weights a M, find the reflection coefficients Γ 1, Γ 2,..., Γ M : { (FP) a m,k = a m 1,k + Γ m am 1,m k Recall:, k = 0,..., m (BP) am,m k = a m 1,m k + Γ m a m 1,k, a m,m = Γ m Multiply (BP) by Γ m and subtract from (FP): a m 1,k = a m,k Γ ma m,m k 1 Γ m 2 Γ m = a m,m, Γ m 1 = a m 1,m 1,..., iterate with m = M 1, M 2,... = a m,k a m,ma m,m k 1 a m,m 2, k = 0,..., m 1 i.e., From a M a m Γ m to lower order see 5 Lattice structure: ECE Lecture Part I 14 / 20
15 (8) Autocorrelation Function & Reflection Coefficients Recall: The 2nd-order statistics of a stationary time series can be represented in terms of autocorrelation function r(k), or equivalently the power spectral density by taking DTFT. Another way is to use {r(0), Γ 1, Γ 2,..., Γ M }. To find the relation between them, recall: m 1 r BT m a m 1 = M 1 k=0 a m 1,kr( m + k) and Γ m = m 1 P m 1 Γ m P m 1 = m 1 k=0 a m 1,kr(k m), where a m 1,0 = 1. ECE Lecture Part I 15 / 20
16 (8) Autocorrelation Function & Reflection Coefficients 1 r(m) = r ( m) = Γ mp m 1 m 1 k=1 a m 1,kr(m k) Given r(0), Γ 1, Γ 2,..., Γ M, can get a m using Levinson-Durbin recursion s.t. r(1),..., r(m) can be generated recursively. 2 Recall if r(0),..., r(m) are given, we can get a m. So Γ 1,..., Γ M can be obtained recursively: Γ m = a m,m 3 These facts imply that the reflection coefficients {Γ k } can uniquely represent the 2nd-order statistics of a w.s.s. process. ECE Lecture Part I 16 / 20
17 Summary Statistical representation of w.s.s. process ECE Lecture Part I 17 / 20
18 Detailed Derivations/Examples ECE Lecture Part I 18 / 20
19 Example of Forward Recursion Case-2 ECE Lecture Part I 19 / 20
20 Proof for m 1 Property ECE Lecture Part I 20 / 20
Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes
Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted
More informationLecture 7: Linear Prediction
1 Lecture 7: Linear Prediction Overview Dealing with three notions: PREDICTION, PREDICTOR, PREDICTION ERROR; FORWARD versus BACKWARD: Predicting the future versus (improper terminology) predicting the
More informationPart III Spectrum Estimation
ECE79-4 Part III Part III Spectrum Estimation 3. Parametric Methods for Spectral Estimation Electrical & Computer Engineering North Carolina State University Acnowledgment: ECE79-4 slides were adapted
More informationParametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)
Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides
More information1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c =
ENEE630 ADSP Part II w/ solution. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix) R a = 4 4 4,R b = 0 0,R c = j 0 j 0 j 0 j 0 j,r d = 0 0 0
More informationParametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes
1 Discrete-time Stochastic Processes Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering University of Maryland, College Park
More informationOrder Recursion Introduction Order versus Time Updates Matrix Inversion by Partitioning Lemma Levinson Algorithm Interpretations Examples
Order Recursion Introduction Order versus Tie Updates Matrix Inversion by Partitioning Lea Levinson Algorith Interpretations Exaples Introduction Rc d There are any ways to solve the noral equations Solutions
More informationAdvanced Digital Signal Processing -Introduction
Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary
More informationStatistical and Adaptive Signal Processing
r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory
More informationBasics on 2-D 2 D Random Signal
Basics on -D D Random Signal Spring 06 Instructor: K. J. Ray Liu ECE Department, Univ. of Maryland, College Park Overview Last Time: Fourier Analysis for -D signals Image enhancement via spatial filtering
More informationLesson 1. Optimal signalbehandling LTH. September Statistical Digital Signal Processing and Modeling, Hayes, M:
Lesson 1 Optimal Signal Processing Optimal signalbehandling LTH September 2013 Statistical Digital Signal Processing and Modeling, Hayes, M: John Wiley & Sons, 1996. ISBN 0471594318 Nedelko Grbic Mtrl
More informationTime Series: Theory and Methods
Peter J. Brockwell Richard A. Davis Time Series: Theory and Methods Second Edition With 124 Illustrations Springer Contents Preface to the Second Edition Preface to the First Edition vn ix CHAPTER 1 Stationary
More informationSOLVING LINEAR SYSTEMS
SOLVING LINEAR SYSTEMS We want to solve the linear system a, x + + a,n x n = b a n, x + + a n,n x n = b n This will be done by the method used in beginning algebra, by successively eliminating unknowns
More informationL7: Linear prediction of speech
L7: Linear prediction of speech Introduction Linear prediction Finding the linear prediction coefficients Alternative representations This lecture is based on [Dutoit and Marques, 2009, ch1; Taylor, 2009,
More informationChapter 6 Page 1 of 10. Lecture Guide. Math College Algebra Chapter 6. to accompany. College Algebra by Julie Miller
Chapter 6 Page 1 of 10 Lecture Guide Math 105 - College Algebra Chapter 6 to accompany College Algebra by Julie Miller Corresponding Lecture Videos can be found at Prepared by Stephen Toner & Nichole DuBal
More informationADAPTIVE FILTER THEORY
ADAPTIVE FILTER THEORY Fourth Edition Simon Haykin Communications Research Laboratory McMaster University Hamilton, Ontario, Canada Front ice Hall PRENTICE HALL Upper Saddle River, New Jersey 07458 Preface
More informationChapter 9. Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析
Chapter 9 Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析 1 LPC Methods LPC methods are the most widely used in speech coding, speech synthesis, speech recognition, speaker recognition and verification
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More informationMultimedia Communications. Differential Coding
Multimedia Communications Differential Coding Differential Coding In many sources, the source output does not change a great deal from one sample to the next. This means that both the dynamic range and
More informationMulti-rate Signal Processing 7. M-channel Maximally Decmiated Filter Banks
Multi-rate Signal Processing 7. M-channel Maximally Decmiated Filter Banks Electrical & Computer Engineering University of Maryland, College Park Acknowledgment: ENEE630 slides were based on class notes
More informationLinear Prediction Theory
Linear Prediction Theory Joseph A. O Sullivan ESE 524 Spring 29 March 3, 29 Overview The problem of estimating a value of a random process given other values of the random process is pervasive. Many problems
More informationMAC Learning Objectives. Learning Objectives (Cont.) Module 10 System of Equations and Inequalities II
MAC 1140 Module 10 System of Equations and Inequalities II Learning Objectives Upon completing this module, you should be able to 1. represent systems of linear equations with matrices. 2. transform a
More informationRevised Simplex Method
DM545 Linear and Integer Programming Lecture 7 Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 2 Motivation Complexity of single pivot operation
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any
More informationCCNY. BME I5100: Biomedical Signal Processing. Stochastic Processes. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Stochastic Processes Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
More informationSIMON FRASER UNIVERSITY School of Engineering Science
SIMON FRASER UNIVERSITY School of Engineering Science Course Outline ENSC 810-3 Digital Signal Processing Calendar Description This course covers advanced digital signal processing techniques. The main
More informationLecture 4 - Spectral Estimation
Lecture 4 - Spectral Estimation The Discrete Fourier Transform The Discrete Fourier Transform (DFT) is the equivalent of the continuous Fourier Transform for signals known only at N instants separated
More informationADAPTIVE FILTER THEORY
ADAPTIVE FILTER THEORY Fifth Edition Simon Haykin Communications Research Laboratory McMaster University Hamilton, Ontario, Canada International Edition contributions by Telagarapu Prabhakar Department
More informationLevinson Durbin Recursions: I
Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions
More informationSection Gaussian Elimination
Section. - Gaussian Elimination A matrix is said to be in row echelon form (REF) if it has the following properties:. The first nonzero entry in any row is a. We call this a leading one or pivot one..
More informationLevinson Durbin Recursions: I
Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: proximal gradient descent Consider the problem min g(x) + h(x) with g, h convex, g differentiable, and h simple
More informationChapter 2 Wiener Filtering
Chapter 2 Wiener Filtering Abstract Before moving to the actual adaptive filtering problem, we need to solve the optimum linear filtering problem (particularly, in the mean-square-error sense). We start
More informationProf. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides
Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering Stochastic Processes and Linear Algebra Recap Slides Stochastic processes and variables XX tt 0 = XX xx nn (tt) xx 2 (tt) XX tt XX
More informationComputational Fluid Dynamics Prof. Sreenivas Jayanti Department of Computer Science and Engineering Indian Institute of Technology, Madras
Computational Fluid Dynamics Prof. Sreenivas Jayanti Department of Computer Science and Engineering Indian Institute of Technology, Madras Lecture 46 Tri-diagonal Matrix Algorithm: Derivation In the last
More informationAdaptive Beamforming Algorithms
S. R. Zinka srinivasa_zinka@daiict.ac.in October 29, 2014 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method Outline
More informationLinear Prediction 1 / 41
Linear Prediction 1 / 41 A map of speech signal processing Natural signals Models Artificial signals Inference Speech synthesis Hidden Markov Inference Homomorphic processing Dereverberation, Deconvolution
More information2.1 Gaussian Elimination
2. Gaussian Elimination A common problem encountered in numerical models is the one in which there are n equations and n unknowns. The following is a description of the Gaussian elimination method for
More informationFeature extraction 2
Centre for Vision Speech & Signal Processing University of Surrey, Guildford GU2 7XH. Feature extraction 2 Dr Philip Jackson Linear prediction Perceptual linear prediction Comparison of feature methods
More information2-D FIR WIENER FILTER REALIZATIONS USING ORTHOGONAL LATTICE STRUCTURES
-D FIR WIENER FILTER REALIZATIONS USING ORTHOGONAL LATTICE STRUCTURES by Ahmet H. Kayran and Ender Ekşioğlu Department of Electrical Engineering Istanbul Technical University aslak, Istanbul, Turkey 866
More informationReview of Discrete-Time System
Review of Discrete-Time System Electrical & Computer Engineering University of Maryland, College Park Acknowledgment: ENEE630 slides were based on class notes developed by Profs. K.J. Ray Liu and Min Wu.
More informationFIR Filters for Stationary State Space Signal Models
Proceedings of the 17th World Congress The International Federation of Automatic Control FIR Filters for Stationary State Space Signal Models Jung Hun Park Wook Hyun Kwon School of Electrical Engineering
More informationExample: Current in an Electrical Circuit. Solving Linear Systems:Direct Methods. Linear Systems of Equations. Solving Linear Systems: Direct Methods
Example: Current in an Electrical Circuit Solving Linear Systems:Direct Methods A number of engineering problems or models can be formulated in terms of systems of equations Examples: Electrical Circuit
More information6.3 Forecasting ARMA processes
6.3. FORECASTING ARMA PROCESSES 123 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss
More informationLecture 18: The Rank of a Matrix and Consistency of Linear Systems
Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Winfried Just Department of Mathematics, Ohio University February 28, 218 Review: The linear span Definition Let { v 1, v 2,..., v n }
More informationGaussian Elimination and Back Substitution
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 4 Notes These notes correspond to Sections 31 and 32 in the text Gaussian Elimination and Back Substitution The basic idea behind methods for solving
More informationLU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark
DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline
More informationOpen Economy Macroeconomics: Theory, methods and applications
Open Economy Macroeconomics: Theory, methods and applications Lecture 4: The state space representation and the Kalman Filter Hernán D. Seoane UC3M January, 2016 Today s lecture State space representation
More informationMathematical Optimisation, Chpt 2: Linear Equations and inequalities
Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson
More information(Refer Slide Time: )
Digital Signal Processing Prof. S. C. Dutta Roy Department of Electrical Engineering Indian Institute of Technology, Delhi FIR Lattice Synthesis Lecture - 32 This is the 32nd lecture and our topic for
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,
More informationLinear Equations in Linear Algebra
1 Linear Equations in Linear Algebra 1.1 SYSTEMS OF LINEAR EQUATIONS LINEAR EQUATION x 1,, x n A linear equation in the variables equation that can be written in the form a 1 x 1 + a 2 x 2 + + a n x n
More informationMATHEMATICS FOR COMPUTER VISION WEEK 2 LINEAR SYSTEMS. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year
1 MATHEMATICS FOR COMPUTER VISION WEEK 2 LINEAR SYSTEMS Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 2 Linear Systems and solutions Systems of linear
More informationMulti-rate Signal Processing 3. The Polyphase Representation
Multi-rate Signal Processing 3. The Polyphase Representation Electrical & Computer Engineering University of Maryland, College Park Acknowledgment: ENEE630 slides were based on class notes developed by
More informationMTH 464: Computational Linear Algebra
MTH 464: Computational Linear Algebra Lecture Outlines Exam 2 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University February 6, 2018 Linear Algebra (MTH
More informationAck: 1. LD Garcia, MTH 199, Sam Houston State University 2. Linear Algebra and Its Applications - Gilbert Strang
Gaussian Elimination CS6015 : Linear Algebra Ack: 1. LD Garcia, MTH 199, Sam Houston State University 2. Linear Algebra and Its Applications - Gilbert Strang The Gaussian Elimination Method The Gaussian
More informationStatistical Signal Processing Detection, Estimation, and Time Series Analysis
Statistical Signal Processing Detection, Estimation, and Time Series Analysis Louis L. Scharf University of Colorado at Boulder with Cedric Demeure collaborating on Chapters 10 and 11 A TT ADDISON-WESLEY
More informationLeast Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on
Least Square Es?ma?on, Filtering, and Predic?on: Sta?s?cal Signal Processing II: Linear Es?ma?on Eric Wan, Ph.D. Fall 2015 1 Mo?va?ons If the second-order sta?s?cs are known, the op?mum es?mator is given
More informationCS412: Lecture #17. Mridul Aanjaneya. March 19, 2015
CS: Lecture #7 Mridul Aanjaneya March 9, 5 Solving linear systems of equations Consider a lower triangular matrix L: l l l L = l 3 l 3 l 33 l n l nn A procedure similar to that for upper triangular systems
More informationMatrices: 2.1 Operations with Matrices
Goals In this chapter and section we study matrix operations: Define matrix addition Define multiplication of matrix by a scalar, to be called scalar multiplication. Define multiplication of two matrices,
More informationMultiplying matrices by diagonal matrices is faster than usual matrix multiplication.
7-6 Multiplying matrices by diagonal matrices is faster than usual matrix multiplication. The following equations generalize to matrices of any size. Multiplying a matrix from the left by a diagonal matrix
More informationChapter 4 Systems of Linear Equations; Matrices
Chapter 4 Systems of Linear Equations; Matrices Section 5 Inverse of a Square Matrix Learning Objectives for Section 4.5 Inverse of a Square Matrix The student will be able to identify identity matrices
More informationMatrices and systems of linear equations
Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.
More informationA Review of Matrix Analysis
Matrix Notation Part Matrix Operations Matrices are simply rectangular arrays of quantities Each quantity in the array is called an element of the matrix and an element can be either a numerical value
More informationMaths for Signals and Systems Linear Algebra for Engineering Applications
Maths for Signals and Systems Linear Algebra for Engineering Applications Lectures 1-2, Tuesday 11 th October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON
More informationHermite normal form: Computation and applications
Integer Points in Polyhedra Gennady Shmonin Hermite normal form: Computation and applications February 24, 2009 1 Uniqueness of Hermite normal form In the last lecture, we showed that if B is a rational
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725 Consider Last time: proximal Newton method min x g(x) + h(x) where g, h convex, g twice differentiable, and h simple. Proximal
More informationLU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b
AM 205: lecture 7 Last time: LU factorization Today s lecture: Cholesky factorization, timing, QR factorization Reminder: assignment 1 due at 5 PM on Friday September 22 LU Factorization LU factorization
More informationChapter 7. Tridiagonal linear systems. Solving tridiagonal systems of equations. and subdiagonal. E.g. a 21 a 22 a A =
Chapter 7 Tridiagonal linear systems The solution of linear systems of equations is one of the most important areas of computational mathematics. A complete treatment is impossible here but we will discuss
More informationAutoregressive Approximation in Nonstandard Situations: Empirical Evidence
Autoregressive Approximation in Nonstandard Situations: Empirical Evidence S. D. Grose and D. S. Poskitt Department of Econometrics and Business Statistics, Monash University Abstract This paper investigates
More informationLecture 12 (Tue, Mar 5) Gaussian elimination and LU factorization (II)
Math 59 Lecture 2 (Tue Mar 5) Gaussian elimination and LU factorization (II) 2 Gaussian elimination - LU factorization For a general n n matrix A the Gaussian elimination produces an LU factorization if
More informationMethods for Solving Linear Systems Part 2
Methods for Solving Linear Systems Part 2 We have studied the properties of matrices and found out that there are more ways that we can solve Linear Systems. In Section 7.3, we learned that we can use
More informationMatrix Factorizations
1 Stat 540, Matrix Factorizations Matrix Factorizations LU Factorization Definition... Given a square k k matrix S, the LU factorization (or decomposition) represents S as the product of two triangular
More informationScientific Computing
Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting
More informationLecture 1: Systems of linear equations and their solutions
Lecture 1: Systems of linear equations and their solutions Course overview Topics to be covered this semester: Systems of linear equations and Gaussian elimination: Solving linear equations and applications
More informationLinear Equation: a 1 x 1 + a 2 x a n x n = b. x 1, x 2,..., x n : variables or unknowns
Linear Equation: a x + a 2 x 2 +... + a n x n = b. x, x 2,..., x n : variables or unknowns a, a 2,..., a n : coefficients b: constant term Examples: x + 4 2 y + (2 5)z = is linear. x 2 + y + yz = 2 is
More informationMatrix Calculations: Determinants and Basis Transformation
Matrix Calculations: Determinants and asis Transformation A. Kissinger Institute for Computing and Information ciences Version: autumn 27 A. Kissinger Version: autumn 27 Matrix Calculations / 32 Outline
More informationADSP ADSP ADSP ADSP. Advanced Digital Signal Processing (18-792) Spring Fall Semester, Department of Electrical and Computer Engineering
Advanced Digital Signal rocessing (18-792) Spring Fall Semester, 201 2012 Department of Electrical and Computer Engineering ROBLEM SET 8 Issued: 10/26/18 Due: 11/2/18 Note: This problem set is due Friday,
More informationEEM 409. Random Signals. Problem Set-2: (Power Spectral Density, LTI Systems with Random Inputs) Problem 1: Problem 2:
EEM 409 Random Signals Problem Set-2: (Power Spectral Density, LTI Systems with Random Inputs) Problem 1: Consider a random process of the form = + Problem 2: X(t) = b cos(2π t + ), where b is a constant,
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More information6. Methods for Rational Spectra It is assumed that signals have rational spectra m k= m
6. Methods for Rational Spectra It is assumed that signals have rational spectra m k= m φ(ω) = γ ke jωk n k= n ρ, (23) jωk ke where γ k = γ k and ρ k = ρ k. Any continuous PSD can be approximated arbitrary
More informationSimultaneous Linear Equations
Simultaneous Linear Equations PHYSICAL PROBLEMS Truss Problem Pressure vessel problem a a b c b Polynomial Regression We are to fit the data to the polynomial regression model Simultaneous Linear Equations
More informationElec4621: Advanced Digital Signal Processing Chapter 8: Wiener and Adaptive Filtering
Elec462: Advanced Digital Signal Processing Chapter 8: Wiener and Adaptive Filtering Dr D S Taubman May 2, 2 Wiener Filtering A Wiener lter is one which provides the Minimum Mean Squared Error (MMSE) prediction,
More information11. Further Issues in Using OLS with TS Data
11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,
More informationMATHEMATICAL METHODS INTERPOLATION
MATHEMATICAL METHODS INTERPOLATION I YEAR BTech By Mr Y Prabhaker Reddy Asst Professor of Mathematics Guru Nanak Engineering College Ibrahimpatnam, Hyderabad SYLLABUS OF MATHEMATICAL METHODS (as per JNTU
More informationCONTENTS NOTATIONAL CONVENTIONS GLOSSARY OF KEY SYMBOLS 1 INTRODUCTION 1
DIGITAL SPECTRAL ANALYSIS WITH APPLICATIONS S.LAWRENCE MARPLE, JR. SUMMARY This new book provides a broad perspective of spectral estimation techniques and their implementation. It concerned with spectral
More informationDM545 Linear and Integer Programming. Lecture 7 Revised Simplex Method. Marco Chiarandini
DM545 Linear and Integer Programming Lecture 7 Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 2 Motivation Complexity of single pivot operation
More informationCITS2401 Computer Analysis & Visualisation
FACULTY OF ENGINEERING, COMPUTING AND MATHEMATICS CITS2401 Computer Analysis & Visualisation SCHOOL OF COMPUTER SCIENCE AND SOFTWARE ENGINEERING Topic 7 Matrix Algebra Material from MATLAB for Engineers,
More informationNumerical Linear Algebra
Numerical Linear Algebra Decompositions, numerical aspects Gerard Sleijpen and Martin van Gijzen September 27, 2017 1 Delft University of Technology Program Lecture 2 LU-decomposition Basic algorithm Cost
More informationProgram Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects
Numerical Linear Algebra Decompositions, numerical aspects Program Lecture 2 LU-decomposition Basic algorithm Cost Stability Pivoting Cholesky decomposition Sparse matrices and reorderings Gerard Sleijpen
More informationLecture 7: Introduction to linear systems
Lecture 7: Introduction to linear systems Two pictures of linear systems Consider the following system of linear algebraic equations { x 2y =, 2x+y = 7. (.) Note that it is a linear system with two unknowns
More information5.7 Cramer's Rule 1. Using Determinants to Solve Systems Assumes the system of two equations in two unknowns
5.7 Cramer's Rule 1. Using Determinants to Solve Systems Assumes the system of two equations in two unknowns (1) possesses the solution and provided that.. The numerators and denominators are recognized
More informationScientific Computing: Dense Linear Systems
Scientific Computing: Dense Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 February 9th, 2012 A. Donev (Courant Institute)
More informationMath 60. Rumbos Spring Solutions to Assignment #17
Math 60. Rumbos Spring 2009 1 Solutions to Assignment #17 a b 1. Prove that if ad bc 0 then the matrix A = is invertible and c d compute A 1. a b Solution: Let A = and assume that ad bc 0. c d First consider
More informationAutoregressive (AR) Modelling
Autoregressive (AR) Modelling A. Uses of AR Modelling () Alications (a) Seech recognition and coding (storage) (b) System identification (c) Modelling and recognition of sonar, radar, geohysical signals
More informationMAC1105-College Algebra. Chapter 5-Systems of Equations & Matrices
MAC05-College Algebra Chapter 5-Systems of Equations & Matrices 5. Systems of Equations in Two Variables Solving Systems of Two Linear Equations/ Two-Variable Linear Equations A system of equations is
More informationDr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)
Dr. Maddah ENMG 617 EM Statistics 11/28/12 Multiple Regression (3) (Chapter 15, Hines) Problems in multiple regression: Multicollinearity This arises when the independent variables x 1, x 2,, x k, are
More information1.Chapter Objectives
LU Factorization INDEX 1.Chapter objectives 2.Overview of LU factorization 2.1GAUSS ELIMINATION AS LU FACTORIZATION 2.2LU Factorization with Pivoting 2.3 MATLAB Function: lu 3. CHOLESKY FACTORIZATION 3.1
More informationMath 471 (Numerical methods) Chapter 3 (second half). System of equations
Math 47 (Numerical methods) Chapter 3 (second half). System of equations Overlap 3.5 3.8 of Bradie 3.5 LU factorization w/o pivoting. Motivation: ( ) A I Gaussian Elimination (U L ) where U is upper triangular
More informationFall 2011, EE123 Digital Signal Processing
Lecture 6 Miki Lustig, UCB September 11, 2012 Miki Lustig, UCB DFT and Sampling the DTFT X (e jω ) = e j4ω sin2 (5ω/2) sin 2 (ω/2) 5 x[n] 25 X(e jω ) 4 20 3 15 2 1 0 10 5 1 0 5 10 15 n 0 0 2 4 6 ω 5 reconstructed
More information