Linear Least Square Problems Dr.-Ing. Sudchai Boonto

Similar documents
Lecture 5 Least-squares

EECS 275 Matrix Computation

Lecture 9: Input Disturbance A Design Example Dr.-Ing. Sudchai Boonto

Lecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018

Nonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto

Least Squares. Stephen Boyd. EE103 Stanford University. October 28, 2017

Least squares problems Linear Algebra with Computer Science Application

Solutions to Review Problems for Chapter 6 ( ), 7.1

Orthogonal Projection. Hung-yi Lee

Pseudoinverse & Moore-Penrose Conditions

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

Pseudoinverse and Adjoint Operators

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

Stochastic Process II Dr.-Ing. Sudchai Boonto

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Chapter 6: Orthogonality

Filtering and Identification

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Review problems for MA 54, Fall 2004.

Least-squares data fitting

Chapter 6 - Orthogonality

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

Lecture 6: Geometry of OLS Estimation of Linear Regession

Lecture 7: Laplace Transform and Its Applications Dr.-Ing. Sudchai Boonto

Consider a subspace V = im(a) of R n, where m. Then,

Assignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.

Final Exam, Linear Algebra, Fall, 2003, W. Stephen Wilson

5. Orthogonal matrices

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015

Inverses. Stephen Boyd. EE103 Stanford University. October 28, 2017

Math 3191 Applied Linear Algebra

Lecture 19 Observability and state estimation

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

Lecture: Quadratic optimization

Lecture 9: Time-Domain Analysis of Discrete-Time Systems Dr.-Ing. Sudchai Boonto

Orthogonality and Least Squares

Conjugate Gradient algorithm. Storage: fixed, independent of number of steps.

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

1 Cricket chirps: an example

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

Exercise Sheet 1.

For each problem, place the letter choice of your answer in the spaces provided on this page.

DS-GA 1002 Lecture notes 10 November 23, Linear models

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

Typical Problem: Compute.

Lecture 8: Discrete-Time Signals and Systems Dr.-Ing. Sudchai Boonto

The SVD-Fundamental Theorem of Linear Algebra

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Quadratic forms. Defn

Singular Value Decomposition

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Math 407: Linear Optimization

Math 308 Practice Final Exam Page and vector y =

ECE 275A Homework #3 Solutions

3. For a given dataset and linear model, what do you think is true about least squares estimates? Is Ŷ always unique? Yes. Is ˆβ always unique? No.

Least squares: the big idea

Lecture 6. Regularized least-squares and minimum-norm methods 6 1

Lecture 10: Linear Matrix Inequalities Dr.-Ing. Sudchai Boonto

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture 1: Review of linear algebra

Check that your exam contains 30 multiple-choice questions, numbered sequentially.

MA 265 FINAL EXAM Fall 2012

Linear Algebra Review

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

STAT 350: Geometry of Least Squares

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

The geometry of least squares

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

8. Least squares. ˆ Review of linear equations. ˆ Least squares. ˆ Example: curve-fitting. ˆ Vector norms. ˆ Geometrical intuition

Econ 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A )

Conjugate Gradient Method

INC 693, 481 Dynamics System and Modelling: Introduction to Modelling Dr.-Ing. Sudchai Boonto Assistant Professor

EL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Homework 4. Convex Optimization /36-725

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Math 3191 Applied Linear Algebra

Linear algebra I Homework #1 due Thursday, Oct. 5

Designing Information Devices and Systems I Discussion 13B

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Fitting Linear Statistical Models to Data by Least Squares I: Introduction

The Full-rank Linear Least Squares Problem

Observability and state estimation

Lecture 7 : Generalized Plant and LFT form Dr.-Ing. Sudchai Boonto Assistant Professor

Fitting Linear Statistical Models to Data by Least Squares II: Weighted

Assignment #9: Orthogonal Projections, Gram-Schmidt, and Least Squares. Name:

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam January 23, 2015

Math 261 Lecture Notes: Sections 6.1, 6.2, 6.3 and 6.4 Orthogonal Sets and Projections

14 Singular Value Decomposition

Linear algebra review

DS-GA 1002 Lecture notes 12 Fall Linear regression

Topics. Review of lecture 2/11 Error, Residual and Condition Number. Review of lecture 2/16 Backward Error Analysis The General Case 1 / 22

Linear Systems. Carlo Tomasi

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Sample ECE275A Midterm Exam Questions

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

P = A(A T A) 1 A T. A Om (m n)

Transcription:

Dr-Ing Sudchai Boonto Department of Control System and Instrumentation Engineering King Mongkuts Unniversity of Technology Thonburi Thailand

Linear Least-Squares Problems Given y, measurement signal, find the least-squares solution x, min x ϵ T ϵ st y = Ax or more standard ofrm min x Ax y 2, where A R m n, y meas R m and x R n is an unknown parameter vector 2/18

Linear Least-Squares Problems cont The cost function of the least-squares minimization problem can be expanded as V (x) = Ax y 2 = (Ax y) T (Ax y) = x T A T Ax x T A T y y T Ax + y T y The gradient x V (x) = V (x) x 1 V (x) x n = 2AT Ax 2A T y 3/18

Linear Least-Squares Problems cont The solution ˆx to the least-square problem is found by setting the gradient equal to 0 Thus A T Aˆx = A T y, normal equation or A T (Aˆx y) = 0 The minimal value of the cost function Ax y 2 can be find from V (x) = ˆx T A T Aˆx ˆx T A T y y T Aˆx + y T y = ˆx T A T (Aˆx y) y T (Aˆx y) with the normal equation, the first of above equation is zero, then min V (x) = y T (Aˆx y) 4/18

Geometric interpretation y ϵ ŷ a 2 A a 1 we are looking for the projection ŷ of the vector y onto the space spanned by the measurement vector a i ŷ is the vector closest to y if the error ϵ is orthogonal to this space 5/18

Geometric interpretation cont ϵ is orthogonal to this space if a T i ϵ = 0, i = 1,, n in matrix from A T ϵ = 0 or A T (y ŷ) = A T (y Aˆx) = 0 6/18

Completion of Squares the completion of squares formula for quadratic polynomials is ( ax 2 + 2bxy + cy 2 = a x + b ) 2 ( ) a y + c b2 y 2 a when a > 0, this tells us the minimum with respect to x for fixed y ( ) min x R ax2 + 2bxy + cy 2 = c b2 y 2, a which is achieved when x = b a y 7/18

Completion of Squares of matrices if A R n n and D R m m are symmetric matrices and B R n m, then [ ] T [ ] [ ] x A B x y B T = (x + A 1 By) T A(x + A 1 By) D y compare with + y T (D B T A 1 B)y ( ax 2 + 2bxy + cy 2 = a x + b ) 2 ( ) a y + c b2 y 2 a 8/18

Completion of Squares of matrices gives a general formula for quadratic optimization; if A > 0, then min x x y T A B T and the minimizing x is B x = y T (D B T A 1 B)y D y ˆx = A 1 By 9/18

Application to Least-square V (x) = Ax y 2 = x T A T Ax x T A T y y T Ax + y T y [ ] T [ x A = T A A T ] [ ] y x 1 y T A y T y 1 }{{} M By the help of Schur complement [ ] T [ ] [ x I 0 A V (x) = T ] [ ] [ ] A 0 I ˆx x 1 ˆx T I 0 y T y y T, Aˆx 0 I 1 for ˆx satisfying and ˆx = (A T A) 1 A T y V (x) = (x ˆx) T A T A(x ˆx) + (y T y y T Aˆx) 10/18

regression or curve fitting model using a linear combination of functions f(t) = x 1 f 1 (t) + x 2 f 2 (t) + + x n f n (t) collect m data samples y i = f(t i ), i = 1,, m write in matrix form y 1 f 1 (t 1 ) f n (t 1 ) x 1 y 2 = f 1 (t 2 ) f n (t 2 ) x 2 y m f 1 (t m ) f n (t m ) x n find least-squares estimate for x by x est = (A T A) 1 A T y called curve fitting or linear regression; functions f i are called regressors 11/18

Solutions if the matrix A has full column rank if A T A is square and invertible and the solution ˆx is ˆx = (A T A) 1 A T y (A T A) 1 A T is called pseudo-inverse of A because ((A T A) 1 A T )A = I 12/18

Example Given 1 1 1 A = 2 1, y = 0, 1 1 0 consider the set of three equations in two unknowns, F x = y The least squares solution is given by ˆx = (A T A) 1 A T y = = [ 3 2 2 2 3 [ ] 1 2 1 ] [ 1 2 1 1 1 1 ] 1 0 0 The least-squares residual is ϵ 2 = Aˆx y 2 = 1 2 13/18

Solutions if the matrix A does not have full column rank A does not have full column rank the problem is change to min x X x 2 with X = {x : x = arg min Az y 2 } x A = [ U 1 U 2 ] [ Σ 0 0 0 ] [ ] V T 1 V T 2 the minimization problem can be written as min z = U 1 ΣV T 1 Az y 2 = min U 1 ΣV T z 1 z y 2 14/18

Solutions if the matrix A does not have full column rank define the partitioned vector the problem becomes ξ 1 = ξ 2 V 1 T V T 2 z min Az y 2 = min U 1 Σξ 1 y 2 z ξ 1 ˆξ 1 = Σ 1 U T 1 y from the normal equation ξ 2 does not change the value of the minimization problem and can be chosen arbitrarily 15/18

Solutions if the matrix A does not have full column rank the solutions become ] ẑ = [V 1 V 2 ˆξ 1 = V 1 Σ 1 U1 T y + V 2 ˆξ2 ˆξ 2 thus X = {x : x = V 1 Σ 1 U T 1 y + V 2 ˆξ2, ˆξ 2 R n r } x 2 = V 1 Σ 1 U T 1 y 2 + V 2 ˆξ2 2 select ˆξ 2 = 0, thus ˆx = V 1 Σ 1 U T 1 y 16/18

Example Given A = [ ] 1 2 1, y = 1 1 1 ˆx = A T (AA T ) 1 y 1 1 = 2 1 1 1 1 2 = 1 [ 3 2 2 2 3 ] [ 1 0 [ ] 1 0 ] 1 2 the 2-norm of this solution is ˆx = 3/2 17/18

Reference 1 Lecture note on Introduction to Linear Dynamical Systems, Stephen Boyd, Stanford University 2 Michel Verhaegen and Vincent Verdult Filtering and System Identification: A Least Squares Approach, Cambridge University Press, 2007 18/18