The Kalman filter is arguably one of the most notable algorithms

Size: px
Start display at page:

Download "The Kalman filter is arguably one of the most notable algorithms"

Transcription

1 LECTURE E NOTES «Kalman Filtering with Newton s Method JEFFREY HUMPHERYS and JEREMY WEST The Kalman filter is arguably one of the most notable algorithms of the 0th century [1]. In this article, we derive the Kalman filter using Newton s method for root finding. We show that the one-step Kalman filter is given by a single iteration of Newton s method on the gradient of a quadratic objective function, and with a judiciously chosen initial guess. This derivation is different from those found in standard texts [] [6], since it provides a more general framework for recursive state estimation. Although not presented here, this approach can also be used to derive the extended Kalman filter for nonlinear systems. BACKGROUND Linear Estimation Suppose that data are generated by the linear model b 5 Ax 1e, (1) where A is a known m 3 n matrix of rank n, e is an m-dimensional random variable with zero mean and known positive-definite covariance Q 5 E3ee T 4, denoted Q. 0, and b [ R m represents known, but inexact, measurements with errors given by e. The vector x [ R n is the set of parameters to be estimated. The estimator x^ is linear if x^ 5 Kb for some n 3 m matrix K. We say that the estimator x^ is unbiased if E3x^ 4 5 x. Since E3x^ 4 5 E3Kb4 5 E3K(Ax 1e)4 5 KAx 1 KE3e4 5 KAx, it follows that the linear estimator x^ 5 Kb is unbiased if and only if KA 5 I. When m. n, many choices of K could be made to satisfy this condition. We want to find the meansquared error-minimizing linear unbiased estimator, or, in other words, the minimizer of the quantity E3 7x^ x7 4 over all matrices K that satisfy the constraint KA 5 I. In What Is the Best Linear Unbiased Estimator, we prove the Gauss- Markov theorem, which states that the optimal K is given Digital Object Identifier /MCS by K 5 (A T Q A) A T Q, and thus the corresponding estimator is given by whose covariance is given by x^ 5 (A T Q A) A T Q b, () P 5 E3(x^ x)(x^ x) T 4 5 (A T Q A). (3) In addition, the Gauss-Markov theorem states that every other linear unbiased estimator of x has larger covariance than (3). Thus we call () the best linear unbiased estimator of (1). Weighted Least Squares It is not a coincidence that the estimator in () has the same form as the solution of the weighted least squares problem 1 x^ 5 argmin x[r n 7b Ax7 W, (4) where W. 0 is a problem-specific weighting matrix and 7 # 7 W describes a weighted -norm defined as 7x7 W J x T Wx. The factor 1/ is not necessary but makes calculations involving the derivative simpler. Scaling the objective function by a constant does not affect the minimizer. Therefore, consider the positive-definite quadratic objective function J(x) 5 1 7b Ax7 W 5 1 xt A T WAx x T A T Wb 1 1 bt Wb. (5) Its minimizer x^ satisfies =J(x^ ) 5 0, which yields the normal equations A T WAx^ 5 A T Wb. (6) Since A is assumed to have full column rank, the weighted Gramian A T WA is nonsingular, and thus (6) has the solution x^ 5 (A T WA) A T Wb. (7) Hence, in comparing () with (7), we see that the best linear unbiased estimator of (1) is found by solving the weighted least squares problem (4) with weight X/10/$ IEEE DECEMBER 010 «IEEE CONTROL SYSTEMS MAGAZINE 101

2 What Is the Best Linear Unbiased Estimator? Suppose that data are generated by the linear model b 5 Ax 1e, where A is a known m 3 n matrix of rank n, and e is an m-dimensional random variable with zero mean and covariance Q. 0. Recall that a linear estimator x^ 5 Kb of x is unbiased if and only if KA 5 I. If n, m, then there are many choices of K that are possible. We want the matrix K that minimizes the meansquared error E3 7 x^ x 7 4. Note, however, that 7x^ x7 5 7Kb x7 5 7KAx 1 Kex7 5 7Ke7 5e T K T Ke. Moreover, since e T K T Ke is a scalar, we have e T K T Ke5 tr(e T K T Ke) 5 tr(kee T K T ). It follows, by the linearity of expectation, that E3 7x^ x7 4 5 tr(ke3ee T 4K T ) 5 tr(kqk T ). Thus the mean-squared error minimizing linear unbiased estimator is found by minimizing tr(kqk T ) subject to the constraint KA 5 I, where K [ R n3m. This is a convex optimization problem, that is, the objective is a convex function and the feasible set is a convex set. Hence, to fi nd the unique minimizer, it suffi ces to fi nd the unique critical point of the Lagrangian L(K, l) 5 tr(kqk T ) tr(l T (KA I)), where the matrix Lagrange multiplier l [ R n3n corresponds to the n constraints KA 5 I. Thus the minimum occurs when 05= K L(K, l) 5 ' 'K tr(kqkt l T (KA I)) that is, when 5 KQ T 1 KQ la T 5 KQ la T, K 5 1 lat Q. Since KA 5 I, we have that l5(a T Q A). Therefore, K 5 (A T Q A) A T Q, and the optimal estimator is given by x^ 5 (A T Q A) A T Q b. To compute the covariance of the estimator we expand x^ as x^ 5 (A T Q A) A T Q (Ax 1e) 5 x 1 (A T Q A) A T Q e. Thus the covariance is given by P 5 E3(x^ x)(x^ x) T 4 5 (A T Q A) A T Q E3ee T 4Q A(A T Q A) 5 (A T Q A). We now show that every linear unbiased estimator other than x^ produces a larger covariance. If x^l 5 Lb is a linear unbiased estimator of the linear model, then since KA 5 I, it follows that there exists a matrix D such that DA 5 0 and L 5 K 1 D. The covariance of x^l is given by E3(x^L x)(x^l x) T 4 5 E3(K 1 D)ee T (K T 1 D T )4 5 (K 1 D)Q(K T 1 D T ) 5 KQK T 1 DQD T 1 KQD T 1 (KQD T ) T. Note, however, that KQD T 5 0. Thus E3(x^L x)(x^l x) T 4 5 KQK T 1 DQD T $ KQK T. Therefore, x^ is the best linear unbiased estimator of x. W 5 Q. 0. Also from (3) it can be noted that the inverse weighted-gramian matrix P 5 (A T Q A) is the covariance of this estimator. Newton s Method Let f : R n S R n be a smooth function. Assume that f(x^ ) 5 0 and that the Jacobian matrix Df(x) is nonsingular in a neighborhood of x^. If the initial guess x 0 is sufficiently close to x^, then Newton s method, which is given by x k11 5 x k Df(x k ) f(x k ), (8) ` produces a sequence 5x k 6 k50 that converges quadratically to x^, that is, there exists a constant c. 0 such that, for every positive integer k, 7x k11 x^ 7 # c7x k x^ 7. Newton s method is widely used in optimization problems since the local extrema of an objective function J(x) can be obtained by finding roots of its gradient f(x) J =J(x). Hence, from (8) we have x k11 5 x k D J(x k ) =J(x k ), (9) where D J(x k ) denotes the Hessian of J evaluated at x k. This iterative process converges, likewise at a quadratic rate, to the isolated local minimizer x^ of J whenever the starting point x 0 is sufficiently close to x^ and the Hessian D J(x) is positive definite in a neighborhood of x^. In practice, we do not invert the Hessian D J(x k ) when computing (9). Indeed, by a factor of roughly two, we can 10 IEEE CONTROL SYSTEMS MAGAZINE» DECEMBER 010

3 In this article, we derive the Kalman filter using Newton s method for root finding. more efficiently compute the Newton update by solving the linear system D J(x k )y k 5=J(x k ), for example by Gaussian elimination, and then setting x k11 5 x k 1 y k ; see [7] for details. To find the minimizer of a positive-definite quadratic form (5), it is not necessary to use an iterative scheme, such as Newton s method, since the gradient is affine in x and the unique minimizer can be found by solving the normal equations (6) directly. However, positivedefinite quadratic forms have the special property that Newton s method (9) converges in a single step for each starting point x [ R n, that is, for all x [ R n, the minimizer x^ satisfies x^ 5 x D J(x) =J(x). (10) This observation provides a key insight used to derive the Kalman filter below. Note also that the Hessian of the quadratic form (5) is D J(x) 5 A T WA, which is constant in x. Thus, throughout we denote the Hessian as D J, dropping the explicit dependence on x. Recursive Least Squares We now consider the least squares problem (4) in the case of recursive implementation. At each time k, we seek the best linear unbiased estimator x^ k of x for the linear model where b k 5 A k x 1e k, (11) b 1 A 1 v 1 b k 5 (, A k 5 (, and e k 5 (. b k A k v k We assume, for each j 5 1, c, k, that the noise terms v j have zero mean and are uncorrelated with E3v j v jt 4 5 R j. 0. We also assume that A 1 has full column rank, and thus each A k has full column rank. The estimate is found by solving the weighted least squares problem x^ k 5 argmin7 ba k x7 Wk, (1) x where the weight is given by the inverse covariance W k 5 E3e k e kt 4 5 diag(r 1, c, R k ). As time marches forward, the number of rows of the linear system increases, thus altering the least squares solution. We now show that it is possible to use the least squares solution x^ k at time k 1 to efficiently compute the least squares solution x^ k at time k. This process is the recursive least squares algorithm. We begin by rewriting (1) as J k (x) 5 1 a k i51 7b i A i x7 Ri. (13) The positive-definite quadratic form (13) can be expressed recursively as J k (x) 5 J k (x) 1 1 7b k A k x7 Rk. The gradient and Hessian of J k are given by and =J k (x) 5=J k (x) 1 A k T R k (A k x b k ) D J k 5 D J k 1 A k T R k A k, (14) respectively. Since A 1 has full column rank, we see that D J From (14), it follows that D J k. 0 for every positive integer k, and hence from (10) the minimizer of (13) becomes x^ k 5 x (D J k ) (=J k (x) 1 A k T R k (A k x b k )), (15) where the starting point x [ R n can be arbitrarily chosen. Since =J k (x^ k ) 5 0, we set x 5 x^ k in (15). Thus (15) becomes x^ k 5 x^ k K k A k T R k (A k x^ k b k ), where K k J (D J k ) is the inverse of the Hessian of J, and from (3) represents the covariance of the estimate. Observing from (14) that K k 5 K k 1 A T k R k A k and using Lemma 1 in Inversion Lemmata, we have K k 5 (K k 1 A T k R k A k ) 5 K k K k A kt (R k 1 A k K k A kt ) A k K k. Thus, to summarize, the recursive least squares method is K k 5 K k K k A kt (R k 1 A k K k A kt ) A k K k x^ k 5 x^ k K k A k T R k (A k x^ k b k ). At each time k, this algorithm gives the best linear unbiased estimator of (11), along with its covariance. DECEMBER 010 «IEEE CONTROL SYSTEMS MAGAZINE 103

4 The Kalman filter is arguably one of the most notable algorithms of the 0th century. The recursive least squares algorithm allows for efficient updating of the weighted least squares solution. Indeed, if p rows are added during the kth step, then the expression R k 1 A k K k A k T is a p 3 p matrix, which is small in size compared to the matrix A k W k A k T, which is needed, say, if we compute the solution directly using (7). STATE-SPACE MODELS AND STATE ESTIMATION Consider the stochastic discrete-time linear dynamic system x k11 5 F k x k 1 G k u k 1 w k, (16) y k 5 H k x k 1 v k, (17) where the variables represent the state x k [ R n, the input u k [ R p, and the output y k [ R q. The terms w k and v k are noise processes, which are assumed to have mean zero and be mutually uncorrelated, with known covariances Q k. 0 and R k. 0, respectively. We assume that the system has the stochastic initial state x 0 5m 0 1 w, where m 0 is the mean and w is a zero-mean random variable with covariance E3w w T 4 5 Q. 0. State Estimation We formulate the state estimation problem, given m known observations y 1, c, y m and k known inputs u 0, c, u k, where k $ m. To find the best linear unbiased estimator of the states x 1, c, x k, we begin by writing (16) (17) as the large linear estimation problem m 0 5 x 0 w, G 0 u 0 5 x 1 F 0 x 0 w 0, y 1 5 H 1 x 1 1 v 1, ( ( G m u m 5 x m F m x m w m, y m 5 H m x m 1 v m, G m u m 5 x m11 F m x m w m, ( ( G k u k 5 x k F k x k w k. Note that the known measurements, namely, the inputs and outputs, are on the left, whereas the unknown states are on the right, together with the noise terms. The parameters to be estimated in this linear estimation problem are the states x 0, x 1, c, x k, which we write as z k 5 3x 0 c xk 4 T [ R (k11)n. Then the linear system takes the form where b k m 5 A k m z k 1e k m, T T T e k m 5 3w w 0 v 1 c T T T wm v m w m c T wk is a zero-mean random variable whose inverse covariance is the positive-definite block-diagonal matrix W k m 5 diag(q, Q 0, R 1, c, Q m, R m, Q m c, Q k ). 4 T Inversion Lemmata We provide some technical lemmata that are used throughout the article. The first lemma is the Sherman Morrison Woodbury formula, which shows how to invert an additively updated matrix when we know the inverse of the original matrix. The second inversion lemma tells us how to invert block matrices. These statements can be verified by direct computation; see also [8, p ]. LEMMA 1 (SHERMAN-MORRISON-WOODBURY) Let A and C be square matrices and B and D be given so that the sum A 1 BCD is nonsingular. If A, C and C 1 DA B are also nonsingular, then (A 1 BCD) 5 A A B(C 1 DA B) DA. LEMMA (SCHUR) Let M be a square matrix with block form M 5 c A B C D d. If A, D, A BD C, and D CA B are nonsingular, then (A M BD C) A B(D CA B) 5 c D C(A BD C) (D CA B) d. 104 IEEE CONTROL SYSTEMS MAGAZINE» DECEMBER 010

5 The second observation is that the least-squares estimation problem can be solved by minimizing a positive-definite quadratic functional. We observe that each column of A k m is a pivot column. Hence, it follows that A k m has full column rank, and thus the weighted Gramian A T k m W k m A k m is nonsingular. Thus, the best linear unbiased estimator z^ k m of z k is given by the weighted least squares solution z^ k m 5 (A T k m W k m A k m ) A T k m W k m b k m. Since the state estimation problem is cast as a linear model, we can equivalently solve the estimation problem by minimizing the positive-definite objective function J k m ) 5 1 7A k mz k b k m 7 Wk m. (18) We can find its minimizer z^ k m by performing a single iteration of Newton s method given by z^ k m 5 z (A T k m W k m A k m ) A T k m W k m (A k m z b k m ), (19) where z [ R (k11)n is arbitrary. In the following, we set m 5 k and choose a canonical z that simplifies (19), thus leaving us with the Kalman filter. Kalman Derivation with Newton s Method Consider the state estimation problem in the case m 5 k, that is, the number of observations equals the number of inputs. It is of particular interest in applications to determine the current state x k given the observations y 1, c, y k and inputs u 0, c, u k. The Kalman filter gives the best linear unbiased estimator x^ k k of x k in terms of the previous state estimator x^ k k and the latest data u k and y k up to that point in time. We begin by rewriting (18) as J k m ) x 0 m 0 00 Q 1 1 m a 00y i H i x i 00 R i i k a 00x i F i x i G i u i 00 Q i. (0) i51 For convenience, we denote z^ k k as z^ k, x^ k k as x^ k, and J k k as J k. For m 5 k, (0) can be expressed recursively as J k ) 5 J k ) 1 1 7y k H k x k 7 Rk 1 1 7x k F k z k G k u k 7 Q k, where F k 5 30 c 0 F k 4. The gradient and Hessian of J k are given by =J k ) 5 and T =J k ) 1 F k Q k (F k z k x k 1 G k u k ) c Q k (F k z k x k 1 G k u k ) 1 H T k R k (H k x k y k ) d Q k Q k D J k 5 c D T J k 1 F T k F k F k Q k F k Q k 1 H T k R d, () k H k respectively. Since D J 0 5 Q. 0, it follows inductively that D J k. 0 for every positive integer k. The proof follows from the observation that z T k D T J k z k 5 z k D J k z k 1 7F k z k x k 7 Qk 1 7H k x k 7 Rk, and thus the right-hand side is nonnegative, being zero only if z k 5 0 and x k 5 0, or, equivalently, z k 5 0. From one iteration of Newton s method, we have that z^ k 5 z k (D J k ) =J k ) () for all z k [ R (k11)n. Since =J k (z^ k ) 5 0 and F k z^ k 5 F k x^ k, we set Thus, z^ k z k 5 c F k x^ k 1 G k u k. d. 0 =J k ) 5 c H T k R k 3H k (F k x^ k 1 G k u k ) y k 4 d, and the bottom row of () becomes x^ k 5 F k x^ k 1 G k u k P k H k T R k 3 3H k (F k x^ k 1 G k u k ) y k 4, where P k is the bottom-right block of the inverse Hessian (D J k ), which from Lemma in Inversion Lemmata, is given by P k 5 (Q k 1 H T k R k H k Q k F k 3 (D T J k 1 F k Q k F k ) T F k Q k ) 5 3(Q k 1 F k (D J k ) T F k ) 1 H T k R k H k 4 T 5 3(Q k 1 F k P k F k ) 1 H T k R k H k 4. DECEMBER 010 «IEEE CONTROL SYSTEMS MAGAZINE 105

6 Note that P k is the covariance of the estimator x^ k. To summarize, we have the recursive estimator for x k given by T P k 5 3(Q k 1 F k P k F k ) 1 H T k R k H k 4 x^ k 5 F k x^ k 1 G k u k P k H k T R k 3 3H k (F k x^ k 1 G k u k ) y k 4, where x^ 0 5m 0, P 0 5 Q. This recursive estimator is the onestep Kalman filter. CONCLUSIONS There are a few key observations in the Newton method derivation of the Kalman filter. First is that the state estimation problem for linear systems (16) (17) is a potentially large linear least squares estimation problem. The second observation is that the least-squares estimation problem can be solved by minimizing a positive-definite quadratic functional. The third point is that a positive-definite quadratic functional can be minimized with a single iteration of Newton s method for any starting point x [ R n. The final observation is that by judiciously choosing the starting point to be the solution of the previous state estimate, several terms cancel and the Newton update simplifies to the one-step Kalman filter. We remark that this approach to state estimation can be generalized to include predictive estimates on future states and smoothed estimates on previous states. We can also use Newton s method to derive the extended Kalman filter. ACKNOWLEDGMENTS The authors thank Randall W. Beard, Magnus Egerstedt, C. Shane Reese, and H. Dennis Tolley for useful conversations. This work was supported in part by the National Science Foundation, Division of Mathematical Sciences, grant numbers CAREER and CSUMS AUTHOR INFORMATION Jeffrey Humpherys received the Ph.D. in mathematics from Indiana University in 00 and was an Arnold Ross Assistant Professor at The Ohio State University from Since 005, he has been at Brigham Young University, where he is currently an associate professor of mathematics. His interests are in applied mathematical analysis and numerical computation. Jeremy West received B.S. degrees in mathematics and computer science and the M.S. in mathematics in 009, all from Brigham Young University. He is currently a Ph.D. student in mathematics at the University of Michigan. His interests are in discrete mathematics. REFERENCES [1] J. L. Casti, Five More Golden Rules: Knots, Codes, Chaos, and other Great Theories of 0th-Century Mathematics. New York: Wiley, 000. [] M. S. Grewal and A. P. Andrews, Kalman Filtering: Theory and Practice Using MATLAB, nd ed. New York: Wiley, 001. [3] D. Simon, Optimal State Estimation: Kalman, H-Infinity, and Nonlinear Approaches. New York: Wiley-Interscience, 006. [4] T. Kailath, A. Sayed, and B. Hassibi, Linear Estimation. Englewood Cliffs, NJ: Prentice-Hall, 000. [5] R. Brown and P. Hwang, Introduction to Random Signals and Applied Kalman Filtering. New York: Wiley, [6] A. Gelb, Applied Optimal Estimation. Cambridge, MA: MIT Press, [7] C. T. Kelley, Solving Nonlinear Equations with Newton s Method. Philadelphia, PA: SIAM, 003. [8] R. A. Horn and C. R. Johnson, Matrix Analysis. Cambridge, U.K.: Cambridge Univ. Press, IEEE CONTROL SYSTEMS MAGAZINE» DECEMBER 010

Derivation of the Kalman Filter

Derivation of the Kalman Filter Derivation of the Kalman Filter Kai Borre Danish GPS Center, Denmark Block Matrix Identities The key formulas give the inverse of a 2 by 2 block matrix, assuming T is invertible: T U 1 L M. (1) V W N P

More information

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any

More information

Least Squares and Kalman Filtering Questions: me,

Least Squares and Kalman Filtering Questions:  me, Least Squares and Kalman Filtering Questions: Email me, namrata@ece.gatech.edu Least Squares and Kalman Filtering 1 Recall: Weighted Least Squares y = Hx + e Minimize Solution: J(x) = (y Hx) T W (y Hx)

More information

ECE133A Applied Numerical Computing Additional Lecture Notes

ECE133A Applied Numerical Computing Additional Lecture Notes Winter Quarter 2018 ECE133A Applied Numerical Computing Additional Lecture Notes L. Vandenberghe ii Contents 1 LU factorization 1 1.1 Definition................................. 1 1.2 Nonsingular sets

More information

Least Squares Estimation Namrata Vaswani,

Least Squares Estimation Namrata Vaswani, Least Squares Estimation Namrata Vaswani, namrata@iastate.edu Least Squares Estimation 1 Recall: Geometric Intuition for Least Squares Minimize J(x) = y Hx 2 Solution satisfies: H T H ˆx = H T y, i.e.

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Constrained Optimization

Constrained Optimization 1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange

More information

MATH 4211/6211 Optimization Quasi-Newton Method

MATH 4211/6211 Optimization Quasi-Newton Method MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:

More information

Chapter 3 Numerical Methods

Chapter 3 Numerical Methods Chapter 3 Numerical Methods Part 2 3.2 Systems of Equations 3.3 Nonlinear and Constrained Optimization 1 Outline 3.2 Systems of Equations 3.3 Nonlinear and Constrained Optimization Summary 2 Outline 3.2

More information

A Characterization of the Hurwitz Stability of Metzler Matrices

A Characterization of the Hurwitz Stability of Metzler Matrices 29 American Control Conference Hyatt Regency Riverfront, St Louis, MO, USA June -2, 29 WeC52 A Characterization of the Hurwitz Stability of Metzler Matrices Kumpati S Narendra and Robert Shorten 2 Abstract

More information

Course Summary Math 211

Course Summary Math 211 Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.

More information

RECURSIVE ESTIMATION AND KALMAN FILTERING

RECURSIVE ESTIMATION AND KALMAN FILTERING Chapter 3 RECURSIVE ESTIMATION AND KALMAN FILTERING 3. The Discrete Time Kalman Filter Consider the following estimation problem. Given the stochastic system with x k+ = Ax k + Gw k (3.) y k = Cx k + Hv

More information

4 Derivations of the Discrete-Time Kalman Filter

4 Derivations of the Discrete-Time Kalman Filter Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time

More information

Computational Methods. Least Squares Approximation/Optimization

Computational Methods. Least Squares Approximation/Optimization Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the

More information

Particle Filters. Outline

Particle Filters. Outline Particle Filters M. Sami Fadali Professor of EE University of Nevada Outline Monte Carlo integration. Particle filter. Importance sampling. Degeneracy Resampling Example. 1 2 Monte Carlo Integration Numerical

More information

The Solution of Linear Systems AX = B

The Solution of Linear Systems AX = B Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has

More information

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

March 8, 2010 MATH 408 FINAL EXAM SAMPLE March 8, 200 MATH 408 FINAL EXAM SAMPLE EXAM OUTLINE The final exam for this course takes place in the regular course classroom (MEB 238) on Monday, March 2, 8:30-0:20 am. You may bring two-sided 8 page

More information

Math 313 Chapter 1 Review

Math 313 Chapter 1 Review Math 313 Chapter 1 Review Howard Anton, 9th Edition May 2010 Do NOT write on me! Contents 1 1.1 Introduction to Systems of Linear Equations 2 2 1.2 Gaussian Elimination 3 3 1.3 Matrices and Matrix Operations

More information

17 Solution of Nonlinear Systems

17 Solution of Nonlinear Systems 17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m

More information

Estimation, Detection, and Identification

Estimation, Detection, and Identification Estimation, Detection, and Identification Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Chapter 5 Best Linear Unbiased Estimators Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

14. Nonlinear equations

14. Nonlinear equations L. Vandenberghe ECE133A (Winter 2018) 14. Nonlinear equations Newton method for nonlinear equations damped Newton method for unconstrained minimization Newton method for nonlinear least squares 14-1 Set

More information

Square-Root Algorithms of Recursive Least-Squares Wiener Estimators in Linear Discrete-Time Stochastic Systems

Square-Root Algorithms of Recursive Least-Squares Wiener Estimators in Linear Discrete-Time Stochastic Systems Proceedings of the 17th World Congress The International Federation of Automatic Control Square-Root Algorithms of Recursive Least-Squares Wiener Estimators in Linear Discrete-Time Stochastic Systems Seiichi

More information

LINEAR SYSTEMS AND MATRICES

LINEAR SYSTEMS AND MATRICES CHAPTER 3 LINEAR SYSTEMS AND MATRICES SECTION 3. INTRODUCTION TO LINEAR SYSTEMS This initial section takes account of the fact that some students remember only hazily the method of elimination for and

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Quasi-Newton methods for minimization

Quasi-Newton methods for minimization Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1

More information

(Extended) Kalman Filter

(Extended) Kalman Filter (Extended) Kalman Filter Brian Hunt 7 June 2013 Goals of Data Assimilation (DA) Estimate the state of a system based on both current and all past observations of the system, using a model for the system

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2

More information

Riccati difference equations to non linear extended Kalman filter constraints

Riccati difference equations to non linear extended Kalman filter constraints International Journal of Scientific & Engineering Research Volume 3, Issue 12, December-2012 1 Riccati difference equations to non linear extended Kalman filter constraints Abstract Elizabeth.S 1 & Jothilakshmi.R

More information

The Discrete Kalman Filtering of a Class of Dynamic Multiscale Systems

The Discrete Kalman Filtering of a Class of Dynamic Multiscale Systems 668 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL 49, NO 10, OCTOBER 2002 The Discrete Kalman Filtering of a Class of Dynamic Multiscale Systems Lei Zhang, Quan

More information

KKT Conditions for Rank-Deficient Nonlinear Least-Square Problems with Rank-Deficient Nonlinear Constraints1

KKT Conditions for Rank-Deficient Nonlinear Least-Square Problems with Rank-Deficient Nonlinear Constraints1 JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS: Vol. 100, No. 1. pp. 145-160, JANUARY 1999 KKT Conditions for Rank-Deficient Nonlinear Least-Square Problems with Rank-Deficient Nonlinear Constraints1

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

Jim Lambers MAT 419/519 Summer Session Lecture 11 Notes

Jim Lambers MAT 419/519 Summer Session Lecture 11 Notes Jim Lambers MAT 49/59 Summer Session 20-2 Lecture Notes These notes correspond to Section 34 in the text Broyden s Method One of the drawbacks of using Newton s Method to solve a system of nonlinear equations

More information

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical

More information

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I

More information

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz C&O367: Nonlinear Optimization (Winter 013) Assignment 4 H. Wolkowicz Posted Mon, Feb. 8 Due: Thursday, Feb. 8 10:00AM (before class), 1 Matrices 1.1 Positive Definite Matrices 1. Let A S n, i.e., let

More information

A Matrix Theoretic Derivation of the Kalman Filter

A Matrix Theoretic Derivation of the Kalman Filter A Matrix Theoretic Derivation of the Kalman Filter 4 September 2008 Abstract This paper presents a matrix-theoretic derivation of the Kalman filter that is accessible to students with a strong grounding

More information

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx.

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx. Two hours To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER CONVEX OPTIMIZATION - SOLUTIONS xx xxxx 27 xx:xx xx.xx Answer THREE of the FOUR questions. If

More information

Review of Vectors and Matrices

Review of Vectors and Matrices A P P E N D I X D Review of Vectors and Matrices D. VECTORS D.. Definition of a Vector Let p, p, Á, p n be any n real numbers and P an ordered set of these real numbers that is, P = p, p, Á, p n Then P

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by

More information

Numerical solutions of nonlinear systems of equations

Numerical solutions of nonlinear systems of equations Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points

More information

Numerical solution of Least Squares Problems 1/32

Numerical solution of Least Squares Problems 1/32 Numerical solution of Least Squares Problems 1/32 Linear Least Squares Problems Suppose that we have a matrix A of the size m n and the vector b of the size m 1. The linear least square problem is to find

More information

GLOBALLY CONVERGENT GAUSS-NEWTON METHODS

GLOBALLY CONVERGENT GAUSS-NEWTON METHODS GLOBALLY CONVERGENT GAUSS-NEWTON METHODS by C. Fraley TECHNICAL REPORT No. 200 March 1991 Department of Statistics, GN-22 University of Washington Seattle, Washington 98195 USA Globally Convergent Gauss-Newton

More information

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review 10.34 Numerical Methods Applied to Chemical Engineering Fall 2015 Quiz #1 Review Study guide based on notes developed by J.A. Paulson, modified by K. Severson Linear Algebra We ve covered three major topics

More information

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson

More information

Extended Kalman Filter Tutorial

Extended Kalman Filter Tutorial Extended Kalman Filter Tutorial Gabriel A. Terejanu Department of Computer Science and Engineering University at Buffalo, Buffalo, NY 14260 terejanu@buffalo.edu 1 Dynamic process Consider the following

More information

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn Today s class Linear Algebraic Equations LU Decomposition 1 Linear Algebraic Equations Gaussian Elimination works well for solving linear systems of the form: AX = B What if you have to solve the linear

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

2.098/6.255/ Optimization Methods Practice True/False Questions

2.098/6.255/ Optimization Methods Practice True/False Questions 2.098/6.255/15.093 Optimization Methods Practice True/False Questions December 11, 2009 Part I For each one of the statements below, state whether it is true or false. Include a 1-3 line supporting sentence

More information

Conjugate Gradient (CG) Method

Conjugate Gradient (CG) Method Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous

More information

6.4 Kalman Filter Equations

6.4 Kalman Filter Equations 6.4 Kalman Filter Equations 6.4.1 Recap: Auxiliary variables Recall the definition of the auxiliary random variables x p k) and x m k): Init: x m 0) := x0) S1: x p k) := Ak 1)x m k 1) +uk 1) +vk 1) S2:

More information

Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise Covariance

Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise Covariance 2016 American Control Conference (ACC) Boston Marriott Copley Place July 6-8, 2016. Boston, MA, USA Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise

More information

Interval solutions for interval algebraic equations

Interval solutions for interval algebraic equations Mathematics and Computers in Simulation 66 (2004) 207 217 Interval solutions for interval algebraic equations B.T. Polyak, S.A. Nazin Institute of Control Sciences, Russian Academy of Sciences, 65 Profsoyuznaya

More information

Two improved classes of Broyden s methods for solving nonlinear systems of equations

Two improved classes of Broyden s methods for solving nonlinear systems of equations Available online at www.isr-publications.com/jmcs J. Math. Computer Sci., 17 (2017), 22 31 Research Article Journal Homepage: www.tjmcs.com - www.isr-publications.com/jmcs Two improved classes of Broyden

More information

Digital Workbook for GRA 6035 Mathematics

Digital Workbook for GRA 6035 Mathematics Eivind Eriksen Digital Workbook for GRA 6035 Mathematics November 10, 2014 BI Norwegian Business School Contents Part I Lectures in GRA6035 Mathematics 1 Linear Systems and Gaussian Elimination........................

More information

Algorithms for nonlinear programming problems II

Algorithms for nonlinear programming problems II Algorithms for nonlinear programming problems II Martin Branda Charles University Faculty of Mathematics and Physics Department of Probability and Mathematical Statistics Computational Aspects of Optimization

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate CONVERGENCE ANALYSIS OF THE LATOUCHE-RAMASWAMI ALGORITHM FOR NULL RECURRENT QUASI-BIRTH-DEATH PROCESSES CHUN-HUA GUO Abstract The minimal nonnegative solution G of the matrix equation G = A 0 + A 1 G +

More information

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance The Kalman Filter Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience Sarah Dance School of Mathematical and Physical Sciences, University of Reading s.l.dance@reading.ac.uk July

More information

Solvability of Linear Matrix Equations in a Symmetric Matrix Variable

Solvability of Linear Matrix Equations in a Symmetric Matrix Variable Solvability of Linear Matrix Equations in a Symmetric Matrix Variable Maurcio C. de Oliveira J. William Helton Abstract We study the solvability of generalized linear matrix equations of the Lyapunov type

More information

Elementary maths for GMT

Elementary maths for GMT Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1

More information

Numerical Methods for Large-Scale Nonlinear Systems

Numerical Methods for Large-Scale Nonlinear Systems Numerical Methods for Large-Scale Nonlinear Systems Handouts by Ronald H.W. Hoppe following the monograph P. Deuflhard Newton Methods for Nonlinear Problems Springer, Berlin-Heidelberg-New York, 2004 Num.

More information

Lecture 6. Regularized least-squares and minimum-norm methods 6 1

Lecture 6. Regularized least-squares and minimum-norm methods 6 1 Regularized least-squares and minimum-norm methods 6 1 Lecture 6 Regularized least-squares and minimum-norm methods EE263 Autumn 2004 multi-objective least-squares regularized least-squares nonlinear least-squares

More information

Lecture 7. Gaussian Elimination with Pivoting. David Semeraro. University of Illinois at Urbana-Champaign. February 11, 2014

Lecture 7. Gaussian Elimination with Pivoting. David Semeraro. University of Illinois at Urbana-Champaign. February 11, 2014 Lecture 7 Gaussian Elimination with Pivoting David Semeraro University of Illinois at Urbana-Champaign February 11, 2014 David Semeraro (NCSA) CS 357 February 11, 2014 1 / 41 Naive Gaussian Elimination

More information

Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently

More information

5.6. PSEUDOINVERSES 101. A H w.

5.6. PSEUDOINVERSES 101. A H w. 5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and

More information

5 Quasi-Newton Methods

5 Quasi-Newton Methods Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min

More information

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...

More information

Matrices and systems of linear equations

Matrices and systems of linear equations Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.

More information

Prediction, filtering and smoothing using LSCR: State estimation algorithms with guaranteed confidence sets

Prediction, filtering and smoothing using LSCR: State estimation algorithms with guaranteed confidence sets 2 5th IEEE Conference on Decision and Control and European Control Conference (CDC-ECC) Orlando, FL, USA, December 2-5, 2 Prediction, filtering and smoothing using LSCR: State estimation algorithms with

More information

Algorithms for nonlinear programming problems II

Algorithms for nonlinear programming problems II Algorithms for nonlinear programming problems II Martin Branda Charles University in Prague Faculty of Mathematics and Physics Department of Probability and Mathematical Statistics Computational Aspects

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Jim Lambers MAT 610 Summer Session Lecture 1 Notes

Jim Lambers MAT 610 Summer Session Lecture 1 Notes Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra

More information

MATRICES. a m,1 a m,n A =

MATRICES. a m,1 a m,n A = MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of

More information

YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL

YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL Journal of Comput. & Applied Mathematics 139(2001), 197 213 DIRECT APPROACH TO CALCULUS OF VARIATIONS VIA NEWTON-RAPHSON METHOD YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL Abstract. Consider m functions

More information

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...

More information

Math 273a: Optimization Netwon s methods

Math 273a: Optimization Netwon s methods Math 273a: Optimization Netwon s methods Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 some material taken from Chong-Zak, 4th Ed. Main features of Newton s method Uses both first derivatives

More information

7.6 The Inverse of a Square Matrix

7.6 The Inverse of a Square Matrix 7.6 The Inverse of a Square Matrix Copyright Cengage Learning. All rights reserved. What You Should Learn Verify that two matrices are inverses of each other. Use Gauss-Jordan elimination to find inverses

More information

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 11 and

More information

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018 1 Linear Systems Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March, 018 Consider the system 4x y + z = 7 4x 8y + z = 1 x + y + 5z = 15. We then obtain x = 1 4 (7 + y z)

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

Optimization and Root Finding. Kurt Hornik

Optimization and Root Finding. Kurt Hornik Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding

More information

Notes on Mathematics

Notes on Mathematics Notes on Mathematics - 12 1 Peeyush Chandra, A. K. Lal, V. Raghavendra, G. Santhanam 1 Supported by a grant from MHRD 2 Contents I Linear Algebra 7 1 Matrices 9 1.1 Definition of a Matrix......................................

More information

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix EE507 - Computational Techniques for EE 7. LU factorization Jitkomut Songsiri factor-solve method LU factorization solving Ax = b with A nonsingular the inverse of a nonsingular matrix LU factorization

More information

the Unitary Polar Factor æ Ren-Cang Li Department of Mathematics University of California at Berkeley Berkeley, California 94720

the Unitary Polar Factor æ Ren-Cang Li Department of Mathematics University of California at Berkeley Berkeley, California 94720 Relative Perturbation Bounds for the Unitary Polar Factor Ren-Cang Li Department of Mathematics University of California at Berkeley Berkeley, California 9470 èli@math.berkeley.eduè July 5, 994 Computer

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2 MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Linear Riccati Dynamics, Constant Feedback, and Controllability in Linear Quadratic Control Problems

Linear Riccati Dynamics, Constant Feedback, and Controllability in Linear Quadratic Control Problems Linear Riccati Dynamics, Constant Feedback, and Controllability in Linear Quadratic Control Problems July 2001 Revised: December 2005 Ronald J. Balvers Douglas W. Mitchell Department of Economics Department

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

JACOBI S ITERATION METHOD

JACOBI S ITERATION METHOD ITERATION METHODS These are methods which compute a sequence of progressively accurate iterates to approximate the solution of Ax = b. We need such methods for solving many large linear systems. Sometimes

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Lemma 8: Suppose the N by N matrix A has the following block upper triangular form:

Lemma 8: Suppose the N by N matrix A has the following block upper triangular form: 17 4 Determinants and the Inverse of a Square Matrix In this section, we are going to use our knowledge of determinants and their properties to derive an explicit formula for the inverse of a square matrix

More information

Quaternion Data Fusion

Quaternion Data Fusion Quaternion Data Fusion Yang Cheng Mississippi State University, Mississippi State, MS 39762-5501, USA William D. Banas and John L. Crassidis University at Buffalo, State University of New York, Buffalo,

More information

Systems of Linear Equations and Matrices

Systems of Linear Equations and Matrices Chapter 1 Systems of Linear Equations and Matrices System of linear algebraic equations and their solution constitute one of the major topics studied in the course known as linear algebra. In the first

More information