Optimal Estimation of Dynamic Systems

Size: px
Start display at page:

Download "Optimal Estimation of Dynamic Systems"

Transcription

1 Probability Concepts in Optimal Estimation of Dynamic Systems Kamesh Subbarao Associate Professor Department of Mechanical and Aerospace Engineering The University of Texas at Arlington Phone: (214) / 58

2 Outline Probability Concepts in. Least squares estimation. Linear batch estimation (linear least squres, weighted least squares, constrained least squares) Linear sequential estimation Nonlinear sequential estimation Advanced topics Examples Introduction to probability, random processes & statistics. Probability concepts in Least squares. Minimum variance estimation (with and without a priori estimates) Maximum likelihood estimation (MLE) Cramer-Rao inequality Bayesian estimation Advanced topics Examples 2 / 58

3 Probability Concepts in Vectors & matrices Vectors & matrices Basic Linear Algebra and Matrix Calculus A quantity A R m n denotes a matrix with m rows and n columns, wherein each entry of the matrix, a ij located at the i th row and the j th column is a real number. A quantity x R n 1 is a column matrix. Vectors can be considered as a special case of matrices. Row vectors when, x R 1 m and column vectors when x R n 1. Without loss of generality, in this course we will consider all vectors as column vectors. A T represents the transpose of the matrix A. The entries of A T are then a ji. The transpose of a column vector is a row vector. If x 1 x = x 2 : then x T = [x 1 x 2.. x n ]. x n 3 / 58

4 Vectors Probability Concepts in Vectors & matrices Basic Linear Algebra and Matrix Calculus Let x = [x 1, x 2,, x n ] T and y = [y 1, y 2,, y n ] T Vector Norm: A measure of the length of a vector (also called norm) x = x T x = [ n i=1 x 2 i ] 1/2 If α is a scalar, αx = α x, where. denotes the absolute value. Unit Vector: ˆx = x. The carat will will used to denote the estimate x of the variable later (all variable embellishments are purely contextual). Dot/Scalar Product: x T y = y T x = n i=1 x iy i (Note: the dot product gives us the norm of the vector). If the dot product is zero, the two vectors are orthogonal. An orthonormal set of vectors when the dot product is zero and the norms of the vectors equal unity. 4 / 58

5 Vectors Probability Concepts in Vectors & matrices Basic Linear Algebra and Matrix Calculus Triangle Inequality: x + y x + y Cauchy-Schwarz Inequality: 0 x T y x y 1 Orthogonal Projection: The orthogonal projection of y on to x is given by, p = x T y x 2 x Cross/Vector Product: z = x y = [x ] y where, 0 x 3 x 2 [x ] = x 3 0 x 1. Observe, [x ] is a skew-symmetric x 2 x 1 0 matrix and is called the cross-product operator here. 5 / 58

6 Probability Concepts in Vectors & matrices Basic Linear Algebra and Matrix Calculus Matrices The system of m linear equations, y 1 = a 11 x 1 + a 12 x a 1n x n y 2 = a 21 x 1 + a 22 x a 2n x n. y m = a m1 x 1 + a m2 x a mn x n can be written in a matrix form as, y = Ax, where y R m, A R m n and x R n. (Note: all of the components of x are being mapped into y via the matrix A). The rank of A is the smaller of the number of linearly independent rows and the number of linearly independent columns. If m = n, the matrix A is square. The solution for x is determined if m = n, under-determined if n > m = rank(a) (an infinity of solutions possible), over-determined if m > n = rank(a) (usually no exact solution exists) If the rank conditions are not satisfied for the solution, we can use the generalized inverse. 6 / 58

7 Probability Concepts in Vectors & matrices Basic Linear Algebra and Matrix Calculus Matrix Addition, Subtraction and Multiplication For addition and subtraction, matrices must have the same dimension. i.e. C = A ± B implies, c ij = a ij ± b ij. Matrix addition and subtraction is commutative and associative. Matrix multiplication requires conformable matrices. If C = AB, then number of columns of A should be equal to the number of rows of B. c ij = n k=1 a ikb kj. Matrix multiplication is associative but not commutative in general. (αa) T = αa T, where α is a scalar (A + B) T = A T + B T (AB) T = B T A T Matrix Inverse: For a square matrix A, the following statements are equivalent A has linearly independent columns A has linearly independent rows The inverse satisfies A 1 A = AA 1 = I A nonsingular matrix is one whose inverse exists (likewise A T is nonsingular): (A 1) 1 = A; (A ) T 1 ( = A 1) T = A T 7 / 58

8 Probability Concepts in Matrix Inverse and Trace Vectors & matrices Basic Linear Algebra and Matrix Calculus For two nonsingular matrices, A and B: (AB) 1 = B 1 A 1 For an orthonormal matrix C, C 1 = C T. The determinant of an orthonormal matrix can be shown to be ±1. An orthonormal matrix preserves the length of a vector. Thus if C is an orthonormal matrix, Cx = x. In general for an orthogonal matrix, D, D T D = DD T = det(d)i Sherman-Morrison Lemma: For conformable matrices A and B: (I + AB) 1 = I A (I + BA) 1 B Matrix Inversion Lemma: For conformable ( matrices A, B, C and D: (A + BCD) 1 = A 1 A 1 B DA 1 B + C 1) 1 DA 1 Trace of a Matrix (square matrices only): Another quantity used in estimation and control theory. It is the sum of all the diagonal elements. Tr(A) = n i=1 a ii Tr(αA) = αtr(a); Tr(A + B) = Tr(A) + Tr(B) Tr(AB) = Tr(BA); Tr(xy T ) = x T y; Tr(Axy T ) = x T Ay xy T is also known as the outer product of the vectors x and y. In general, xy T yx T 8 / 58

9 Probability Concepts in Matrix Definiteness Vectors & matrices Basic Linear Algebra and Matrix Calculus Definiteness: Often required when we look for sufficiency, stability & convergence tests when dealing with functions involving multiple variables. A real and square matrix A is Positive definite if x T Ax > 0 for all non-zero x Positive semi-definite if x T Ax 0 for all non-zero x Negative definite if x T Ax < 0 for all non-zero x Negative semi-definite if x T Ax 0 for all non-zero x Indefinite when no definiteness can be asserted Notice that the definiteness of a matrix is inferred through a scalar measure. The scalar measure, x T Ax is termed a Quadratic form. Thus for two real square matrices, if one writes, A > B, the implication is A B is positive-definite or x T (A B) x > 0. 9 / 58

10 Probability Concepts in Basic Linear Algebra Vectors & matrices Basic Linear Algebra and Matrix Calculus Matrix Rank and Nullity: The rank of a matrix is given by the dimension of the range of the matrix corresponding to the number of linearly independent rows or columns. An m n matrix is rank-deficient if the rank of A is less than min(m, n). Suppose that the rank of a n n matrix A is given by rank(a) = r. Then a set of (n r) nonzero unit vectors ˆx i can be found such that: Aˆx i = 0, i = 1, 2,..., n r (n r) is the nullity, which is the maximum number of linearly independent null vectors of A. Eigenvalues/Eigenvectors of a Matrix: Ap = λp. λ is the eigenvalue and p is the nonzero (right) eigenvector. If the eigenvalues are distinct, then the set of eigenvectors is linearly independent. Then, the matrix A can be diagonalized as, Λ = P 1 AP, where Λ = diag [λ 1 λ 2 λ n] and P = [p 1 p 2 p n ] Note, if A is symmetric then P is orthogonal. 10 / 58

11 Probability Concepts in Basic Linear Algebra Vectors & matrices Basic Linear Algebra and Matrix Calculus QR Decomposition: This is useful in least squares and the Square Root Information Filter (SRIF). The QR decomposition of an m n matrix A is given by, A = QR, where Q is an m m orthogonal matrix and R is an upper triangular m n matrix with all elements R ij = 0 for i > j. If A has full column rank, then the first n columns of Q form an orthonormal basis for the range of A. Singular Value Decomposition: This decomposes an m n matrix A into a diagnoal matrix and two orthogonal matrices: A = USV, where U is an m m unitary matrix, S is an m n matrix wherein S ij = 0 for i j and V is an n n unitary matrix. Let A = USV be the reduced representation where the (n + 1) and higher rows of S (correspondingly the columns of U) are eliminated The elements of S = diag[s 1 s 2 s n] are termed the singular values of A and ordered smallest to largest. These values tell us how well one can invert a matrix. Condition Number = sn. Large condition numbers indicate a near s 1 singular matrix. 11 / 58

12 Probability Concepts in Vectors & matrices Basic Linear Algebra and Matrix Calculus Basic Linear Algebra & Matrix Calculus LU and Cholesky Decompositions: The LU decomposition factors an n n matrix A into a product of a lower triangular matrix L and an upper triangular matrix U, so that A = LU. For symmetric positive definite matrices, A = LL T, wherein L is known as the matrix square root and the factorization is known as the Cholesky decomposition. Matrix Calculus: Consider a scalar function f (x), where x is an n 1 vector. The Jacobian (gradient) of f (x) is an n 1 vector given by, xf f x = f x 1 f x 2 : f x n 12 / 58

13 Probability Concepts in Matrix Calculus Vectors & matrices Basic Linear Algebra and Matrix Calculus The Hessian of f (x) is an n 1 vector given by, 2 f 2 f 2 f x1 2 x 1 x 2 x 1 x n 2 f 2 f 2 f 2 xf 2 f x x = x T 2 x 1 x2 2 x 2 x n f 2 f 2 f x n x 1 x n x 2 xn 2 Note, that the Hessian of a scalar function is a symmetric matrix. If f (x) is an m 1 vector and x is an n 1 vector, then the Jacobian matrix is an m n matrix. f 1 f 1 f 1 x 1 x 2 x n xf f f 2 f 2 f 2 x = x 1 x 2 x n f m f m f m 13 / 58

14 Probability Concepts in Matrix Calculus Vectors & matrices Basic Linear Algebra and Matrix Calculus A list of derivatives involving some products are given below, x (Ax) = A A (x T Ay) = xy T A (x T A T y) = yx T x (x T A T x) = (A + A T )x A very comprehensive list of identities and other Matrix manipulation tools can be found here. Matrix identitites: Matrix manipulations: http: //home.online.no/ pjacklam/matlab/doc/mtt/index.html More Free Stuff: jburkardt/m_src/m_src.html 14 / 58

15 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Linear For a measurable quantity, x, the following two equations hold, measured value = true value + measurement error x = x + v measured value = estimated value + residual error x = ˆx + e Note, The actual measurement error (v) and the true value (x) is never known in practice. The errors in the process/mechanism that physically generate this error are usually approximated by some known process (Gaussian noise with known variance). The statistical properties are utilized to weight the relative importance of the various measurements used in the estimation scheme. The residual error is known exactly and is easily computed once an estimated value has been found. The residual error drives the estimator. 15 / 58

16 Probability Concepts in Linear Batch Estimation Linear Batch Estimation Linear Sequential Estimation Consider a batch of measurements obtained at discrete instants of time: {(ỹ 1, t 1 ), (ỹ 2, t 2 ), (ỹ 3, t 3 ),, (ỹ m, t m)} We wish to model these measurements via a mathematical model. (REMEMBER: YOU ARE PROPOSING THE MODEL!) y(t) = n x i h i (t), i=1 m n where, h i (t) {h 1 (t), h 2 (t), h 3 (t),, h n(t)} are a set of independent specified basis functions. x i are the constants whose values are to be determined. We seek to obtain optimum x-values based upon a measure of how well the model predicts the measurements. Errors in the prediction are usually due to, measurement errors incorrect x-values modelling errors, i.e. the proposed model was bad. 16 / 58

17 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Linear Batch Estimation Lets relate the measurements (ỹ j ) and the estimated outputs (ŷ j ). ỹ j ỹ(t j ) = n x i h i (t j ) + v j i=1 j = 1, 2,..., m The estimated outputs are computed using the estimated values of x and the basis functions. ŷ j ŷ(t j ) = n ˆx i h i (t j ) i=1 j = 1, 2,..., m What about v j? Clearly, ỹ j = n i=1 ˆx ih i (t j ) + e j, where e j is the residual error, e j = ỹ j ŷ j. We can compactly represent the above by combining the measurements at all time instants and stacking them up as, ỹ = H ˆx + e (1) 17 / 58

18 Probability Concepts in Linear Batch Estimation Linear Batch Estimation Linear Sequential Estimation where, ỹ = [ỹ 1 ỹ 2 ỹ m] T = measured y values e = [ẽ 1 ẽ 2 ẽ m] T = residual errors ˆx = [ˆx 1 ˆx 2 ˆx n] T = estimated x values h 1 (t 1 ) h 2 (t 1 ) h n(t 1 ) h 1 (t 2 ) h 2 (t 2 ) h n(t 2 ) H =... h 1 (t m) h 2 (t m) h n(t m) Similarly one can also develop the following equations, ỹ = Hx + v (2) ŷ = H ˆx (3) Equations (1) and (2) are commonly referred to as the observation equations. 18 / 58

19 Probability Concepts in Linear Batch Estimation Linear Batch Estimation Linear Sequential Estimation Gauss least squares principle selects the optimum ˆx by minimizing the sum square of the residual errors, given by J = 1 2 et e = 1 2 (ỹ H ˆx)T (ỹ H ˆx) or J = J(ˆx) = 1 2 ( ) ỹ T ỹ 2ỹ T H ˆx + ˆx T H T H ˆx The multiplier 1/2 has a statistical significance (will be discussed later). Now, use matrix calculus differentiation rules to obtain the necessary condition for the minimum. ( 1 Necessary condition: ˆx J = H T H ˆx H T ỹ = 0, i.e., ˆx = H H) T HT ỹ Sufficient condition: 2ˆx J 2 J ˆx ˆx T = HT H > 0 (H T H is positive definite) The inverse of H T H is required. A good choice of the basis functions is important. 19 / 58

20 Probability Concepts in Linear Batch Estimation Linear Batch Estimation Linear Sequential Estimation Example 1: Scalar dynamical system. ẏ = ay + bu. u is external input; a and b are constants. Consider the discrete equivalent for a constant sampling interval t, y k+1 = Φy k + Γu k. Note, Φ = e a t and Γ = t 0 be a t dt = b a ( e a t 1 ) Given measurements, y(t k ) for an impulse input u 1 = 100, u k = 0, fork 2 obtain estimates of Φ and Γ values See Example 1 Try different impulse values and analyze the performance of the batch estimator. Try different noise standard deviations and compare performance. 20 / 58

21 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Weighted If each measurement is made with different precisions, it is better if this aspect is captured by weighting the measurements accordingly. The choice of the weights is non-unique but it turns out that the inverse of the measurement error variance is an intuitive choice. J = 1 2 et W e Necessary condition: ˆx J = H T W H ˆx H T W ỹ = 0, i.e., ( 1 ˆx = H T W H) H T W ỹ Sufficient condition: 2ˆx J 2 J ˆx ˆx T = HT W H > 0 (H T W H is positive definite) W is typically chosen to be a diagonal matrix. A subset of the w ii are chosen much larger than the others to reflect the preciseness of that specific subset of measurements. 21 / 58

22 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Constrained A limiting case of the weighted least squares is when a perfect measurement is obtained. The corresponding weight w ii = and can cause difficulties. How then to impose equality constraints in estimation problems? Suppose the observations are partitioned into sub-systems ỹ 1 and ỹ 2 and let ỹ 2 correspond to the perfect measurements. or ỹ 1... ỹ 2 = H 1... H 2 ˆx + e (4) ỹ 1 = H 1ˆx + e 1 (5) ỹ 2 = H 2ˆx (6) Let ỹ 1 R m 1 1, ỹ 2 R m 2 1, ˆx R n 1, n m 2 and n m / 58

23 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Constrained Notice, there is no residual error in equation (6). This has to be satisfied exactly. Thus, the cost funtion to be minimized becomes, J = 1 2 et 1 W 1 e 1 = 1 2 (ỹ 1 H 1ˆx) T W 1 (ỹ 1 H 1ˆx) subject to the equality constraint, ỹ 2 H 2ˆx = 0. We use the method of Lagrange Multipliers to solve this. We begin by augmenting the cost function with a vector of Lagrange Multipliers (λ)as follows, Necessary conditions: J a = 1 2 (ỹ 1 H 1ˆx) T W 1 (ỹ 1 H 1ˆx) + λ T (ỹ 2 H 2ˆx) ˆx J a = (H T 1 W 1 H 1 ) ˆx H T 1 ỹ 1 HT 2 λ = 0 (7) ˆλ Ja = ỹ 2 H 2ˆx = 0 ỹ 2 = H 2ˆx (8) 23 / 58

24 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Constrained Solve for ˆx from equation (7) i.e ˆx = ( H T 1 W 1 H 1 ) 1 H T 1 W 1 ỹ 1 + ( H T 1 W 1 H 1 ) 1 H T 2 λ Now subsitute the result in equation (8), ) ] 1 1 ) ] 1 λ = [H 2 (H T1 W 1 H 1 H T2 [ỹ 2 H 2 (H T1 W 1 H 1 H T1 W 1 ỹ 1 Substitute the value of λ back in equation (7) to finally solve for ˆx Let ) 1 ) ] 1 1 K = (H T 1 W 1 H 1 H T 2 [H 2 (H T 1 W 1 H 1 H T 2 and Thus, x = ( H T 1 W 1 H 1 ) 1 H T 1 W 1 ỹ 1 ˆx = x + K (ỹ 2 H 2 x) Note, if H 2 is a square matrix, K = H 1 2 COMFORTING! 24 / 58

25 Probability Concepts in Linear Sequential Estimation Linear Batch Estimation Linear Sequential Estimation Consider the case when measurements become sequentially available in subsets and estimates are determined immediately upon receipt of such a subset of data. When a new data subset arrives, it is desirable to determine the new estimates based upon all previous measurements including the current subset. Let us examine the case with two subsets for simplicity, ỹ 1 = H 1 x + v 1 (9) ỹ 2 = H 2 x + v 2 (10) The least squares estimate based on the first measurement subset is ( ) 1 ˆx 1 = H T 1 W 1 H 1 H T 1 W 1 ỹ 1 To consider both ỹ 1 and ỹ 2 simultaneously, we merge the two into a single observation equation, ỹ = Hx + v. 25 / 58

26 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Linear Sequential ỹ = Hx + v where, ỹ 1 H 1 v 1 ỹ =..., H =..., ˆv =... ỹ 2 H 2 v 2 Assuming the merged weight matrix is block diagonal so that,. W W = W 2 The optimal estimate based upon the first two measurements is given by, ( 1 ˆx 2 = H T W H) H T W ỹ Since W is block diagonal, ] 1 ) ˆx 2 = [H T 1 W 1 H 1 + H T 2 W 2 H 2 (H T 1 W 1 ỹ 1 + H T 2 W 2 ỹ 2 26 / 58

27 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Linear Sequential [ ] 1 ] 1 Let P 1 H T 1 W 1 H 1 and P2 [H T 1 W 1 H 1 + H T 2 W 2 H 2 ] Notice, P 1 2 = P [H T 2 W 2 H 2 After some manipulations, one can obtain, ˆx 1 = P 1 H T 1 W 1 ỹ 1 ˆx 2 = ˆx 1 + K 2 (ỹ 2 H 2ˆx 1 ).. (ỹ ) ˆx k+1 = ˆx k + K k+1 k+1 H k+1ˆx k with, K k+1 = P k+1 H T k+1w k+1 P 1 k+1 = P 1 k + H T k+1w k+1 H k+1 Thus we modify the previous best correction ˆx k by an additional correction to account for the information contained in the k + 1 measurement subset. This is the Kalman Update Equation for computing the improved estimate ˆx k+1. K k+1 is termed the Kalman Gain Matrix. 27 / 58

28 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear Most real world estimation problems are nonlinear Linear versions of the estimation problem and associated developments apply only to a subset of problems encountered in practice. Most nonlinear estimation problems can be accurately solved by a judiciously chosen successive approximation procedure. Most widely used Successive Approximation Procedure Nonlinear Least Squares - also known as Gaussian Least Squares Differential Correction (Early application by Gauss in 1800s to determine planetary orbits from telescope measurements of the line of sight angles to the planets) 28 / 58

29 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear Assume m observable quantities modelled as y j = f j (x 1, x 2,..., x n); j = 1, 2,..., m; m n where, the f j (x 1, x 2,..., x n) are m arbitrary independent functions of the unknown parameters x i. We require that f (x i ) and at least its first partial derivatives be single-valued, continuous and at least once differentiable. Suppose a set of observed values of the variables y j are available, Measurement Model: ỹ = f (x) + v y j {y 1, y 2,..., y m} 29 / 58

30 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear ỹ = [ỹ 1 ỹ 2... ỹ m] T = measured y values f (x) = [f 1 f 2... f m] T = independent functions x = [x 1 x 2... x n] T = true x values v = [v 1 v 2... v m] T = measurement errors Estimated y-values: ŷ = f (ˆx) e = ỹ ŷ y ŷ = [ŷ 1 ŷ 2... ŷ m] T = estimated y values ˆx = [ˆx 1 ˆx 2... ˆx n] T = estimated x values e = [e 1 e 2... e m] T = residual errors 30 / 58

31 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear Rewrite Measurement Model: ỹ = f (ˆx) + e As before, we seek an estimate (ˆx) for x that minimizes J = 1 2 et W e = 1 2 [ỹ f (ˆx)]T W [ỹ f (ˆx)] W is an m m weighting matrix used to weight the relative importance of each measurement. Gauss Procedure: Assume current estimates of the unknown x-values are available x c = [x 1c x 2c... x nc] T 31 / 58

32 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear Whatever the unknown objective x-values, ˆx are, we assume they are related to the respective current estimates, by an also unknown set of corrections x. ˆx = x c + x Linearize f (ˆx) about x c. f (ˆx) f (x c) + H x where, H = f x xc 32 / 58

33 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear The Gradient Matrix H is known as the Jacobian matrix. The measurement residual after the correction can now be linearly approximated as y ỹ f (ˆx) ỹ f (x c) H x = y c H x where, the residual before the correction is y c ỹ f (x c) Objective: Minimize weighted sum squares J Strategy: To determine approximate corrections in x, select particular corrections that lead to minimum sum of squares of the linearly predicted residuals, J p: J = 1 2 y T W y J p 1 2 ( y c H x)t W ( y c H x) 33 / 58

34 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear Following the minimization procedure as before, one obtains x = (H T WH) 1 H T W y c An initial guess x c is essential to begin the procedure. A stopping condition with an accuracy dependent tolerance is given by δj J i J i 1 J i < ε W where, ε is some prescribed small value. If the condition is not satisfied then the update procedure is iterated with the new estimate. 34 / 58

35 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear This procedure of differential correction has been widely used on a lot of problems. Convergence difficulties usually stem from any of the following sources. initial x-estimate if too far from the minimizing ˆx (for the nonlinearity of particular application) numerical difficulties in solving for corrections - arithmetic errors corrupting the algorithm or H matrix having fewer than n linearly independent rows or columns (i.e. rank deficient) The initial estimate convergence difficulty can be overcome by using Levenberg-Marquardt algorithm that combines the least squares differential correction process with a gradient search. 35 / 58

36 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear - Algorithm 36 / 58

37 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear - Example Consider the earlier example with the model, [ y k+1 = e a t] [ b ( y k + e a t 1) ] u k a Suppose we wish to determine, a and b directly. Clearly the parameters appear nonlinearly in the above equation so we use the nonlinear least squares algorithm. f k a f k b x = [a b] T ỹ = [ỹ 2 ỹ 3 ỹ 101 ] T [ f k = e a t] [ b ( y k + e a t 1) ] u k a The appropriate partials are given by, [ = t e a t] y k + = 1 a ( ) e a t 1 u k [ b a 2 ( 1 e a t) + b a tea t ] u k 37 / 58

38 Probability Concepts in Linear Batch Estimation Linear Sequential Estimation Nonlinear - Example Then the H matrix is given by, t [ e a t] [ b ( y 1 + ) 1 e a t + b ] a 2 a tea t t [ e a t] [ b ( y 2 + ) 1 e a t + b H = a 2 a tea t. t [ e a t] [ b ( y ) 1 e a t + b a 2 a tea t 1 ( u 1 e a t 1 ) u 1 a ( e a t 1 ) u 2 ] u 2 1 a ] u a. ( e a t 1 ) u 100 Using a starting guess x = [5 5] T and a stopping criterion of ɛ = , the NLS algorithm converges in 5 to 6 iterations, with â = and ˆb = Converting these to the discrete equivalents, ˆΦ = and ˆΓ = The example also illustrates why model choice is important in least squares estimation algorithms. 38 / 58

39 Probability Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Consider a single throw of a true die. The probability of occurrence of each of the events, 1, 2, 3, 4, 5 or 6 is exactly the same on a given throw. For a loaded or biased die, the probability of occurrence of certain events would be greater than others. If a given discrete-values experiment is conducted N times and N j is the number of times that the j th event x(j) occurred, then intuitively, the probability of occurrence of x(j) can be written as, p(x(j)) = lim N N j N For example, for the throw of a single die the probability of obtaining a value of 3 is given by p(3) = 1/6. A discrete-valued random variable, x, is defined as a function having finite number of possible values, x(j); with the associated probability of x(j) occuring being denoted by p(x(j)). More compactly, henceforth, x(j) and p(x(j)) would be referred to as x and p(x). (Attn: context is important!!) 39 / 58

40 Probability Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Expanding this for the case of a single throw of two dice. We now have 36 outcomes. Clearly the probability of the sum of the two dice being 7 is the highest. When n 2, it is difficult to produce all combinations. A generating function is helpful, f (x) = (x + x 2 + x 3 + x 4 + x 5 + x 6) n The coefficients of the powers of x are used to form the count. The probability is then the obtained by dividing the count by 6 n. Consider another experiment involving 4 flips of a coin. The number of ways, heads appears for the total of 16 (= 2 4 ) outcomes is calculated as follows. The number of ways to obtain x heads in n flips is the number of combinations of n things taken x at a time. Number of ways n C x = n! x!(n x)! For example if n = 4 and x = 2, the number of ways is equal to 6. The probability is 6 16 = / 58

41 Probability Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation A compound event is defined as the occurrence of either x(j) or x(k). The probability of such an event is computed as p(x(j) x(k)) = p(x(j)) + p(x(k)) p(x(j) x(k)) Note, x(j) x(k) denotes x(j) or x(k) and x(j) x(k) denotes x(j) and x(k) The probability of obtaining one event and another event is known as the joint probability. If x(j) and x(k) are mutually exclusive, then p(x(j) x(k)) = 0. For the discrete valued events, p(x(j)) is also termed the probability mass function. 0 p(x(j)) 1 p(x(j)) = 1 If events x(j) and x(k) are independent, then we have p(x(j) x(k)) = p(x(j))p(x(k)). j The conditional probability of x(j) given x(k) is denoted as p(x(j) x(k)). 41 / 58

42 Probability Concepts in Conditional Probability Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Suppose event x(k) has occurred. Then x(j) occurs if and only if x(j) and x(k)) occur. Thus, the probability of x(j), given that we know x(k) occurred should intuitively depend on p(x(j) x(k)). Similarly, p(x(j) x(k)) = p(x(k) x(j)) = p(x(j) x(k)) p(x(k)) p(x(j) x(k)) p(x(j)) Combining the two above gives us the Bayes rule. p(x(j) x(k))p(x(k)) = p(x(k) x(j))p(x(j)) p(x(j) x(k)) = p(x(k) x(j))p(x(j)) p(x(k)) This rule can be used to show some interesting counterintuitive results. For example, say 1 out of 1000 people have a rare disease. Tests show that 99% are positive when they have a disease and 2% are positive when they dont. What s the probability that someone actually has a disease when the test if positive? (0.047) 42 / 58

43 Probability Concepts in Random variables and statistics Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation The random variable x is usually described in terms of its moments. The first two moments of x are given by the mean µ of x: µ j x(j)p(x(j)) and the variance (σ 2 ) of x: σ 2 j (x(j) µ) 2 p(x(j)) σ is the standard deviation of x. The expected value of a function f (x) of a discrete random variable x is defined as, E{f (x)} = f (x(j))p(x(j)) j Clearly, the mean and variance are the expected values of the functions, f (x) = x and f (x) = (x µ) 2 respectively. The expectation operator E{ } is a linear operator, i.e., E{af (x) + bg(x)} = ae{f (x)} + be{g(x)} 43 / 58

44 Probability Concepts in Gaussian Random Processes Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation The most widely used distribution for state estimation involes the Gaussian random process. The Gaussian or normal density function for x: p(x) = 1 ] [ σ 2π exp (x µ)2 2σ 2 with mean given by µ and variance given by σ 2. For a multi-dimensional case for a vector x with an associated probability density [ function having mean µ 1 and covariance R, p(x) = exp 1 ] 1/2 [det(2πr)] 2 (x µ)t R 1 (x µ) The mean and standard deviation suffice to define the distribution. Thus, a simple notation for a Gaussian random variable is x N (µ, R) A stochastic process is simply a collection of random vectors defined on the same probability space. A zero-mean Gaussian white-noise process has the following properties, E{x} = 0 E{x(τ)x T (τ )} = Rδ(τ τ) where δ(τ τ) is the delta function. 44 / 58

45 Probability Concepts in Gaussian Random Processes Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation A process is supposed to be stationary if its random variable statistics do not vary in time, i.e., the probability statistics at time τ have the same mean and covariance as the probability statistics at time τ 45 / 58

46 Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Minimum Variance Estimation (a priori state estimates absent) For a measurable quantity, x, assume the following linear observation model, ỹ = Hx + v where, v N (0, R). We desire to estimate x as a linear combination of the measurements, ỹ as, ˆx = Mỹ + n We seek an optimum choice of M and n. The minimum variance definition of optimum M and n is that the variance of all estimates, ˆx i, from their respective true values is minimized. J i = 1 2 E{(ˆx i x i ) 2 } i = 1, 2,..., n From the first two equations and assuming perfect measurements (no measurement error, i.e. ỹ = Hx), we note that ˆx = MHx + n Thus for perfect estimates, MH = I and n = / 58

47 Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Minimum Variance Estimation (a priori state estimates absent) These are the constraints determined for the ideal estimation process, MH = I and n = 0. Then the desired estima tor is simply, ˆx = Mỹ The estimator is derived as follows. The error covariance matrix of the unbiased estimator is, P = E{(ˆx x) (ˆx x) T }. We wish to determine M that minimizes the above subject to the constraint that MH = I. We form the generalized loss function to be minimized: J = 1 ] [E{(ˆx 2 Tr x) (ˆx x) T } + Tr [Λ (I MH)] Λ is a matrix of Lagrange Multipliers. Solving the above minimization we obtain, ( ) 1 M = H T R 1 H H T R 1 Thus the desired estimator is: ( 1 ˆx = H T R H) 1 H T R 1 ỹ 47 / 58

48 Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Minimum Variance Estimation (a priori state estimates absent) ˆx = ( H T R 1 H) 1 H T R 1 ỹ The above is also referred to as the Gauss-Markov Theorem. The minimum variance estimator is identical to the least squares estimator provided that the weight matrix is identified as the inverse of the observation error covariance. Note, the sequential least squares solution takes on a similar form if R 1 has a block structure. 48 / 58

49 Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Minimum Variance Estimation (with a priori state estimates) We will now extend the above to allow a rigorous incorporation of a priori estimates, ˆx a. Assume the following linear observation model, ỹ = Hx + v where, v N (0, R). The covariance of R = E{vv T } Suppose x is unknown (a random variable). The a priori state estimates are given as ˆx a = x + w with associated (assumed known) a priori error covariance matrix, cov{w} Q = E{ww T }. Assumption: E{wv T } = 0 i.e., the measurement errors and the a priori estimate errors are uncorrelated. We desire to estimate, x as a linear combination of the measurements ỹ and the a priori estimates ˆx a, i.e., ˆx = Mỹ + N ˆx a + n As before we derive the constraints for perfect state estimates, MH + N = I and n = 0. Thus, ˆx = Mỹ + N ˆx a 49 / 58

50 Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Minimum Variance Estimation (with a priori state estimates) We form the generalized loss function to be minimized: J = 1 ] [E{(ˆx 2 Tr x) (ˆx x) T } + Tr [Λ (I MH N)] Λ is a matrix of Lagrange Multipliers. Solving the above minimization we obtain, ( M = H T R 1 H + Q 1) 1 H T R 1 and Thus N = ( H T R 1 H + Q 1) 1 Q 1 ˆx = (H T R 1 H + Q 1) 1 (H T R 1 ỹ + Q 1ˆx ) a 50 / 58

51 Probability Concepts in Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Minimum Variance Estimation (with a priori state estimates) Consider the limiting cases, ˆx = (H T R 1 H + Q 1) 1 (H T R 1 ỹ + Q 1ˆx ) a A priori knowledge is very poor, Q, Q 1 0 and R is finite. The estimator reduces to the standard minimum variance estimator. Lousy measurements, R, R 1 0 and Q is finite then, ˆx = ˆx a. Notice the parallels with sequential least squares estimation. 51 / 58

52 Probability Concepts in Maximum Likelihood Estimation Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation This method was introduced by R. A. Fisher, a geneticist and statistician in the 1920s. Maximum Likelihood Estimation yields estimates for the unknown quantities which maximize the probability (likelihood) of obtaining the observed set of data. Despite the fundamental differences with the minimum variance estimator, both approaches give the exact same results for least squares estimates under the assumption of zero-mean Gaussian noise measurement-error process. Assume the following linear observation model with deterministic x, ỹ = Hx + v v is a zero-mean Gaussian error process with covariance R. The mean (µ) of the measurements (ỹ), and the covariance, µ = E{Hx} + E{v} µ = Hx cov{ỹ} = E{(ỹ µ) (ỹ µ) T } = R 52 / 58

53 Probability Concepts in Maximum Likelihood Estimation Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Thus ỹ N (µ, R). We now use the multidimensional or the multivariate normal distribution for the likelihood (probability) function. [ 1 L(ỹ; x) = exp 1 ] (2π) m/2 1/2 [det(r)] 2 (ỹ Hx)T R 1 (ỹ Hx) The log likelihood function is given by, ln L(ỹ; x) = m 2 ln(2π) 1 2 ln [det(r)] 1 2 (ỹ Hx)T R 1 (ỹ Hx) For the optimization, we can ignore the first two terms on the right hand side as they dont depend on x. Note, maximizing negative of the above log likelihood to obtain ˆx is equivalent to minimizing, J(ˆx) = 1 2 (ỹ Hx)T R 1 (ỹ Hx) Thus, ˆx = estimate) ( H T R 1 H) 1 H T R 1 ỹ. (same as the minimum variance 53 / 58

54 Probability Concepts in Cramer Rao Inequality Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation This inequality can be used to give a lower bound on the expected errors between the estimated quantities and the true values from the known properties of the measurement errors. Let f (ỹ; x) be the probability density function of the sample ỹ. The Cramer-Rao inequality for an unbiased estimate ˆx is given by, P E{(ˆx x) (ˆx x) T } F 1 where, the Fisher Information Matrix is given by, { [ ] [ ] } T F = E ln f (ỹ; x) ln f (ỹ; x) x x The matrix can also be computed using the Hessian matrix, given by { } 2 F = E ln f (ỹ; x) x x T 54 / 58

55 Probability Concepts in Cramer Rao Inequality Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation For the Gauss-Markov theorem and the log likelihood function, i.e. Jx), we compute ( ) F = 2 x x J(x) = H T R 1 H T Hence the Cramer-Rao inequality is given by, Also, ˆx x = P ( ) 1 H T R 1 H ( H T R 1 H) 1 H T R 1 v Using E{vv T } = R gives us the error covariance, P = ( H T R 1 H) 1. The equality is satisfied. So, the least squares estimate from the Gauss-Markov Theorem is the most efficient possible estimate. 55 / 58

56 Probability Concepts in Bayesian Estimation Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Up until now we assumed the parameters to be estimated to be unknown constants. In Bayesian estimation, we consider the parameters are random variables with some a priori distribution. Bayesian estimation combines this a priori information with the measurements through a conditional density function of x given the measurements ỹ. This conditional function is known as the a posteriori distribution of x. Thus, Bayesian Estimation requires the probability density functions of both the measurement noise and the unknown parameters. From Bayes rule, f (x ỹ) = f (ỹ x) f (x) f (ỹ) (11) Since, ỹ is treated known, f (ỹ) is just a normalization factor to ensure that f (x ỹ) is a probability density function. Thus, f (ỹ) = f (ỹ x) f (x) dx. If this integral doesnt exist, then we let f (ỹ) = 1 so that f (x ỹ) is proper. 56 / 58

57 Probability Concepts in Bayesian Estimation Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Maximum a posteriori (MAP) estimation finds an estimate for x that maximizes equation (11). Since f (ỹ) doesnt depend on ˆx explicitly, this is equivalent to maximizing f (ỹ ˆx) f (ˆx). We again use the natural logrithm. J MAP (ˆx) = ln [f (ỹ; ˆx)] + ln [f (ˆx)] The first term in the sum is actually the log-likelihood function and the second term gives the a priori information on the to-be-determined parameters. Thus the MAP estimator maximizes, J MAP (ˆx) = ln [L((ỹ; ˆx)] + ln [f (ˆx)] Note, if the a priori distribution f (ˆx) is uniform, the MAP estimation is equivalent to the maximum likelihood estimation. 57 / 58

58 Probability Concepts in Bayesian Estimation Minimum variance estimation without a priori state estimates Minimum variance estimation with a priori state estimates Maximum Likelihood Estimation Cramer Rao Inequality Bayesian Estimation Let, [ 1 L(ỹ; ˆx) = exp 1 ] (2π) m/2 1/2 [det(r)] 2 (ỹ H ˆx)T R 1 (ỹ H ˆx) [ 1 f (ˆx) = exp 1 ] (2π) n/2 1/2 [det(q)] 2 (ˆx a ˆx)T Q 1 (ˆx a ˆx) Applying the maximization as per MAP estimation gives us the following estimator, ˆx = (H T R 1 H + Q 1) 1 (H T R 1 ỹ + Q 1ˆx ) a which is the same result obtained through the minimum variance. The solution through MAP estimation is much simpler though. Another approach for Bayesian Estimation is a Minimum Risk (MR) Estimator. Some practical difficulties such as convergence to the ML estimates for uniform a priori distributions is not guaranteed. For the Gaussian cases, this is identical to the MAP estimator. 58 / 58

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

MAT 610: Numerical Linear Algebra. James V. Lambers

MAT 610: Numerical Linear Algebra. James V. Lambers MAT 610: Numerical Linear Algebra James V Lambers January 16, 2017 2 Contents 1 Matrix Multiplication Problems 7 11 Introduction 7 111 Systems of Linear Equations 7 112 The Eigenvalue Problem 8 12 Basic

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

Jim Lambers MAT 610 Summer Session Lecture 1 Notes

Jim Lambers MAT 610 Summer Session Lecture 1 Notes Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra

More information

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes

More information

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization

More information

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations Linear Algebra in Computer Vision CSED441:Introduction to Computer Vision (2017F Lecture2: Basic Linear Algebra & Probability Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Mathematics in vector space Linear

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT Math Camp II Basic Linear Algebra Yiqing Xu MIT Aug 26, 2014 1 Solving Systems of Linear Equations 2 Vectors and Vector Spaces 3 Matrices 4 Least Squares Systems of Linear Equations Definition A linear

More information

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search

More information

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017 Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Basic Concepts in Matrix Algebra

Basic Concepts in Matrix Algebra Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz

Introduction to Mobile Robotics Compact Course on Linear Algebra. Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Introduction to Mobile Robotics Compact Course on Linear Algebra Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Vectors Arrays of numbers Vectors represent a point in a n dimensional space

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

3 (Maths) Linear Algebra

3 (Maths) Linear Algebra 3 (Maths) Linear Algebra References: Simon and Blume, chapters 6 to 11, 16 and 23; Pemberton and Rau, chapters 11 to 13 and 25; Sundaram, sections 1.3 and 1.5. The methods and concepts of linear algebra

More information

Matrices A brief introduction

Matrices A brief introduction Matrices A brief introduction Basilio Bona DAUIN Politecnico di Torino Semester 1, 2014-15 B. Bona (DAUIN) Matrices Semester 1, 2014-15 1 / 44 Definitions Definition A matrix is a set of N real or complex

More information

4 Derivations of the Discrete-Time Kalman Filter

4 Derivations of the Discrete-Time Kalman Filter Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

Linear Algebra Formulas. Ben Lee

Linear Algebra Formulas. Ben Lee Linear Algebra Formulas Ben Lee January 27, 2016 Definitions and Terms Diagonal: Diagonal of matrix A is a collection of entries A ij where i = j. Diagonal Matrix: A matrix (usually square), where entries

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

Linear Algebra Primer

Linear Algebra Primer Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................

More information

Stat 206: Linear algebra

Stat 206: Linear algebra Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe

More information

A Little Necessary Matrix Algebra for Doctoral Studies in Business & Economics. Matrix Algebra

A Little Necessary Matrix Algebra for Doctoral Studies in Business & Economics. Matrix Algebra A Little Necessary Matrix Algebra for Doctoral Studies in Business & Economics James J. Cochran Department of Marketing & Analysis Louisiana Tech University Jcochran@cab.latech.edu Matrix Algebra Matrix

More information

Regression. Oscar García

Regression. Oscar García Regression Oscar García Regression methods are fundamental in Forest Mensuration For a more concise and general presentation, we shall first review some matrix concepts 1 Matrices An order n m matrix is

More information

ECE 275A Homework 6 Solutions

ECE 275A Homework 6 Solutions ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =

More information

MIT Final Exam Solutions, Spring 2017

MIT Final Exam Solutions, Spring 2017 MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

Introduction to Matrices

Introduction to Matrices POLS 704 Introduction to Matrices Introduction to Matrices. The Cast of Characters A matrix is a rectangular array (i.e., a table) of numbers. For example, 2 3 X 4 5 6 (4 3) 7 8 9 0 0 0 Thismatrix,with4rowsand3columns,isoforder

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.) page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation

More information

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations. POLI 7 - Mathematical and Statistical Foundations Prof S Saiegh Fall Lecture Notes - Class 4 October 4, Linear Algebra The analysis of many models in the social sciences reduces to the study of systems

More information

Exercise Sheet 1.

Exercise Sheet 1. Exercise Sheet 1 You can download my lecture and exercise sheets at the address http://sami.hust.edu.vn/giang-vien/?name=huynt 1) Let A, B be sets. What does the statement "A is not a subset of B " mean?

More information

7. Dimension and Structure.

7. Dimension and Structure. 7. Dimension and Structure 7.1. Basis and Dimension Bases for Subspaces Example 2 The standard unit vectors e 1, e 2,, e n are linearly independent, for if we write (2) in component form, then we obtain

More information

A Review of Linear Algebra

A Review of Linear Algebra A Review of Linear Algebra Gerald Recktenwald Portland State University Mechanical Engineering Department gerry@me.pdx.edu These slides are a supplement to the book Numerical Methods with Matlab: Implementations

More information

Foundations of Computer Vision

Foundations of Computer Vision Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Lecture 4: Types of errors. Bayesian regression models. Logistic regression Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

SUMMARY OF MATH 1600

SUMMARY OF MATH 1600 SUMMARY OF MATH 1600 Note: The following list is intended as a study guide for the final exam. It is a continuation of the study guide for the midterm. It does not claim to be a comprehensive list. You

More information

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra Massachusetts Institute of Technology Department of Economics 14.381 Statistics Guido Kuersteiner Lecture Notes on Matrix Algebra These lecture notes summarize some basic results on matrix algebra used

More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2 Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch

More information

Basic Elements of Linear Algebra

Basic Elements of Linear Algebra A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Computational Methods. Least Squares Approximation/Optimization

Computational Methods. Least Squares Approximation/Optimization Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Linear Algebra, Summer 2011, pt. 2

Linear Algebra, Summer 2011, pt. 2 Linear Algebra, Summer 2, pt. 2 June 8, 2 Contents Inverses. 2 Vector Spaces. 3 2. Examples of vector spaces..................... 3 2.2 The column space......................... 6 2.3 The null space...........................

More information

Linear Algebra and Matrices

Linear Algebra and Matrices Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Math 302 Outcome Statements Winter 2013

Math 302 Outcome Statements Winter 2013 Math 302 Outcome Statements Winter 2013 1 Rectangular Space Coordinates; Vectors in the Three-Dimensional Space (a) Cartesian coordinates of a point (b) sphere (c) symmetry about a point, a line, and a

More information

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Jim Lambers MAT 610 Summer Session Lecture 2 Notes Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the

More information

Mathematical Foundations of Applied Statistics: Matrix Algebra

Mathematical Foundations of Applied Statistics: Matrix Algebra Mathematical Foundations of Applied Statistics: Matrix Algebra Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/105 Literature Seber, G.

More information

ECEN 615 Methods of Electric Power Systems Analysis Lecture 18: Least Squares, State Estimation

ECEN 615 Methods of Electric Power Systems Analysis Lecture 18: Least Squares, State Estimation ECEN 615 Methods of Electric Power Systems Analysis Lecture 18: Least Squares, State Estimation Prof. om Overbye Dept. of Electrical and Computer Engineering exas A&M University overbye@tamu.edu Announcements

More information

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review 10.34 Numerical Methods Applied to Chemical Engineering Fall 2015 Quiz #1 Review Study guide based on notes developed by J.A. Paulson, modified by K. Severson Linear Algebra We ve covered three major topics

More information

Linear Algebra Highlights

Linear Algebra Highlights Linear Algebra Highlights Chapter 1 A linear equation in n variables is of the form a 1 x 1 + a 2 x 2 + + a n x n. We can have m equations in n variables, a system of linear equations, which we want to

More information

1 Matrices and vector spaces

1 Matrices and vector spaces Matrices and vector spaces. Which of the following statements about linear vector spaces are true? Where a statement is false, give a counter-example to demonstrate this. (a) Non-singular N N matrices

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

ME751 Advanced Computational Multibody Dynamics

ME751 Advanced Computational Multibody Dynamics ME751 Advanced Computational Multibody Dynamics Review: Elements of Linear Algebra & Calculus September 9, 2016 Dan Negrut University of Wisconsin-Madison Quote of the day If you can't convince them, confuse

More information

Lecture 15 Review of Matrix Theory III. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore

Lecture 15 Review of Matrix Theory III. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore Lecture 15 Review of Matrix Theory III Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore Matrix An m n matrix is a rectangular or square array of

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna

More information

Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices

Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices Lecture 3: Simple Linear Regression in Matrix Format To move beyond simple regression we need to use matrix algebra We ll start by re-expressing simple linear regression in matrix form Linear algebra is

More information

Lecture Summaries for Linear Algebra M51A

Lecture Summaries for Linear Algebra M51A These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture

More information

Linear Algebra for Machine Learning. Sargur N. Srihari

Linear Algebra for Machine Learning. Sargur N. Srihari Linear Algebra for Machine Learning Sargur N. srihari@cedar.buffalo.edu 1 Overview Linear Algebra is based on continuous math rather than discrete math Computer scientists have little experience with it

More information

RECURSIVE ESTIMATION AND KALMAN FILTERING

RECURSIVE ESTIMATION AND KALMAN FILTERING Chapter 3 RECURSIVE ESTIMATION AND KALMAN FILTERING 3. The Discrete Time Kalman Filter Consider the following estimation problem. Given the stochastic system with x k+ = Ax k + Gw k (3.) y k = Cx k + Hv

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Review Packet 1 B 11 B 12 B 13 B = B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43

Review Packet 1 B 11 B 12 B 13 B = B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43 Review Packet. For each of the following, write the vector or matrix that is specified: a. e 3 R 4 b. D = diag{, 3, } c. e R 3 d. I. For each of the following matrices and vectors, give their dimension.

More information

Matrices and systems of linear equations

Matrices and systems of linear equations Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.

More information

Orthonormal Transformations and Least Squares

Orthonormal Transformations and Least Squares Orthonormal Transformations and Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 30, 2009 Applications of Qx with Q T Q = I 1. solving

More information

[POLS 8500] Review of Linear Algebra, Probability and Information Theory

[POLS 8500] Review of Linear Algebra, Probability and Information Theory [POLS 8500] Review of Linear Algebra, Probability and Information Theory Professor Jason Anastasopoulos ljanastas@uga.edu January 12, 2017 For today... Basic linear algebra. Basic probability. Programming

More information

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Evan Kwiatkowski, Jan Mandel University of Colorado Denver December 11, 2014 OUTLINE 2 Data Assimilation Bayesian Estimation

More information

1 Matrices and Systems of Linear Equations. a 1n a 2n

1 Matrices and Systems of Linear Equations. a 1n a 2n March 31, 2013 16-1 16. Systems of Linear Equations 1 Matrices and Systems of Linear Equations An m n matrix is an array A = (a ij ) of the form a 11 a 21 a m1 a 1n a 2n... a mn where each a ij is a real

More information

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0. Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix Definition: Let L : V 1 V 2 be a linear operator. The null space N (L) of L is the subspace of V 1 defined by N (L) = {x

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

Lecture 7. Econ August 18

Lecture 7. Econ August 18 Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar

More information

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015 Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See

More information