4 Derivations of the Discrete-Time Kalman Filter

Size: px
Start display at page:

Download "4 Derivations of the Discrete-Time Kalman Filter"

Transcription

1 Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time Kalman Filter We derive here the basic equations of the Kalman filter (KF), for discrete-time linear systems We consider several derivations under different assumptions and viewpoints: For the Gaussian case, the KF is the optimal (MMSE) state estimator In the non-gaussian case, the KF is derived as the best linear (LMMSE) state estimator We also provide a deterministic (least-squares) interpretation We start by describing the basic state-space model 1

2 41 The Stochastic State-Space Model A discrete-time, linear, time-varying state space system is given by: x k+1 = F k x k + G k w k (state evolution equation) z k = H k x k + v k (measurement equation) for k 0 (say), and initial conditions x 0 Here: F k, G k, H k are known matrices x k IR n is the state vector w k IR n w is the state noise z k IR m is the observation vector v k the observation noise The initial conditions are given by x 0, usually a random variable The noise sequences (w k, v k ) and the initial conditions x 0 are stochastic processes with known statistics The Markovian model Recall that a stochastic process {X k } is a Markov process if p(x k+1 X k, X k 1, ) = p(x k+1 X k ) For the state x k to be Markovian, we need the following assumption Assumption A1: The state-noise process {w k } is white in the strict sense, namely all w k s are independent of each other Furthermore, this process is independent of x 0 The following is then a simple exercise: Proposition: Under A1, the state process {x k, k 0} is Markov 2

3 Note: Linearity is not essential: The Marko property follows from A1 also for the nonlinear state equation x k+1 = f(x k, w k ) The measurement process z k is usually not Markov The pdf of the state can (in principle) be computed recursively via the following (Chapman-Kolmogorov) equation: p(x k+1 ) = p(x k+1 x k )p(x k )dx k where p(x k+1 x k ) is determined by p(w k ) The Gaussian model Assume that the noise sequences {w k }, {v k } and the initial conditions x 0 are jointly Gaussian It easily follows that the processes {x k } and {z k } are (jointly) Gaussian as well If, in addition, A1 is satisfied (namely {w k } is white and independent of x 0 ), then x k is a Markov process This model is often called the Gauss-Markov Model 3

4 Second-Order Model We often assume that only the first and second order statistics of the noise is known Consider our linear system: x k+1 = F k x k + G k w k, k 0 z k = H k x x + v k, under the following assumptions: w k a 0-mean white noise: E(w k ) = 0, cov(w k, w l ) = Q k δ kl v k a 0-mean white noise: E(v k ) = 0, cov(v k, v l ) = R k δ kl cov(w k, v l ) = 0: uncorrelated noise x 0 is uncorrelated with the other noise sequences denote x 0 = E(x 0 ), cov(x 0 ) = P 0 We refer to this model as the standard second-order model It is sometimes useful to allow correlation between v k and w k : cov(w k, v l ) E(w k vl T ) = S k δ kl This gives the second-order model with correlated noise A short-hand notation for the above correlations: w k w l Q k δ kl S k δ kl 0 cov( v k, v l ) = S k T δ kl R k δ kl P 0 x 0 x 0 Note that the Gauss-Markov model is a special case of this model 4

5 Mean and covariance propagation For the standard second-order model, we easily obtain recursive formulas for the mean and covariance of the state The mean obviously satisfies: x k+1 = F k x k + G k w k = F k x k Consider next the covariance: P k = E((xk x k )(x k x) T ) Note that x k+1 x k+1 = F k (x k x k ) + G k w k, and w k and x k are uncorrelated (why?) Therefore P k+1 = F k P k Fk T + G k Q k G T k This equation is in the form of a Lyapunov difference equation Since z k = H k x x + v k, it is now easy to compute its covariance, and also the joint covariances of (x k, z k ) In the Gaussian case, the pdf of x k is completely specified by the mean and covariance: x k N(x k, P k ) 5

6 42 The KF for the Gaussian Case Consider the linear Gaussian (or Gauss-Markov) model x k+1 = F k x k + G k w k, k 0 z k = H k x x + v k where: {w k } and {v k } are independent, zero-mean Gaussian white processes with covariances E(v k v T l ) = R k δ kl, E(w k w T l ) = Q k δ kl The initial state x 0 is a Gaussian RV, independent of the noise processes, with x 0 N(x 0, P 0 ) Let Z k = (z 0,, z k ) Our goal is to compute recursively the following optimal (MMSE) estimator of x k : ˆx + k ˆx k k = E(x k Z k ) Also define the one-step predictor of x k : ˆx k ˆx k k 1 = E(x k Z k 1 ) and the respective covariance matrices: P + k P k P k k P k k 1 = E{x k ˆx + k )(x k ˆx + k )T Z k } = E{x k ˆx k )(x k ˆx k )T Z k 1 } Note that P + k (and similarly P k ) can be viewed in two ways: (i) It is the covariance matrix of the (posterior) estimation error, e k = x k ˆx + k In particular, MMSE = trace(p + k ) 6

7 (ii) It is the covariance matrix of the conditional RV (x k Z k ), namely an RV with distribution p(x k Z k ) (since ˆx + k is its mean) Finally, denote P 0 = P 0, ˆx 0 = x 0 Recall the formulas for conditioned Gaussian vectors: If x and z are jointly Gaussian, then p x z N(m, Σ), with m = m x + Σ xz Σ 1 zz (z m z ), Σ = Σ xx Σ xz Σ 1 zz Σ zx The same formulas hold when everything is conditioned, in addition, on another random vector According to the terminology above, we say in this case that the conditional RV (x z) is Gaussian Proposition: For the model above, all random processes (noises, x k, z k ) are jointly Gaussian Proof: All can be expressed as linear combinations of the noise seqeunces, which are jointly Gaussian (why?) It follows that (x k Z m ) is Gaussian (for any k, m) In particular: (x k Z k ) N(ˆx + k, P + k ), (x k Z k 1 ) N(ˆx k, P k ) 7

8 Filter Derivation Suppose, at time k, that (ˆx k, P k ) is given We shall compute (ˆx + k, P + k ) and (ˆx k+1, P k+1 ), using the following two steps Measurement update step: Since z k = H k x k + v k, then the conditional vector ( ) xk ( Z k 1 ) is Gaussian, with mean and covariance: z k ˆx k H kˆx k, P k P k HT k H k P k M k where M k = Hk P k HT k + R k To compute (x k Z k ) = (x k z k, Z k 1 ), we apply the above formula for conditional expectation of Gaussian RVs, with everything pre-conditioned on Z k 1 It follows that (x k Z k ) is Gaussian, with mean and covariance: ˆx + k = E(x k Z k ) = ˆx k + P k HT k (M k ) 1 (z k H kˆx k ) P + k = cov(x k Z k ) = P k P k HT k (M k ) 1 H k P k Time update step Recall that x k+1 = F k x k + G k w k Further, x k and w k are independent given Z k (why?) Therefore, P k+1 ˆx k+1 = E(x k+1 Z k ) = F kˆx + k = cov(x k+1 Z k ) = F k P + k F T k + G k Q k G T k 8

9 Remarks: 1 The KF computes both the estimate ˆx + k and its MSE/covariance P + k (and similarly for ˆx k ) Note that the covariance computation is needed as part of the estimator computation However, it is also of independent importance as is assigns a measure of the uncertainly (or confidence) to the estimate 2 It is remarkable that the conditional covariance matrices P + k and P k do not depend on the measurements {z k } They can therefore be computed in advance, given the system matrices and the noise covariances 3 As usual in the Gaussian case, P + k is also the unconditional error covariance: P + k = cov(x k ˆx + k ) = E[(x k ˆx + k )(x k ˆx + k )T ] In the non-gaussian case, the unconditional covariance will play the central role as we compute the LMMSE estimator 4 Suppose we need to estimate some s k = Cxk Then the optimal estimate is ŝ k = E(s k Z k ) = C ˆx + k 5 The following output prediction error z k = zk H kˆx k z k E(z k Z k 1 ) is called the innovation, and { z k } is the important innovations process Note that M k = H k P k HT k + R k is just the covariance of z k 9

10 43 Best Linear Estimator Innovations Approach a Linear Estimators Recall that the best linear (or LMMSE) estimator of x given y is an estimator of the form ˆx = Ay + b, which minimizes the mean square error E( x ˆx 2 ) It is given by: ˆx = m x + Σ xy Σ 1 yy (y m y ) where Σ xy and Σ yy are the covariance matrices It easily follows that ˆx is unbiased: E(ˆx) = m x, and the corresponding (minimal) error covariance is cov(x ˆx) = E(x ˆx)(x ˆx) T = Σ xx Σ xy Σ 1 yy Σ T xy We shall find it convenient to denote this estimator ˆx as E L (x y) Note that this is not the standard conditional expectation Recall further the orthogonality principle: E((x E L (x y))l(y)) = 0 for any linear function L(y) of y The following property will be most useful It follows simply by using y = (y 1 ; y 2 ) in the formulas above: Suppose cov(y 1, y 2 ) = 0 Then E L (x y 1, y 2 ) = E L (x y 1 ) + [E L (x y 2 ) E(x)] Furthermore, cov(x E L (x y 1, y 2 )) = (Σ xx Σ xy1 Σ 1 y 1 y 1 Σ T xy 1 ) Σ xy2 Σ 1 y 2 y 2 Σ T xy 2 10

11 b The innovations process Consider a discrete-time stochastic process {z k } k 0 The (wide-sense) innovations process is defined as z k = z k E L (z k Z k 1 ), where Z k 1 = (z 0 ; z k 1 ) The innovation RV z k may be regarded as containing only the new statistical information which is not already in Z k 1 The following properties follow directly from those of the best linear estimator: (1) E( z k ) = 0, and E( z k Z T k 1 ) = 0 (2) z k is a linear function of Z k (3) Thus, cov( z k, z l ) = E( z k z T l ) = 0 for k l This implies that the innovations process is a zero-mean white noise process Denote Z k = ( z 0 ; ; z k ) It is easily verified that Z k and Z k are linear functions of each other This implies that E L (x Z k ) = E L (x Z k ) for any RV x It follows that (taking E(x) = 0 for simplicity): E L (x Z k ) = E L (x Z k ) = E L (x Z k 1 ) + E L (x z k ) = k E L (x z l ) l=0 11

12 c Derivation of the KF equations We proceed to derive the Kalman filter as the best linear estimator for our linear, non-gaussian model We slightly generalize the model that was treated so far by allowing correlation between the state noise and measurement noise Thus, we consider the model x k+1 = F k x k + G k w k, k 0 z k = H k x x + v k, with [w k ; v k ] a zero-mean white noise sequence with covariance E( w k [wl T, vl T ]) = Q k v k S T k S k R k δ kl x 0 has mean x 0, covariance P 0, and is uncorrelated with the noise sequence We use here the following notation: Z k = (z 0 ; ; z k ) ˆx k k 1 = E L (x k Z k 1 ) ˆx k k = E L (x k Z k ) x k k 1 = x k ˆx k k 1 x k k = x k ˆx k k P k k 1 = cov( x k k 1 ) P k k = cov( x k k ) and defne the innovations process z k = zk E L (z k Z k 1 ) = z k H kˆx k k 1 Note that z k = H k x k k 1 + v k 12

13 Measurement update: From our previous discussion of linear estimation and innovations, ˆx k k = E L (x k Z k ) = E L (x k Z k ) = E L (x k Z k 1 ) + E L (x k z k ) E(x k ) This relation is the basis for the innovations approach The rest follows essentially by direct computations, and some use of the orthogonality principle First, E L (x k z k ) E(x k ) = cov(x k, z k )cov( z k ) 1 z k The two covariances are next computed: cov(x k, z k ) = cov(x k, H k x k k 1 + v k ) = P k k 1 Hk T, where E(x k x T k k 1 ) = P k k 1 follows by orthogonality, and we also used the fact that v k and x k are not correlated Similarly, cov( z k ) = cov(h k x k k 1 + v k ) = H k P k k 1 H T k + R k = Mk By substituting in the estimator expression we obtain ˆx k k = ˆx k k 1 + P k k 1 Hk T M 1 k z k Time update: This step is less trivial than before due to the correlation between v k and w k We have ˆx k+1 k = E L (x k+1 Z k ) = E L (F k x k + G k w k Z k ) = F kˆx k k + G k E L (w k z k ) In the last equation we used E L (w k Z k 1 ) = 0 since w k is uncorrelated with Z k 1 Thus ˆx k+1 k = F kˆx k k + G k E(w k z T k )cov( z k ) 1 z k = F kˆx k k + G k S k M 1 k z k where E(w k z T k ) = E(w kv T k ) = S k follows from z k = H k x k k 1 + v k 13

14 Combined update: Combining the measurement and time updates, we obtain the one-step update for ˆx k k 1 : ˆx k+1 k = F kˆx k k 1 + K k z k where K k = (Fk P k k 1 H k + G k S k )M 1 k z k = z k H kˆx k k 1 M k = H k P k k 1 H T k + R k Covariance update: The relation between P k k and P k k 1 is exactly as before The recursion for P k+1 k is most conveniently obtained in terms of P k k 1 directly From the previous relations we obtain x k+1 k = (F k K k H k ) x k k 1 + G k w k K k v k Since x k is uncorrelated with w k and v k, P k+1 k = (F k K k H k )P k k 1 (F k K k H k ) T + G k Q k G T k +K k R k K T k (G ks k K T k + K ks T k GT k ) This completes the filter equations for this case 14

15 Addendum: A Hilbert space interpretation The definitions and results concerning linear estimators can be nicely interpreted in terms of a Hilbert space formulation Consider for simplicity all RVs in this section to have 0 mean Recall that a Hilbert space is a (complete) inner-product space That is, it is a linear vector space V, with a real-valued inner product operation v 1, v 2 which is bi-linear, symmetric, and non-degenerate ( v, v = 0 iff v = 0) (Completeness means that every Cauchy sequence has a limit) The derived norm is defined as v 2 = v, v The following facts are standard: 1 A subspace S is a linearly-closed subset of V Alternatively, it is the linear span of some set of vectors {v α } 2 The orthogonal projection Π S v of a vector v unto the subspace S is the closest element to v in S, ie, the vector v S which minimizes v v Such a vector exists and is unique, and satisfies (v Π S v) S, ie, v Π S v, s = 0 for s S 3 If S = span{s 1,, s k }, then Π S v = k i=1 α is i, where [α 1,, α k ] = [ v, s 1,, v, s k ][ s i, s j i,j=1k ] 1 4 If S = S 1 S 2 (S is the direct sum of two orthogonal subspaces S 1 and S 2 ), then Π S v = Π S1 v + Π S2 v If {s 1,, s k } is an orthogonal basis of S, then Π S v = k v, s i s i, s i 1 s i i=1 15

16 5 Given a set of (independent) vectors {v 1, v 2 }, the following Gram-Schmidt procedure provides an orthogonal basis: ṽ k = v k Π span{v1 v k 1 }v k k 1 = v k v k, ṽ i ṽ i, ṽ i 1 v i i=1 We can fit the previous results on linear estimation to this framework by noting the following correspondence: Our Hilbert space is the space of all zero-mean random variables x (on a given probability space) which are square-integrable: E(x 2 ) = 0 The inner product in defined as x, y = E(xy) The optimal linear estimator E L (x k Z k ), with Z k = (z 0,, z k ), is the orthogonal projection of the vector x k on the subspace spanned by Z k (If x k is vector-valued, we simply consider the projection of each element separately) The innovations process {z k } is an orthogonalized version of {z k } The Hilbert space formulation provides a nice insight, and can also provide useful technical results, especially in the continuous-time case However, we shall not go deeper into this topic 16

17 44 The Kalman Filter as a Least-Squares Problem Consider the following deterministic optimization problem Cost function (to be minimized): J k = 1 2 (x 0 x 0 ) T P 1 0 (x 0 x 0 ) k l=0 l=0 (z l H l x l ) T R 1 l (z l H l x l ) k 1 wl T Q 1 l w l Constraints: Variables: x l+1 = F l x l + G l w l, l = 0, 1,, k 1 x 0, x k ; w 0, w k 1 Here x 0, {z l } are given vectors, and P 0, R l, Q l symmetric positive-definite matrices Let (x (k) o,, x (k) ) denote the optimal solution of this problem We claim that x(k) k can be computed exactly as ˆx k k in the corresponding KF problem This claim can be established by writing explicitly the least-squares solution for k 1 and k, and manipulating the matrix expressions We will take here a quicker route, using the Gaussian insight Theorem The minimizing solution (x (k) o,, x (k) ) of the above LS problem is the k maximizer of the conditional probability (that is, the M AP estimator): p(x 0,, x k Z k ), wrt(x o,, x k ) k 17

18 related to the Gaussian model: x k+1 = F k x k + G k w k, x 0 N(x 0, P 0 ) z k = H k x k + v k, w k N(0, Q k ), v k N(0, P k ) with w k, v k white and independent of x 0 Proof: Write down the distribution p(x 0 x k, Z k ) Immediate Consequence: Since for Gaussian RV s MAP=MMSE, then (x 0,, x k ) (k) are equivalent to the expected means: In particular, x (k) k = x + k Remark: The above theorem (but not the last consequence) holds true even for the non-linear model: x k+1 = F k (x k ) + G k w k 18

19 45 KF Equations Basic Versions a The basic equations Initial Conditions: ˆx 0 = x 0 = E(x0 ), P 0 = P 0 = cov(x0 ) Measurement update: ˆx + k = ˆx k + K k(z k H kˆx k ) P + k = P k K kh k P k where K k is the Kalman Gain matrix: K k = P k HT k (H k P k HT k + R k ) 1 Time update: ˆx k+1 = F kˆx + k [+B k u k ] P k+1 = F kp + k F T k + G k Q k G T k b One-step iterations The two-step equations may obviously be combined into a one-step update which computes ˆx + k+1 from ˆx+ k (or ˆx k+1 from ˆx k ) For example, ˆx k+1 = F kˆx k + F kk k (z k H kˆx k ) P k+1 = F k (P k K kh k P k )F k T + G k Q k G T k L k = Fk K k is also known as the Kalman gain The iterative equation for P k Riccati Equation is called the (discrete-time, time-varying) Matrix 19

20 c Other important quantities The measurement prediction, the innovations process, and the innovations covariance are given by ẑ k = E(zk Z k 1 ) = H kˆx k (+I ku k ) z k = zk ẑ k = H k x k M k = cov( zk ) = H k P k HT k + R k d Alternative Forms for the covariance update The measurement update for the (optimal) covariance P k may be expressed in the following equivalent formulas: P + k = P k K kh k P k We mention two alternative forms: = (I K k H k )P k = P k P k HT k M 1 k H kp k = P k K km k K T k 1 The Joseph form: Noting that x k ˆx + k = (I K kh k )(x k ˆx k ) K kv k it follows immediately that P + k = (I K kh k )P k (I K kh k ) T + K k R k K T k This form may be more computationally expensive, but has the following advantages: 20

21 It holds for any gain K k (not just the optimal) that is used in the estimator equation ˆx + k = ˆx k + K k z k Numerically, it is guaranteed to preserve positive-definiteness (P + k > 0) 2 Information form: (P + k ) 1 = (P k ) 1 + H k R 1 k H k The equivalence may obtained via the useful Matrix Inversion Lemma: (A + BCD) 1 = A 1 A 1 B(DA 1 B + C 1 ) 1 DA 1 where A, C are square nonsingular matrices (possibly of different size) P 1 is called the Information Matrix It forms the basis for the information filter, which only computes the inverse covariances e Relation to Deterministic Observers The one-step recursion for ˆx k control theory Given a (deterministic) system: is similar in form to the algebraic state observer from x k+1 = F k x k + B k u k z k = H k x k a state observer is defined by ˆx k+1 = F kˆx k + B k u k + L k (z k H kˆx k ) where L k are gain matrices to be chosen, with the goal of obtaining x k 0 as k = (xk ˆx k ) 21

22 Since x k+1 = (F k L k H k ) x k, we need to choose L k so that the linear system defined by A k = (F k L k H k ) is asymptotically stable This is possible when the original system is detectable The Kalman gain automatically satisfies this stability requirement (whenever the detectability condition is satisfied) 22

9 Multi-Model State Estimation

9 Multi-Model State Estimation Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State

More information

Lecture 4: Least Squares (LS) Estimation

Lecture 4: Least Squares (LS) Estimation ME 233, UC Berkeley, Spring 2014 Xu Chen Lecture 4: Least Squares (LS) Estimation Background and general solution Solution in the Gaussian case Properties Example Big picture general least squares estimation:

More information

RECURSIVE ESTIMATION AND KALMAN FILTERING

RECURSIVE ESTIMATION AND KALMAN FILTERING Chapter 3 RECURSIVE ESTIMATION AND KALMAN FILTERING 3. The Discrete Time Kalman Filter Consider the following estimation problem. Given the stochastic system with x k+ = Ax k + Gw k (3.) y k = Cx k + Hv

More information

Kalman Filtering. Namrata Vaswani. March 29, Kalman Filter as a causal MMSE estimator

Kalman Filtering. Namrata Vaswani. March 29, Kalman Filter as a causal MMSE estimator Kalman Filtering Namrata Vaswani March 29, 2018 Notes are based on Vincent Poor s book. 1 Kalman Filter as a causal MMSE estimator Consider the following state space model (signal and observation model).

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

The Kalman Filter ImPr Talk

The Kalman Filter ImPr Talk The Kalman Filter ImPr Talk Ged Ridgway Centre for Medical Image Computing November, 2006 Outline What is the Kalman Filter? State Space Models Kalman Filter Overview Bayesian Updating of Estimates Kalman

More information

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter D. Richard Brown III Worcester Polytechnic Institute 09-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 09-Apr-2009 1 /

More information

ESTIMATION THEORY. Chapter Estimation of Random Variables

ESTIMATION THEORY. Chapter Estimation of Random Variables Chapter ESTIMATION THEORY. Estimation of Random Variables Suppose X,Y,Y 2,...,Y n are random variables defined on the same probability space (Ω, S,P). We consider Y,...,Y n to be the observed random variables

More information

6.4 Kalman Filter Equations

6.4 Kalman Filter Equations 6.4 Kalman Filter Equations 6.4.1 Recap: Auxiliary variables Recall the definition of the auxiliary random variables x p k) and x m k): Init: x m 0) := x0) S1: x p k) := Ak 1)x m k 1) +uk 1) +vk 1) S2:

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance The Kalman Filter Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience Sarah Dance School of Mathematical and Physical Sciences, University of Reading s.l.dance@reading.ac.uk July

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

LINEAR MMSE ESTIMATION

LINEAR MMSE ESTIMATION LINEAR MMSE ESTIMATION TERM PAPER FOR EE 602 STATISTICAL SIGNAL PROCESSING By, DHEERAJ KUMAR VARUN KHAITAN 1 Introduction Linear MMSE estimators are chosen in practice because they are simpler than the

More information

ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form

ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER KRISTOFFER P. NIMARK The Kalman Filter We will be concerned with state space systems of the form X t = A t X t 1 + C t u t 0.1 Z t

More information

Solutions to Homework Set #6 (Prepared by Lele Wang)

Solutions to Homework Set #6 (Prepared by Lele Wang) Solutions to Homework Set #6 (Prepared by Lele Wang) Gaussian random vector Given a Gaussian random vector X N (µ, Σ), where µ ( 5 ) T and 0 Σ 4 0 0 0 9 (a) Find the pdfs of i X, ii X + X 3, iii X + X

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26 Estimation: Regression and Least Squares This note explains how to use observations to estimate unobserved random variables.

More information

X t = a t + r t, (7.1)

X t = a t + r t, (7.1) Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical

More information

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011 UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and

More information

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE53 Handout #34 Prof Young-Han Kim Tuesday, May 7, 04 Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) Linear estimator Consider a channel with the observation Y XZ, where the

More information

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Ch. 12 Linear Bayesian Estimators

Ch. 12 Linear Bayesian Estimators Ch. 1 Linear Bayesian Estimators 1 In chapter 11 we saw: the MMSE estimator takes a simple form when and are jointly Gaussian it is linear and used only the 1 st and nd order moments (means and covariances).

More information

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I

More information

X n = c n + c n,k Y k, (4.2)

X n = c n + c n,k Y k, (4.2) 4. Linear filtering. Wiener filter Assume (X n, Y n is a pair of random sequences in which the first component X n is referred a signal and the second an observation. The paths of the signal are unobservable

More information

A Crash Course on Kalman Filtering

A Crash Course on Kalman Filtering A Crash Course on Kalman Filtering Dan Simon Cleveland State University Fall 2014 1 / 64 Outline Linear Systems Probability State Means and Covariances Least Squares Estimation The Kalman Filter Unknown

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2 MA 575 Linear Models: Cedric E Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2 1 Revision: Probability Theory 11 Random Variables A real-valued random variable is

More information

Least Squares Estimation Namrata Vaswani,

Least Squares Estimation Namrata Vaswani, Least Squares Estimation Namrata Vaswani, namrata@iastate.edu Least Squares Estimation 1 Recall: Geometric Intuition for Least Squares Minimize J(x) = y Hx 2 Solution satisfies: H T H ˆx = H T y, i.e.

More information

ENGR352 Problem Set 02

ENGR352 Problem Set 02 engr352/engr352p02 September 13, 2018) ENGR352 Problem Set 02 Transfer function of an estimator 1. Using Eq. (1.1.4-27) from the text, find the correct value of r ss (the result given in the text is incorrect).

More information

Statistics 910, #15 1. Kalman Filter

Statistics 910, #15 1. Kalman Filter Statistics 910, #15 1 Overview 1. Summary of Kalman filter 2. Derivations 3. ARMA likelihoods 4. Recursions for the variance Kalman Filter Summary of Kalman filter Simplifications To make the derivations

More information

The Kalman filter. Chapter 6

The Kalman filter. Chapter 6 Chapter 6 The Kalman filter In the last chapter, we saw that in Data Assimilation, we ultimately desire knowledge of the full a posteriori p.d.f., that is the conditional p.d.f. of the state given the

More information

State Estimation of Linear and Nonlinear Dynamic Systems

State Estimation of Linear and Nonlinear Dynamic Systems State Estimation of Linear and Nonlinear Dynamic Systems Part I: Linear Systems with Gaussian Noise James B. Rawlings and Fernando V. Lima Department of Chemical and Biological Engineering University of

More information

Appendix A : Introduction to Probability and stochastic processes

Appendix A : Introduction to Probability and stochastic processes A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of

More information

EE226a - Summary of Lecture 13 and 14 Kalman Filter: Convergence

EE226a - Summary of Lecture 13 and 14 Kalman Filter: Convergence 1 EE226a - Summary of Lecture 13 and 14 Kalman Filter: Convergence Jean Walrand I. SUMMARY Here are the key ideas and results of this important topic. Section II reviews Kalman Filter. A system is observable

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

Least Squares and Kalman Filtering Questions: me,

Least Squares and Kalman Filtering Questions:  me, Least Squares and Kalman Filtering Questions: Email me, namrata@ece.gatech.edu Least Squares and Kalman Filtering 1 Recall: Weighted Least Squares y = Hx + e Minimize Solution: J(x) = (y Hx) T W (y Hx)

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

4.4. Orthogonality. Note. This section is awesome! It is very geometric and shows that much of the geometry of R n holds in Hilbert spaces.

4.4. Orthogonality. Note. This section is awesome! It is very geometric and shows that much of the geometry of R n holds in Hilbert spaces. 4.4. Orthogonality 1 4.4. Orthogonality Note. This section is awesome! It is very geometric and shows that much of the geometry of R n holds in Hilbert spaces. Definition. Elements x and y of a Hilbert

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49 State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing

More information

Projection Theorem 1

Projection Theorem 1 Projection Theorem 1 Cauchy-Schwarz Inequality Lemma. (Cauchy-Schwarz Inequality) For all x, y in an inner product space, [ xy, ] x y. Equality holds if and only if x y or y θ. Proof. If y θ, the inequality

More information

Inner Product Spaces

Inner Product Spaces Inner Product Spaces Introduction Recall in the lecture on vector spaces that geometric vectors (i.e. vectors in two and three-dimensional Cartesian space have the properties of addition, subtraction,

More information

FIR Filters for Stationary State Space Signal Models

FIR Filters for Stationary State Space Signal Models Proceedings of the 17th World Congress The International Federation of Automatic Control FIR Filters for Stationary State Space Signal Models Jung Hun Park Wook Hyun Kwon School of Electrical Engineering

More information

Exercise Sheet 1.

Exercise Sheet 1. Exercise Sheet 1 You can download my lecture and exercise sheets at the address http://sami.hust.edu.vn/giang-vien/?name=huynt 1) Let A, B be sets. What does the statement "A is not a subset of B " mean?

More information

11 - The linear model

11 - The linear model 11-1 The linear model S. Lall, Stanford 2011.02.15.01 11 - The linear model The linear model The joint pdf and covariance Example: uniform pdfs The importance of the prior Linear measurements with Gaussian

More information

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes Shannon meets Wiener II: On MMSE estimation in successive decoding schemes G. David Forney, Jr. MIT Cambridge, MA 0239 USA forneyd@comcast.net Abstract We continue to discuss why MMSE estimation arises

More information

Kalman Filters with Uncompensated Biases

Kalman Filters with Uncompensated Biases Kalman Filters with Uncompensated Biases Renato Zanetti he Charles Stark Draper Laboratory, Houston, exas, 77058 Robert H. Bishop Marquette University, Milwaukee, WI 53201 I. INRODUCION An underlying assumption

More information

Kalman Filter and Parameter Identification. Florian Herzog

Kalman Filter and Parameter Identification. Florian Herzog Kalman Filter and Parameter Identification Florian Herzog 2013 Continuous-time Kalman Filter In this chapter, we shall use stochastic processes with independent increments w 1 (.) and w 2 (.) at the input

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

ECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018

ECE534, Spring 2018: Solutions for Problem Set #4 Due Friday April 6, 2018 ECE534, Spring 2018: s for Problem Set #4 Due Friday April 6, 2018 1. MMSE Estimation, Data Processing and Innovations The random variables X, Y, Z on a common probability space (Ω, F, P ) are said to

More information

Optimal control and estimation

Optimal control and estimation Automatic Control 2 Optimal control and estimation Prof. Alberto Bemporad University of Trento Academic year 2010-2011 Prof. Alberto Bemporad (University of Trento) Automatic Control 2 Academic year 2010-2011

More information

Kalman Filter. Man-Wai MAK

Kalman Filter. Man-Wai MAK Kalman Filter Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: S. Gannot and A. Yeredor,

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Evan Kwiatkowski, Jan Mandel University of Colorado Denver December 11, 2014 OUTLINE 2 Data Assimilation Bayesian Estimation

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7 Linear Algebra and its Applications-Lab 1 1) Use Gaussian elimination to solve the following systems x 1 + x 2 2x 3 + 4x 4 = 5 1.1) 2x 1 + 2x 2 3x 3 + x 4 = 3 3x 1 + 3x 2 4x 3 2x 4 = 1 x + y + 2z = 4 1.4)

More information

x log x, which is strictly convex, and use Jensen s Inequality:

x log x, which is strictly convex, and use Jensen s Inequality: 2. Information measures: mutual information 2.1 Divergence: main inequality Theorem 2.1 (Information Inequality). D(P Q) 0 ; D(P Q) = 0 iff P = Q Proof. Let ϕ(x) x log x, which is strictly convex, and

More information

Statistical Filtering and Control for AI and Robotics. Part II. Linear methods for regression & Kalman filtering

Statistical Filtering and Control for AI and Robotics. Part II. Linear methods for regression & Kalman filtering Statistical Filtering and Control for AI and Robotics Part II. Linear methods for regression & Kalman filtering Riccardo Muradore 1 / 66 Outline Linear Methods for Regression Gaussian filter Stochastic

More information

Kalman filter using the orthogonality principle

Kalman filter using the orthogonality principle Appendix G Kalman filter using the orthogonality principle This appendix presents derivation steps for obtaining the discrete Kalman filter equations using a method based on the orthogonality principle.

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

UCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE 53 Handout #46 Prof. Young-Han Kim Thursday, June 5, 04 Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei). Discrete-time Wiener process. Let Z n, n 0 be a discrete time white

More information

Organization. I MCMC discussion. I project talks. I Lecture.

Organization. I MCMC discussion. I project talks. I Lecture. Organization I MCMC discussion I project talks. I Lecture. Content I Uncertainty Propagation Overview I Forward-Backward with an Ensemble I Model Reduction (Intro) Uncertainty Propagation in Causal Systems

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53 State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State

More information

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017) UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable

More information

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. Orthogonal sets Let V be a vector space with an inner product. Definition. Nonzero vectors v 1,v

More information

Lecture Note 1: Probability Theory and Statistics

Lecture Note 1: Probability Theory and Statistics Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would

More information

1 Compact and Precompact Subsets of H

1 Compact and Precompact Subsets of H Compact Sets and Compact Operators by Francis J. Narcowich November, 2014 Throughout these notes, H denotes a separable Hilbert space. We will use the notation B(H) to denote the set of bounded linear

More information

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 205 Motivation When working with an inner product space, the most

More information

State Estimation for Nonlinear Systems using Restricted Genetic Optimization

State Estimation for Nonlinear Systems using Restricted Genetic Optimization State Estimation for Nonlinear Systems using Restricted Genetic Optimization Santiago Garrido, Luis Moreno, and Carlos Balaguer Universidad Carlos III de Madrid, Leganés 28911, Madrid (Spain) Abstract.

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: ony Jebara Kalman Filtering Linear Dynamical Systems and Kalman Filtering Structure from Motion Linear Dynamical Systems Audio: x=pitch y=acoustic waveform Vision: x=object

More information

MATH Linear Algebra

MATH Linear Algebra MATH 304 - Linear Algebra In the previous note we learned an important algorithm to produce orthogonal sequences of vectors called the Gramm-Schmidt orthogonalization process. Gramm-Schmidt orthogonalization

More information

Covariance Matrix Simplification For Efficient Uncertainty Management

Covariance Matrix Simplification For Efficient Uncertainty Management PASEO MaxEnt 2007 Covariance Matrix Simplification For Efficient Uncertainty Management André Jalobeanu, Jorge A. Gutiérrez PASEO Research Group LSIIT (CNRS/ Univ. Strasbourg) - Illkirch, France *part

More information

Lecture 7: Optimal Smoothing

Lecture 7: Optimal Smoothing Department of Biomedical Engineering and Computational Science Aalto University March 17, 2011 Contents 1 What is Optimal Smoothing? 2 Bayesian Optimal Smoothing Equations 3 Rauch-Tung-Striebel Smoother

More information

Linear Algebra. Preliminary Lecture Notes

Linear Algebra. Preliminary Lecture Notes Linear Algebra Preliminary Lecture Notes Adolfo J. Rumbos c Draft date May 9, 29 2 Contents 1 Motivation for the course 5 2 Euclidean n dimensional Space 7 2.1 Definition of n Dimensional Euclidean Space...........

More information

Lessons in Estimation Theory for Signal Processing, Communications, and Control

Lessons in Estimation Theory for Signal Processing, Communications, and Control Lessons in Estimation Theory for Signal Processing, Communications, and Control Jerry M. Mendel Department of Electrical Engineering University of Southern California Los Angeles, California PRENTICE HALL

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part X Factor analysis When we have data x (i) R n that comes from a mixture of several Gaussians, the EM algorithm can be applied to fit a mixture model. In this setting,

More information

Lecture 3: Functions of Symmetric Matrices

Lecture 3: Functions of Symmetric Matrices Lecture 3: Functions of Symmetric Matrices Yilin Mo July 2, 2015 1 Recap 1 Bayes Estimator: (a Initialization: (b Correction: f(x 0 Y 1 = f(x 0 f(x k Y k = αf(y k x k f(x k Y k 1, where ( 1 α = f(y k x

More information

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π Solutions to Homework Set #5 (Prepared by Lele Wang). Neural net. Let Y X + Z, where the signal X U[,] and noise Z N(,) are independent. (a) Find the function g(y) that minimizes MSE E [ (sgn(x) g(y))

More information

Topics in Probability and Statistics

Topics in Probability and Statistics Topics in Probability and tatistics A Fundamental Construction uppose {, P } is a sample space (with probability P), and suppose X : R is a random variable. The distribution of X is the probability P X

More information

ECE534, Spring 2018: Solutions for Problem Set #3

ECE534, Spring 2018: Solutions for Problem Set #3 ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation. So: How can we evaluate E(X Y )? What is a neural network? How is it useful?

Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation. So: How can we evaluate E(X Y )? What is a neural network? How is it useful? Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation The Story So Far... So: How can we evaluate E(X Y )? What is a neural network? How is it useful? EE 278: Using Neural Networks for

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

A Theoretical Overview on Kalman Filtering

A Theoretical Overview on Kalman Filtering A Theoretical Overview on Kalman Filtering Constantinos Mavroeidis Vanier College Presented to professors: IVANOV T. IVAN STAHN CHRISTIAN Email: cmavroeidis@gmail.com June 6, 208 Abstract Kalman filtering

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms L06. LINEAR KALMAN FILTERS NA568 Mobile Robotics: Methods & Algorithms 2 PS2 is out! Landmark-based Localization: EKF, UKF, PF Today s Lecture Minimum Mean Square Error (MMSE) Linear Kalman Filter Gaussian

More information

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1) EXERCISE SET 5. 6. The pair (, 2) is in the set but the pair ( )(, 2) = (, 2) is not because the first component is negative; hence Axiom 6 fails. Axiom 5 also fails. 8. Axioms, 2, 3, 6, 9, and are easily

More information

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical

More information

Statistical Signal Processing Detection, Estimation, and Time Series Analysis

Statistical Signal Processing Detection, Estimation, and Time Series Analysis Statistical Signal Processing Detection, Estimation, and Time Series Analysis Louis L. Scharf University of Colorado at Boulder with Cedric Demeure collaborating on Chapters 10 and 11 A TT ADDISON-WESLEY

More information

Linear Algebra. Preliminary Lecture Notes

Linear Algebra. Preliminary Lecture Notes Linear Algebra Preliminary Lecture Notes Adolfo J. Rumbos c Draft date April 29, 23 2 Contents Motivation for the course 5 2 Euclidean n dimensional Space 7 2. Definition of n Dimensional Euclidean Space...........

More information

ECE Lecture #9 Part 2 Overview

ECE Lecture #9 Part 2 Overview ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by

More information

Lecture 6: April 19, 2002

Lecture 6: April 19, 2002 EE596 Pat. Recog. II: Introduction to Graphical Models Spring 2002 Lecturer: Jeff Bilmes Lecture 6: April 19, 2002 University of Washington Dept. of Electrical Engineering Scribe: Huaning Niu,Özgür Çetin

More information

Gaussian random variables inr n

Gaussian random variables inr n Gaussian vectors Lecture 5 Gaussian random variables inr n One-dimensional case One-dimensional Gaussian density with mean and standard deviation (called N, ): fx x exp. Proposition If X N,, then ax b

More information

UCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE53 Handout #7 Prof. Young-Han Kim Tuesday, May 6, 4 Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei). Neural net. Let Y = X + Z, where the signal X U[,] and noise Z N(,) are independent.

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by

More information

1 Kalman Filter Introduction

1 Kalman Filter Introduction 1 Kalman Filter Introduction You should first read Chapter 1 of Stochastic models, estimation, and control: Volume 1 by Peter S. Maybec (available here). 1.1 Explanation of Equations (1-3) and (1-4) Equation

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Lecture 6: Bayesian Inference in SDE Models

Lecture 6: Bayesian Inference in SDE Models Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs

More information

ECE531 Lecture 12: Linear Estimation and Causal Wiener-Kolmogorov Filtering

ECE531 Lecture 12: Linear Estimation and Causal Wiener-Kolmogorov Filtering ECE531 Lecture 12: Linear Estimation and Causal Wiener-Kolmogorov Filtering D. Richard Brown III Worcester Polytechnic Institute 16-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 16-Apr-2009

More information