Kalman Filtering. Namrata Vaswani. March 29, Kalman Filter as a causal MMSE estimator
|
|
- Adam Webb
- 6 years ago
- Views:
Transcription
1 Kalman Filtering Namrata Vaswani March 29, 2018 Notes are based on Vincent Poor s book. 1 Kalman Filter as a causal MMSE estimator Consider the following state space model (signal and observation model). t = H t X t + W t, W t N (0, R) (1) X t = F t X t 1 + U t, U t N (0, Q) (2) where X 0, {U t, t = 1,... }, {W t, t = 0,... } are mutually independent and X 0 N (0, Σ 0 ). Recall that Cov[X E[(X E[X )(X E[X ) T. Define ˆX t s E[X t 0:s Σ t s Cov[X t 0:s = E[(X t ˆX t s )( ) T 0:s (3) Thus ˆX t s is the MMSE estimator of X t from observations 0:s and Σ t s is its error covariance conditioned in 0:s. We claim that that ˆX t t and ˆX t t 1 satisfy the following recursion (Kalman filter). ˆX t t 1 = F t ˆXt 1 t 1 Σ t t 1 = F t Σ t 1 t 1 F T t + Q, with initialization, ˆX0 1 = 0, Σ 0 1 = Σ 0. K t = Σ t t 1 H T t (H t Σ t t 1 H T t + R) 1 ˆX t t = ˆX t t 1 + K t ( t H t ˆXt t 1 ) Σ t t = [I K t H t Σ t t 1 (4)
2 1.1 Conditional Gaussian Distribution In the proof we will need the following result for jointly Gaussian random variables. If X, have the joint PDF [ then X N ( [ µ µ X, [ Σ Σ X Σ X Σ X E[X = µ X + Σ X Σ 1 ( µ ) Cov[X = Σ X Σ X Σ 1 Σ X, (5) Proof: One way to prove this is to write out the expression for the conditional PDF and use the block matrix inversion lemma. A shorter and nicer proof is as follows. The idea is to define a r.v. Z that is a linear function of X and and is such that Cov(Z, ) = 0. Because it is a linear function, Z and are also jointly Gaussian and hence cov = 0 will imply independence. Then if one writes of X as a linear function of Z and, getting the above quantities becomes very easy because E[f(Z) = E[f(Z). 1. Let Z = X + B. We want Cov(Z, ) = 0. But notice that Cov(Z, ) = E[(X µ X )( µ ) T +BE[( µ )( µ ) T = Σ X +BΣ. Thus if we let B = Σ X Σ 1 we will get Cov(Z, ) = 0 and so using joint-guassianity Z and are independent. 2. Thus, Z = X Σ X Σ 1 and so X = Z +Σ X Σ 1. Also, Z and are independent. Thus, E[X = E[Z + Σ X Σ 1 ) = µ X Σ X Σ 1 µ + Σ X Σ 1 (6) and since Cov(Z ) = Cov(Z), Cov( ) = 0 and Cov(Z, B ) = E[(Z µ Z ) ( µ ) T = E[(Z µ Z )( µ ) T = 0, we get Cov[X = Cov(Z B, Z B ) where B := Σ X Σ 1 = Cov(Z ) + Cov(B ) Cov(Z, B ) Cov(B, Z ) = Cov(Z) = Σ X + Σ X Σ 1 Σ Σ 1 Σ X Σ X Σ 1 Σ X Σ X Σ 1 Σ X = Σ X Σ X Σ 1 Σ X (7) 1.2 Proof of KF as MMSE estimator 1. Use induction. Base case for t = 0 follows directly from the signal model. Assume that (4) holds at t 1.
3 2. To compute ˆX t t 1, take E[ 0:t 1 on the signal model, (2). Then use the fact that U t is independent of 0:t 1 = f(x 0, U 0:t 1, W 0:t 1 ) to show that E[U t 0:t 1 = E[U t = 0 (since U t is zero mean). 3. To compute Σ t t 1, take Cov[ 0:t 1 on the signal model, (2). Show that the crossterms which contain E[(X t 1 ˆX t 1 t 1 )U T t 0:t 1 or its transpose are zero using the following approach. (a) Let Z X t 1 ˆX t 1 t 1. (b) It is easy to see that both Z and 0:t 1 are functions of X 0, U 0:t 1, W 0:t 1, i.e. {Z, 0:t 1 } = f((x 0, U 0:t 1, W 0:t 1 ). Thus, U t is independent of {Z, 0:t 1 } (using the model assumption). (c) U independent of {Z, } implies (i) U independent of ; (ii) U independent of Z given (conditionally independent). Proof: U independent of {Z, } implies that f(y, u) = f(z, y, u)dz = f(z, y)f(u)dz = f(y)f(u). Thus (i) follows. To show (ii), notice that f(z, u y) = f(z y)f(u z, y) by chain rule. U independent of {Z, } implies that f(u z, y) = f(u). Also, (i) implies that f(u) = f(u y). Thus, f(z, u y) = f(z y)f(u y) and thus (ii) holds. (d) Thus U t, Z are conditionally independent given 0:t 1. (e) Thus the cross term, E[ZU T t 0:t 1 = E[Z 0:t 1 E[U t 0:t 1 T = E[Z 0:t 1 E[U t T = 0 and the same holds for its transpose. The second equality follows because U t is independent of 0:t 1 while the third follows because U t is zero mean. 4. For the update step, first derive the expression for the joint pdf of X t, t conditioned on 0:t 1, f(x t, y t y 0:t 1 ), and then use the conditional mean and covariance formula for joint Gaussians given in (5), to obtain the expression for ˆX t t and Σ t t. To obtain the joint pdf expression, use the following approach. (a) Since the three random variables are jointly Gaussian, f(x t, y t y 0:t 1 ) will also be Gaussian. Thus we only need expressions for its mean and covariance. Clearly, [ [ f(x t, y t y 0:t 1 ) = N ( Σ X = Σ t t 1, x t y t ; [ ˆXt t 1 H ˆX t t 1, Σ X Σ T X Σ X Σ X ), where Σ X = Cov[X t, t 0:t 1 = E[(X t ˆX t t 1 )(H(X t ˆX t t 1 ) + W t ) T 0:t 1 = Σ t t 1 H T + cross 1 = Σ t t 1 H T + 0, Σ = Cov[ t 0:t 1 = HΣ t t 1 H T + R + cross 2 + cross T 2 = HΣ t t 1 H T + R
4 (b) To show that cross 1 = 0 and cross 2 = 0, use the fact that (X t ˆX t t 1 ) and W t are conditionally independent given 0:t 1 ; W t is independent of 0:t 1 and W t is zero mean. These can be shown in a fashion analogous to step 3. This finishes the proof of the induction step and thus of the result. A few important things to notice are as follows. 1. Since the expressions for Σ t t 1 and Σ t t do not depend on 0:t 1, thus these are also equal to the unconditional error covariances, i.e. Σ t t 1 E[(X t ˆX t t 1 )( ) T 0:t 1 = E[(X t ˆX t t 1 )( ) T and similarly, Σ t t E[(X t ˆX t t )( ) T 0:t = E[(X t ˆX t t )( ) T 2. It can be shown that the innovations, Z t t H ˆX t t 1, are pairwise uncorrelated. Since they are jointly Gaussian, this means they are mutually independent. (a) To show uncorrelated-ness one needs to use E[E[Z Z 1, Z 2 Z 1 = E[Z Z 1 (law of iterated expectations applied to Z = Z Z 1 ). (b) Consider an s < t, E[Z t Z T s = E[E[Z t Z T s 0:s = E[E[Z t 0:s Z T s = E[E[( t H ˆX t t 1 ) 0:s Z T s = E[(E[ t 0:s E[H ˆX t t 1 0:s )Z T s = E[(H ˆX t s E[H ˆX t t 1 0:s )Z T s = E[(H ˆX t s H ˆX t s )Z T s = 0 (8) We used E[E[Z Z 1, Z 2 Z 1 = E[Z Z 1 to show that E[H ˆX t t 1 0:s = H ˆX t s. Here Z X t, Z 1 0:s, Z 2 s+1:t. 3. Another expression for K t : 1.3 Kalman filter with control input Consider a state space model of the form t = HX t + r( 1, 2,... t 1 ) + W t, W t N (0, R) X t = F X t 1 + q( 1, 2,... t 1 ) + GU t, U t N (0, Q) with X 0, {U t } t=1, {W t } t=0 being mutually independent and X 0 N (0, Σ 0 ). The above is a state space model, but with a nonzero feedback control input in both equations. To derive the Kalman recursion for this model, use the exact same procedure
5 outlined above. Since at time t, the control inputs are functions of 0:t 1, when we take E[ 0:t 1 of either the signal model or the observation model, r and q just get pulled out as constants. Thus the expressions for the error covariances does not change at all. 2 Kalman filter as a causal linear MMSE estimator Consider the state space model of (1), (2), but with the difference that X 0, U t, W t s are no longer Gaussian, but are just some zero mean random variables with the given covariances. Also, instead of being mutually independent, they are only pairwise uncorrelated. For this model, the Kalman filter of (4) is the causal linear MMSE estimator, i.e. ˆXt t 1 is the linear MMSE of X t from 0:t 1, ˆXt t is the linear MMSE of X t from 0:t, and the unconditional error covariances, E[(X t ˆX t t 1 )(.) T = Σ t t 1 and E[(X t ˆX t t )(.) T = Σ t t. We show the above using the orthogonality principle: ˆX is the linear MMSE of X from observations 0:n 1 if and only if Outline of Proof. E[(X ˆX) = 0 and E[(X ˆX) T l = 0, l = 0, 1,... n 1, (9) 1. Use induction. The base case is easy, since you are using no observations. Assume that our result holds for t 1, i.e. E[(X t 1 ˆX t 1 t 1 )(.) T = Σ t 1 t The expression for ˆX t t 1 is given in (4). Thus, E[(X t ˆX t t 1 ) T l ˆXt 1 t 1 is linear MMSE of X t 1 from 0:t 1 and = F E[(X t 1 ˆX t 1 t 1 )l T + F E[U t l T = 0 + 0, l = 0, 1,... t 1(10) The first term is zero because of the induction hypothesis and the orthogonality principle. The second term is zero because of uncorrelated-ness of U t and l (follows because l is a linear function of X 0, U 1:l, W 0:l ). Also, E[(X t ˆX t t 1 ) = F E[X t 1 ˆX t 1 t 1 + E[U t = 0 (11) This follows by induction assumption and orthogonality principle and since U t is zero mean. Thus by the orthogonality principle, ˆXt t 1 is the L-MMSE of X t from 0:t It is easy to show that E[(X t ˆX t t 1 )(.) T = F Σ t 1 t 1 F T + Q = Σ t t 1 by showing that the cross-terms are zero. The cross-terms are of the form E[(X t 1 ˆX t 1 t 1 )U T t (or its transpose). They are zero since (X t 1 ˆX t 1 t 1 ) is a linear function of X 0, U 1:t 1, W 0:t 1, all of which are uncorrelated with U t, and U t is zero mean.
6 4. The expression for ˆX t t is given in (4). Using the expression for t, X t ˆX t t = X t [(I K t H) ˆX t t 1 + K t t = (I K t H)(X t ˆX t t 1 ) + K t W t (12) Thus, E[(X t ˆX t t ) T l = (I K t H)E[(X t ˆX t t 1 )l T + K t E[W t l T (13) For l = 0, 1,... t 1, one can use the previous step; uncorrelated-ness of W t and 0:t 1 and W t being zero mean, to show that the above is zero. Consider l = t. E[(X t ˆX t t )t T = (I K t H)E[(X t ˆX t t 1 )t T K t E[W t t T = (I K t H)E[(X t ˆX t t 1 )(HX t + W t ) T K t E[W t W T t = (I K t H)Σ t t 1 H T K t R = 0 (14) (a) The last equality follows from the expression for K t. (b) The second-last one follows because i. E[(X t ˆX t t 1 )X T t = E[(X t ˆX t t 1 )(X t ˆX t t 1 + ˆX t t 1 ) T = Σ t t 1 +E[(X t ˆX t t 1 ) ˆX T t t 1. ii. E[(X t ˆX t t 1 ) ˆX T t t 1 = E[E[(X t ˆX t t 1 ) 0:t 1 ˆX T t t 1 = E[0. ˆX T t t 1 = 0 iii. W t and (X t ˆX t t 1 ) are uncorrelated and W t is zero mean. (c) The third-last one follows using (1) and the fact that W t and X t are uncorrelated and W t is zero mean. Also, using (12), E[(X t ˆX t t ) = (I K t H)E[X t ˆX t t 1 + K t E[W t = 0 (15) The first term is zero by prediction step claim and orthogonality principle; the second term is zero since W t is zero mean. Thus by orthogonality principle, ˆXt t is L-MMSE of X t from 0:t. 5. To get the expression for E[(X t ˆX t t )(.) T, expand it out using the expression for ˆX t t from (4) and then simplify it using (1) and the expression for K t. (a) E[(X t ˆX t t )(.) T = (I K t H)E[(X t ˆX t t 1 )(.) T (I K t H) T + K t E[W t W T t K T t + cross 3 + cross T 3 = (I K t H)Σ t t 1 (I K t H) T + K t RK T t (b) The cross term, cross 3, contains E[(X t ˆX t t 1 )W T t which can be shown to be zero since W t and (X t ˆX t t 1 ) are uncorrelated and W t is zero mean. (c) Use the expression for K t and simplify to show that (I K t H)Σ t t 1 (I K t H) T + K t RK T t = (I K t H)Σ t t 1 : homework problem.
4 Derivations of the Discrete-Time Kalman Filter
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time
More informationSolutions to Homework Set #6 (Prepared by Lele Wang)
Solutions to Homework Set #6 (Prepared by Lele Wang) Gaussian random vector Given a Gaussian random vector X N (µ, Σ), where µ ( 5 ) T and 0 Σ 4 0 0 0 9 (a) Find the pdfs of i X, ii X + X 3, iii X + X
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter
ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter D. Richard Brown III Worcester Polytechnic Institute 09-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 09-Apr-2009 1 /
More informationStatistics 910, #15 1. Kalman Filter
Statistics 910, #15 1 Overview 1. Summary of Kalman filter 2. Derivations 3. ARMA likelihoods 4. Recursions for the variance Kalman Filter Summary of Kalman filter Simplifications To make the derivations
More informationLecture Note 12: Kalman Filter
ECE 645: Estimation Theory Spring 2015 Instructor: Prof. Stanley H. Chan Lecture Note 12: Kalman Filter LaTeX prepared by Stylianos Chatzidakis) May 4, 2015 This lecture note is based on ECE 645Spring
More information9 Multi-Model State Estimation
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State
More informationLecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel
Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More informationShannon meets Wiener II: On MMSE estimation in successive decoding schemes
Shannon meets Wiener II: On MMSE estimation in successive decoding schemes G. David Forney, Jr. MIT Cambridge, MA 0239 USA forneyd@comcast.net Abstract We continue to discuss why MMSE estimation arises
More informationUCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011
UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and
More informationChapter 5 Class Notes
Chapter 5 Class Notes Sections 5.1 and 5.2 It is quite common to measure several variables (some of which may be correlated) and to examine the corresponding joint probability distribution One example
More informationECE Lecture #9 Part 2 Overview
ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationLecture 3: Functions of Symmetric Matrices
Lecture 3: Functions of Symmetric Matrices Yilin Mo July 2, 2015 1 Recap 1 Bayes Estimator: (a Initialization: (b Correction: f(x 0 Y 1 = f(x 0 f(x k Y k = αf(y k x k f(x k Y k 1, where ( 1 α = f(y k x
More informationECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1
EE 650 Lecture 4 Intro to Estimation Theory Random Vectors EE 650 D. Van Alphen 1 Lecture Overview: Random Variables & Estimation Theory Functions of RV s (5.9) Introduction to Estimation Theory MMSE Estimation
More informationStatistics Homework #4
Statistics 910 1 Homework #4 Chapter 6, Shumway and Stoffer These are outlines of the solutions. If you would like to fill in other details, please come see me during office hours. 6.1 State-space representation
More information6.4 Kalman Filter Equations
6.4 Kalman Filter Equations 6.4.1 Recap: Auxiliary variables Recall the definition of the auxiliary random variables x p k) and x m k): Init: x m 0) := x0) S1: x p k) := Ak 1)x m k 1) +uk 1) +vk 1) S2:
More informationLecture 4: Least Squares (LS) Estimation
ME 233, UC Berkeley, Spring 2014 Xu Chen Lecture 4: Least Squares (LS) Estimation Background and general solution Solution in the Gaussian case Properties Example Big picture general least squares estimation:
More informationLeast Squares Estimation Namrata Vaswani,
Least Squares Estimation Namrata Vaswani, namrata@iastate.edu Least Squares Estimation 1 Recall: Geometric Intuition for Least Squares Minimize J(x) = y Hx 2 Solution satisfies: H T H ˆx = H T y, i.e.
More informationDetection and Estimation Theory
Detection and Estimation Theory Instructor: Prof. Namrata Vaswani Dept. of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/ namrata Slide 1 What is Estimation and Detection
More informationCh. 12 Linear Bayesian Estimators
Ch. 1 Linear Bayesian Estimators 1 In chapter 11 we saw: the MMSE estimator takes a simple form when and are jointly Gaussian it is linear and used only the 1 st and nd order moments (means and covariances).
More informationUCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE53 Handout #34 Prof Young-Han Kim Tuesday, May 7, 04 Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) Linear estimator Consider a channel with the observation Y XZ, where the
More informationECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form
ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER KRISTOFFER P. NIMARK The Kalman Filter We will be concerned with state space systems of the form X t = A t X t 1 + C t u t 0.1 Z t
More informationSummary of EE226a: MMSE and LLSE
1 Summary of EE226a: MMSE and LLSE Jean Walrand Here are the key ideas and results: I. SUMMARY The MMSE of given Y is E[ Y]. The LLSE of given Y is L[ Y] = E() + K,Y K 1 Y (Y E(Y)) if K Y is nonsingular.
More informationLeast Squares and Kalman Filtering Questions: me,
Least Squares and Kalman Filtering Questions: Email me, namrata@ece.gatech.edu Least Squares and Kalman Filtering 1 Recall: Weighted Least Squares y = Hx + e Minimize Solution: J(x) = (y Hx) T W (y Hx)
More informationf X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx
INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationKalman Filter. Lawrence J. Christiano
Kalman Filter Lawrence J. Christiano Background The Kalman filter is a powerful tool, which can be used in a variety of contexts. can be used for filtering and smoothing. To help make it concrete, we will
More informationGaussians. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics
Gaussians Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Outline Univariate Gaussian Multivariate Gaussian Law of Total Probability Conditioning
More informationSolutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π
Solutions to Homework Set #5 (Prepared by Lele Wang). Neural net. Let Y X + Z, where the signal X U[,] and noise Z N(,) are independent. (a) Find the function g(y) that minimizes MSE E [ (sgn(x) g(y))
More informationTSRT14: Sensor Fusion Lecture 8
TSRT14: Sensor Fusion Lecture 8 Particle filter theory Marginalized particle filter Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 8 Gustaf Hendeby Spring 2018 1 / 25 Le 8: particle filter theory,
More informationProblem Set 7 Due March, 22
EE16: Probability and Random Processes SP 07 Problem Set 7 Due March, Lecturer: Jean C. Walrand GSI: Daniel Preda, Assane Gueye Problem 7.1. Let u and v be independent, standard normal random variables
More informationGaussian random variables inr n
Gaussian vectors Lecture 5 Gaussian random variables inr n One-dimensional case One-dimensional Gaussian density with mean and standard deviation (called N, ): fx x exp. Proposition If X N,, then ax b
More informationENGR352 Problem Set 02
engr352/engr352p02 September 13, 2018) ENGR352 Problem Set 02 Transfer function of an estimator 1. Using Eq. (1.1.4-27) from the text, find the correct value of r ss (the result given in the text is incorrect).
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationECE531 Screencast 5.5: Bayesian Estimation for the Linear Gaussian Model
ECE53 Screencast 5.5: Bayesian Estimation for the Linear Gaussian Model D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III / 8 Bayesian Estimation
More informationMoment Generating Function. STAT/MTHE 353: 5 Moment Generating Functions and Multivariate Normal Distribution
Moment Generating Function STAT/MTHE 353: 5 Moment Generating Functions and Multivariate Normal Distribution T. Linder Queen s University Winter 07 Definition Let X (X,...,X n ) T be a random vector and
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationOpen Economy Macroeconomics: Theory, methods and applications
Open Economy Macroeconomics: Theory, methods and applications Lecture 4: The state space representation and the Kalman Filter Hernán D. Seoane UC3M January, 2016 Today s lecture State space representation
More informationMATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)
MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation) Last modified: March 7, 2009 Reference: PRP, Sections 3.6 and 3.7. 1. Tail-Sum Theorem
More information1 Kalman Filter Introduction
1 Kalman Filter Introduction You should first read Chapter 1 of Stochastic models, estimation, and control: Volume 1 by Peter S. Maybec (available here). 1.1 Explanation of Equations (1-3) and (1-4) Equation
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationRECURSIVE ESTIMATION AND KALMAN FILTERING
Chapter 3 RECURSIVE ESTIMATION AND KALMAN FILTERING 3. The Discrete Time Kalman Filter Consider the following estimation problem. Given the stochastic system with x k+ = Ax k + Gw k (3.) y k = Cx k + Hv
More informationA Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition
A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition Junfeng Wu, Karl Henrik Johansson and Ling Shi E-mail: jfwu@ust.hk Stockholm, 9th, January 2014 1 / 19 Outline Outline
More informationLecture 5: Control Over Lossy Networks
Lecture 5: Control Over Lossy Networks Yilin Mo July 2, 2015 1 Classical LQG Control The system: x k+1 = Ax k + Bu k + w k, y k = Cx k + v k x 0 N (0, Σ), w k N (0, Q), v k N (0, R). Information available
More information5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors
EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we
More informationStability of Modified-CS over Time for recursive causal sparse reconstruction
Stability of Modified-CS over Time for recursive causal sparse reconstruction Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/ namrata
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationChapter 4 : Expectation and Moments
ECE5: Analysis of Random Signals Fall 06 Chapter 4 : Expectation and Moments Dr. Salim El Rouayheb Scribe: Serge Kas Hanna, Lu Liu Expected Value of a Random Variable Definition. The expected or average
More informationAppendix A : Introduction to Probability and stochastic processes
A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of
More informationEE226a - Summary of Lecture 13 and 14 Kalman Filter: Convergence
1 EE226a - Summary of Lecture 13 and 14 Kalman Filter: Convergence Jean Walrand I. SUMMARY Here are the key ideas and results of this important topic. Section II reviews Kalman Filter. A system is observable
More informationMiscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier
Miscellaneous Regarding reading materials Reading materials will be provided as needed If no assigned reading, it means I think the material from class is sufficient Should be enough for you to do your
More informationReview (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology
Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna
More informationThe Kalman Filter ImPr Talk
The Kalman Filter ImPr Talk Ged Ridgway Centre for Medical Image Computing November, 2006 Outline What is the Kalman Filter? State Space Models Kalman Filter Overview Bayesian Updating of Estimates Kalman
More informationRecursive Estimation
Recursive Estimation Raffaello D Andrea Spring 08 Problem Set 3: Extracting Estimates from Probability Distributions Last updated: April 9, 08 Notes: Notation: Unless otherwise noted, x, y, and z denote
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More information(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.
54 We are given the marginal pdfs of Y and Y You should note that Y gamma(4, Y exponential( E(Y = 4, V (Y = 4, E(Y =, and V (Y = 4 (a With U = Y Y, we have E(U = E(Y Y = E(Y E(Y = 4 = (b Because Y and
More informationA Theoretical Overview on Kalman Filtering
A Theoretical Overview on Kalman Filtering Constantinos Mavroeidis Vanier College Presented to professors: IVANOV T. IVAN STAHN CHRISTIAN Email: cmavroeidis@gmail.com June 6, 208 Abstract Kalman filtering
More informationEcon 2120: Section 2
Econ 2120: Section 2 Part I - Linear Predictor Loose Ends Ashesh Rambachan Fall 2018 Outline Big Picture Matrix Version of the Linear Predictor and Least Squares Fit Linear Predictor Least Squares Omitted
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationA TERM PAPER REPORT ON KALMAN FILTER
A TERM PAPER REPORT ON KALMAN FILTER By B. Sivaprasad (Y8104059) CH. Venkata Karunya (Y8104066) Department of Electrical Engineering Indian Institute of Technology, Kanpur Kanpur-208016 SCALAR KALMAN FILTER
More informationUCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE 53 Handout #46 Prof. Young-Han Kim Thursday, June 5, 04 Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei). Discrete-time Wiener process. Let Z n, n 0 be a discrete time white
More informationNext tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2
Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution Defn: Z R 1 N(0,1) iff f Z (z) = 1 2π e z2 /2 Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) (a column
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationSUPPLEMENT TO TESTING FOR REGIME SWITCHING: A COMMENT (Econometrica, Vol. 80, No. 4, July 2012, )
Econometrica Supplementary Material SUPPLEMENT TO TESTING FOR REGIME SWITCHING: A COMMENT Econometrica, Vol. 80, No. 4, July 01, 1809 181 BY ANDREW V. CARTER ANDDOUGLAS G. STEIGERWALD We present proof
More informationFinal Examination Solutions (Total: 100 points)
Final Examination Solutions (Total: points) There are 4 problems, each problem with multiple parts, each worth 5 points. Make sure you answer all questions. Your answer should be as clear and readable
More informationHere represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities.
19 KALMAN FILTER 19.1 Introduction In the previous section, we derived the linear quadratic regulator as an optimal solution for the fullstate feedback control problem. The inherent assumption was that
More informationThe Kalman filter. Chapter 6
Chapter 6 The Kalman filter In the last chapter, we saw that in Data Assimilation, we ultimately desire knowledge of the full a posteriori p.d.f., that is the conditional p.d.f. of the state given the
More informationUCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE53 Handout #7 Prof. Young-Han Kim Tuesday, May 6, 4 Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei). Neural net. Let Y = X + Z, where the signal X U[,] and noise Z N(,) are independent.
More informationThe Multivariate Gaussian Distribution [DRAFT]
The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationKalman Filter. Man-Wai MAK
Kalman Filter Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: S. Gannot and A. Yeredor,
More informationCh. 5 Joint Probability Distributions and Random Samples
Ch. 5 Joint Probability Distributions and Random Samples 5. 1 Jointly Distributed Random Variables In chapters 3 and 4, we learned about probability distributions for a single random variable. However,
More informationExam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition)
Exam P Review Sheet log b (b x ) = x log b (y k ) = k log b (y) log b (y) = ln(y) ln(b) log b (yz) = log b (y) + log b (z) log b (y/z) = log b (y) log b (z) ln(e x ) = x e ln(y) = y for y > 0. d dx ax
More informationParticle Filtered Modified-CS (PaFiMoCS) for tracking signal sequences
Particle Filtered Modified-CS (PaFiMoCS) for tracking signal sequences Samarjit Das and Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/
More informationCovariance and Correlation
Covariance and Correlation ST 370 The probability distribution of a random variable gives complete information about its behavior, but its mean and variance are useful summaries. Similarly, the joint probability
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationCS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro
CS37300 Class Notes Jennifer Neville, Sebastian Moreno, Bruno Ribeiro 2 Background on Probability and Statistics These are basic definitions, concepts, and equations that should have been covered in your
More informationStability of LS-CS-residual and modified-cs for sparse signal sequence reconstruction
Stability of LS-CS-residual and modified-cs for sparse signal sequence reconstruction Namrata Vaswani ECE Dept., Iowa State University, Ames, IA 50011, USA, Email: namrata@iastate.edu Abstract In this
More informationProblem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51
Yi Lu Correlation and Covariance Yi Lu ECE 313 2/51 Definition Let X and Y be random variables with finite second moments. the correlation: E[XY ] Yi Lu ECE 313 3/51 Definition Let X and Y be random variables
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49
State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing
More informationLimited circulation. For review only
Event-based Sensor Data Scheduling: Trade-off Between Communication Rate and Estimation Quality Process Sensor Estimator Networ Junfeng Wu, Qing-Shan Jia, Karl Henri Johansson, Ling Shi Abstract We consider
More informationLecture Note 1: Probability Theory and Statistics
Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would
More informationMultivariate Distributions (Hogg Chapter Two)
Multivariate Distributions (Hogg Chapter Two) STAT 45-1: Mathematical Statistics I Fall Semester 15 Contents 1 Multivariate Distributions 1 11 Random Vectors 111 Two Discrete Random Variables 11 Two Continuous
More informationRecitation 2: Probability
Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions
More informationA Crash Course on Kalman Filtering
A Crash Course on Kalman Filtering Dan Simon Cleveland State University Fall 2014 1 / 64 Outline Linear Systems Probability State Means and Covariances Least Squares Estimation The Kalman Filter Unknown
More informationSIMON FRASER UNIVERSITY School of Engineering Science
SIMON FRASER UNIVERSITY School of Engineering Science Course Outline ENSC 810-3 Digital Signal Processing Calendar Description This course covers advanced digital signal processing techniques. The main
More information2D Image Processing. Bayes filter implementation: Kalman filter
2D Image Processing Bayes filter implementation: Kalman filter Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationECE Lecture #10 Overview
ECE 450 - Lecture #0 Overview Introduction to Random Vectors CDF, PDF Mean Vector, Covariance Matrix Jointly Gaussian RV s: vector form of pdf Introduction to Random (or Stochastic) Processes Definitions
More informationProbability Theory Review Reading Assignments
Probability Theory Review Reading Assignments R. Duda, P. Hart, and D. Stork, Pattern Classification, John-Wiley, 2nd edition, 2001 (appendix A.4, hard-copy). "Everything I need to know about Probability"
More informationKalman Filters. Derivation of Kalman Filter equations
Kalman Filters Derivation of Kalman Filter equations Mar Fiala CRV 10 Tutorial Day May 29/2010 Kalman Filters Predict x ˆ + = F xˆ 1 1 1 B u P F + 1 = F P 1 1 T Q Update S H + K ~ y P = H P 1 T T 1 = P
More informationLecture 17: Differential Entropy
Lecture 17: Differential Entropy Differential entropy AEP for differential entropy Quantization Maximum differential entropy Estimation counterpart of Fano s inequality Dr. Yao Xie, ECE587, Information
More informationLesson 4: Stationary stochastic processes
Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@univaq.it Stationary stochastic processes Stationarity is a rather intuitive concept, it means
More informationSTAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong
STAT 443 Final Exam Review L A TEXer: W Kong 1 Basic Definitions Definition 11 The time series {X t } with E[X 2 t ] < is said to be weakly stationary if: 1 µ X (t) = E[X t ] is independent of t 2 γ X
More informationUniversity of Illinois ECE 313: Final Exam Fall 2014
University of Illinois ECE 313: Final Exam Fall 2014 Monday, December 15, 2014, 7:00 p.m. 10:00 p.m. Sect. B, names A-O, 1013 ECE, names P-Z, 1015 ECE; Section C, names A-L, 1015 ECE; all others 112 Gregory
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationCOS Lecture 16 Autonomous Robot Navigation
COS 495 - Lecture 16 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization
More information