A Probability Review

Similar documents
EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

Multiple Random Variables

3. Probability and Statistics

Probability Background

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

Introduction to Probability and Stocastic Processes - Part I

Multivariate Random Variable

Multivariate Statistics

conditional cdf, conditional pdf, total probability theorem?

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors

BASICS OF PROBABILITY

5 Operations on Multiple Random Variables

Review: mostly probability and some statistics

The Multivariate Gaussian Distribution [DRAFT]

Let X and Y denote two random variables. The joint distribution of these random

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

Stat 5101 Notes: Algorithms

Lecture 11. Multivariate Normal theory

EE4601 Communication Systems

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Lecture Note 1: Probability Theory and Statistics

Stat 5101 Notes: Algorithms (thru 2nd midterm)

Continuous Random Variables

Chapter 5. The multivariate normal distribution. Probability Theory. Linear transformations. The mean vector and the covariance matrix

L2: Review of probability and statistics

ACM 116: Lectures 3 4

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Random Variables and Their Distributions

Section 8.1. Vector Notation

Elliptically Contoured Distributions

01 Probability Theory and Statistics Review

Multivariate Gaussians. Sargur Srihari

Chapter 5 continued. Chapter 5 sections

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Multiple Random Variables

REVIEW OF MAIN CONCEPTS AND FORMULAS A B = Ā B. Pr(A B C) = Pr(A) Pr(A B C) =Pr(A) Pr(B A) Pr(C A B)

18 Bivariate normal distribution I

Lecture 2. Spring Quarter Statistical Optics. Lecture 2. Characteristic Functions. Transformation of RVs. Sums of RVs

Recitation 2: Probability

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π

Basics on Probability. Jingrui He 09/11/2007

MAS223 Statistical Inference and Modelling Exercises

Quick Tour of Basic Probability Theory and Linear Algebra

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

ECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1

ME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 PROBABILITY. Prof. Steven Waslander

ECE 636: Systems identification

Recall that if X 1,...,X n are random variables with finite expectations, then. The X i can be continuous or discrete or of any other type.

The Delta Method and Applications

Lecture 2: Repetition of probability theory and statistics

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

The Multivariate Normal Distribution. In this case according to our theorem

Chapter 5,6 Multiple RandomVariables

Introduction to Probability Theory

Multivariate Distributions

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)

Introduction to Normal Distribution

Statistics, Data Analysis, and Simulation SS 2015

Lecture 2: Review of Probability

[POLS 8500] Review of Linear Algebra, Probability and Information Theory

Formulas for probability theory and linear models SF2941

2 Functions of random variables

Notes on Random Vectors and Multivariate Normal

Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

component risk analysis

Stat 206: Sampling theory, sample moments, mahalanobis

ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions

where r n = dn+1 x(t)

ECE 4400:693 - Information Theory

Probability and Distributions

Lecture 3: Pattern Classification

We introduce methods that are useful in:

The Gaussian distribution

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

Review (Probability & Linear Algebra)

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Bayes Decision Theory

4. Distributions of Functions of Random Variables

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 1

Lecture 2: Review of Basic Probability Theory

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Statistics for scientists and engineers

Gaussian vectors and central limit theorem

Variations. ECE 6540, Lecture 02 Multivariate Random Variables & Linear Algebra

Statistical Pattern Recognition

TAMS39 Lecture 2 Multivariate normal distribution

4. CONTINUOUS RANDOM VARIABLES

1 Review of Probability and Distributions

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

1 Random Variable: Topics

1 Introduction. Systems 2: Simulating Errors. Mobile Robot Systems. System Under. Environment

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Chapter 4. Multivariate Distributions. Obviously, the marginal distributions may be obtained easily from the joint distribution:

Transcription:

A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1

A Probability Review Reading: Go over handouts 2 5 in EE 420x notes Basic probability rules: (1) Pr{Ω} = 1, Pr{ } = 0, 0 Pr{A} 1; Pr{ i=1 A i} = i=1 Pr{A i} if A i A }{{} j = for all i j; A i and A j disjoint (2) Pr{A B} = Pr{A} + Pr{B} Pr{A B}, Pr{A c } = 1 Pr{A}; (3) If A B, then Pr{A B} = Pr{A} Pr{B}; (4) Pr{A B} = Pr{A B} Pr{B} (conditional probability) or Pr{A B} = Pr{A B} Pr{B} (chain rule); EE 527, Detection and Estimation Theory, # 0b 2

(5) Pr{A} = Pr{A B 1 } Pr{B 1 } + + Pr{A B n } Pr{B n } if B 1, B 2,, B n form a partition of the full space Ω; (6) Bayes rule: Pr{A B} = Pr{B A} Pr{A} Pr{B} EE 527, Detection and Estimation Theory, # 0b 3

Reminder: Independence, Correlation and Covariance For simplicity, we state all the definitions for pdfs; the corresponding definitions for pmfs are analogous Two random variables X and Y are independent if f X,Y (x, y) = f X (x) f Y (y) Correlation between real-valued random variables X and Y : E X,Y {X Y } = + + x y f X,Y (x, y) dx dy Covariance between real-valued random variables X and Y : cov X,Y (X, Y ) = E X,Y {(X µ X ) (Y µ Y )} = + + (x µ X ) (y µ Y ) f X,Y (x, y) dx dy EE 527, Detection and Estimation Theory, # 0b 4

where µ X = E X (X) = µ Y = E Y (Y ) = + + x f X (x) dx y f Y (y) dy Uncorrelated random variables: Random variables X and Y are uncorrelated if c X,Y = cov X,Y (X, Y ) = 0 (1) If X and Y are real-valued RVs, then (1) can be written as E X,Y {X Y } = E X {X} E Y {Y } Mean Vector and Covariance Matrix: Consider a random vector X = X 1 X 2 X N EE 527, Detection and Estimation Theory, # 0b 5

The mean of this random vector is defined as µ 1 E µ X = µ 2 = E [X X 1 1] X{X} = E [X X 2 2] µ N E XN [X N ] Denote the covariance between X i and X k, cov Xi,X k (X i, X k ), by c i,k ; hence, the variance of X i is c i,i = cov Xi,X k (X i, X i ) = var Xi (X i ) = σx 2 }{{} i The covariance matrix of X is defined as more notation C X = c 1,1 c 1,2 c 1,N c 2,1 c 2,2 c 2,N c N,1 c N,2 c N,N The above definitions apply to both real and complex vectors X Covariance matrix of a real-valued random vector X: C X = E X {(X E X [X]) (X E X [X]) T } = E X [X X T ] E X [X] (E X [X]) T For real-valued X, c i,k = c k,i and, therefore, C X is a symmetric matrix EE 527, Detection and Estimation Theory, # 0b 6

Linear Transform of Random Vectors Linear Transform For real-valued Y, X, A, Y = g(x) = A X Mean Vector: µ Y = E X {A X} = A µ X (2) Covariance Matrix: C Y = E Y {Y Y T } µ Y µ T Y = E X {A X X T A T } A µ X µ T X A T ( ) = A E X {X X T } µ X µ }{{ T X} C X A T = A C X A T (3) EE 527, Detection and Estimation Theory, # 0b 7

Reminder: Iterated Expectations In general, we can find E X,Y [g(x, Y )] using iterated expectations: E X,Y [g(x, Y )] = E Y {E X Y [g(x, Y ) Y ]} (4) where E X Y denotes the expectation with respect to f X Y (x y) and E Y denotes the expectation with respect to f Y (y) Proof E Y {E X Y [g(x, Y ) Y ]} = = = = + + + + E X Y [g(x, Y ) y] f Y (y) dy ( + ) g(x, y)f X Y (x y) dx f Y (y) dy + + = E X,Y [g(x, Y )] g(x, y)f X Y (x y)f Y (y) dx dy g(x, y) f X,Y (x, y) dx dy EE 527, Detection and Estimation Theory, # 0b 8

Reminder: Law of Conditional Variances Define the conditional variance of X given Y = y to be the variance of X with respect to f X Y (x y), ie ] var X Y (X Y = y) = E X Y [(X E X Y [X y]) 2 y = E X Y [X 2 y] (E X Y [X y]) 2 The random variable var X Y (X Y ) is a function of Y only, taking values var(x Y = y) Its expected value with respect to Y is E Y {var X Y (X Y )} = E Y { E X Y [X 2 Y ] (E X Y [X Y ]) 2} iterated exp = E X,Y [X 2 ] E Y {(E X Y [X Y ]) 2 } = E X [X 2 ] E Y {(E X Y [X Y ]) 2 } Since E X Y [X Y ] is a random variable (and a function of Y only), it has variance: var Y {E X Y [X Y ]} = E Y {(E X Y [X Y ]) 2 } (E Y {E X Y [X Y ]}) 2 iterated exp = E Y {(E X Y [X Y ]) 2 } (E X,Y [X]) 2 = E Y {E X Y [X Y ] 2 } (E X [X]) 2 EE 527, Detection and Estimation Theory, # 0b 9

Adding the above two expressions yields the law of conditional variances: E Y {var X Y (X Y )} + var Y {E X Y [X Y ]} = var X (X) (5) Note: (4) and (5) hold for both real- and complex-valued random variables EE 527, Detection and Estimation Theory, # 0b 10

Useful Expectation and Covariance Identities for Real-valued Random Variables and Vectors E X,Y [a X + b Y + c] = a E X [X] + b E Y [Y ] + c var X,Y (a X + b Y + c) = a 2 var X (X) + b 2 var Y (Y ) +2 a b cov X,Y (X, Y ) where a, b, and c are constants and X and Y are random variables A vector/matrix version of the above identities: E X,Y [A X + B Y + c] = A E X [X] + B E Y [Y ] + c cov X,Y (A X + B Y + c) = A cov X (X) A T + B cov Y (Y ) B T +A cov X,Y (X, Y ) B T + B cov X,Y (Y, X) A T where T denotes a transpose and cov X,Y (X, Y ) = E X,Y {(X E X [X]) (Y E Y [Y ]) T } Useful properties of crosscovariance matrices: cov X,Y,Z (X, Y + Z) = cov X,Y (X, Y ) + cov X,Z (X, Z) EE 527, Detection and Estimation Theory, # 0b 11

cov Y,X (Y, X) = [cov X,Y (X, Y )] T cov X (X) = cov X (X, X) var X (X) = cov X (X, X) cov X,Y (A X + b, P Y + q) = A cov X,Y (X, Y ) P T (To refresh memory about covariance and its properties, see p 12 of handout 5 in EE 420x notes For random vectors, see handout 7 in EE 420x notes, particularly pp 1 15) Useful theorems: (1) (handout 5 in EE 420x notes) E X (X) = E Y [E X Y (X Y )] shown on p 8 E X Y [g(x) h(y ) y] = h(y) E X Y [g(x) y] E X,Y [g(x) h(y )] = E Y {h(y ) E X Y [g(x) Y ]} The vector version of (1) is the same just put bold letters EE 527, Detection and Estimation Theory, # 0b 12

(2) var X (X) = E Y [var X Y (X Y )] + var Y (E X Y [X Y ]); and the vector/matrix version is cov X (X) }{{} variance/covariance matrix of X = E Y [cov X Y (X Y )] +cov Y (E X Y [X Y ]) shown on p (3) Generalized law of conditional variances: cov X,Y (X, Y ) = E Z [cov X,Y Z (X, Y Z)] +cov Z (E X Z [X Z], E Y Z [Y Z]) (4) Transformation: Y = g(x) one-to-one Y 1 = g 1 (X 1,, X n ) Y n = g n (X 1,, X n ) then f Y (y) = f X (h 1 (y 1 ),, h n (y n )) J EE 527, Detection and Estimation Theory, # 0b 13

where h( ) is the unique inverse of g( ) and J = x y T = x 1 x 1 y n y 1 x n y 1 x n y n Print and read the handout Probability distributions from the Course readings section on WebCT Bring it with you to the midterm exams EE 527, Detection and Estimation Theory, # 0b 14

Jointly Gaussian Real-valued RVs Scalar Gaussian random variables: f X (x) = 1 2 π σ 2 X exp [ (x µ X) 2 2 σ 2 X ] Definition Two real-valued RVs X and Y are jointly Gaussian if their joint pdf is of the form f X,Y (x, y) = 1 exp { 2πσ X σ Y 1 ρ 2 X,Y 1 2 (1 ρ 2 X,Y ) (x µ X ) (y µ Y ) ] 2 ρ X,Y σ X σ Y [ (x µx ) 2 } σ 2 X + (y µ Y ) 2 σ 2 Y (6) This pdf is parameterized by µ X, µ Y, σ 2 X, σ 2 Y, and ρ X,Y Here, σ X = σ 2 X and σ Y = σ 2 Y Note: We will soon define a more general multivariate Gaussian pdf EE 527, Detection and Estimation Theory, # 0b 15

If X and Y are jointly Gaussian, contours of equal joint pdf are ellipses defined by the quadratic equation (x µ X ) 2 σ 2 X + (y µ Y ) 2 (x µ X ) (y µ Y ) σ 2 2 ρ X,Y = const 0 Y σ X σ Y Examples: In the following examples, we plot contours of the joint pdf f X,Y (x, y) for zero-mean jointly Gaussian RVs for various values of σ X, σ Y, and ρ X,Y EE 527, Detection and Estimation Theory, # 0b 16

EE 527, Detection and Estimation Theory, # 0b 17

If X and Y Gaussian, eg are jointly Gaussian, the conditional pdfs are ( X {Y = y} N ρ X,Y σ X y E Y [Y ] σ Y +E X [X], (1 ρ 2 X,Y ) σ 2 X If X and Y are jointly Gaussian and uncorrelated, ie ρ X,Y = 0, they are also independent (7) ) EE 527, Detection and Estimation Theory, # 0b 18

Gaussian Random Vectors Real-valued Gaussian random vectors: f X (x) = 1 (2 π) N/2 C X 1/2 exp [ 1 2 (x µ X) T C 1 X (x µ X ) ] Complex-valued Gaussian random vectors: f Z (z) = 1 π N C Z exp [ (z µ Z ) H C 1 Z (z µ Z ) ] Notation for real- and complex-valued Gaussian random vectors: X N r (µ X, C X ) [or simply N (µ X, C X )] X N c (µ X, C X ) complex real An affine transform of a Gaussian vector is also a Gaussian random vector, ie if then Y = A X + b Y N r (A µ X + b, A C X A T ) Y N c (A µ X + b, A C X A H ) real complex EE 527, Detection and Estimation Theory, # 0b 19

The Gaussian random vector W N r (0, σ 2 I n ) (where I n denotes the identity matrix of size n) is called white; pdf contours of a white Gaussian random vector are spheres centered at the origin Suppose that W [n], n = 0, 1,, N 1 are independent, identically distributed (iid) zeromean univariate Gaussian N (0, σ 2 ) Then, for W = [W [0], W [1],, W [N 1]] T, f W (w) = N (w 0, σ 2 I) Suppose now that, for these W [n], Y [n] = θ + W [n] where θ is a constant What is the joint pdf of Y [0], Y [1],, and Y [N 1]? This pdf is the pdf of the vector Y = [Y [0], Y [1],, Y [N 1]] T : Y = 1 θ + W where 1 is an N 1 vector of ones Now, Since θ is a constant, f Y (y) = N (y 1 θ, σ 2 I) f Y (y) = f Y θ (y θ) EE 527, Detection and Estimation Theory, # 0b 20

Gaussian Random Vectors A real-valued random vector X = [X 1, X 2,, X n ] T with mean µ and covariance matrix Σ with determinant Σ > 0 (ie Σ is positive definite) is a Gaussian random vector (or X 1, X 2,, X n Gaussian RVs) if and only if its joint pdf is are jointly f X (x) = 1 2πΣ 1/2 exp[ 1 2 (x µ)t Σ 1 (x µ)] (8) Verify that, for n = 2, this joint pdf reduces to the twodimensional pdf in (6) Notation: We use X N (µ, Σ) to denote a Gaussian random vector Since Σ is positive definite, Σ 1 is also positive definite and, for x µ, (x µ) T Σ 1 (x µ) > 0 which means that the contours of the multivariate Gaussian pdf in (8) are ellipsoids The Gaussian random vector X N (0, σ 2 I n ) (where I n denotes the identity matrix of size n) is called white contours of the pdf of a white Gaussian random vector are spheres centered at the origin EE 527, Detection and Estimation Theory, # 0b 21

Properties of Real-valued Gaussian Random Vectors Property 1: For a Gaussian random vector, uncorrelation implies independence This is easy to verify by setting Σ i,j = 0 for all i j in the joint pdf, then Σ becomes diagonal and so does Σ 1 ; then, the joint pdf reduces to the product of marginal pdfs f Xi (x i ) = N (µ i, Σ i,i ) = N (µ i, σ 2 ) Clearly, this property Xi holds for blocks of RVs (subvectors) as well Property 2: A linear transform of a Gaussian random vector X N (µ X, Σ X ) yields a Gaussian random vector: Y = A X N (A µ X, A Σ X A T ) It is easy to show that E Y [Y ] = A µ X and cov Y (Y ) = Σ Y = A Σ X A T So E Y [Y ] = E X [A X] = A E X [X] = A µ X and Σ Y = E Y [(Y E Y [Y ]) (Y E Y [Y ]) T ] = E X [(A X A µ X ) (A X A µ X ) T ] = A E X [(X µ X ) (X µ X ) T ] A T = A Σ X A T EE 527, Detection and Estimation Theory, # 0b 22

Of course, if we use the definition of a Gaussian random vector in (8), we cannot yet claim that Y is a Gaussian random vector (For a different definition of a Gaussian random vector, we would be done right here) Proof We need to verify that the joint pdf of Y indeed has the right form Here, we decide to take the equivalent (easier) task and verify that the characteristic function of Y has the right form Definition Suppose X f X (X) Then the characteristic function of X is given by Φ X (ω) = E X [exp(j ω T X)] where ω is an n-dimensional real-valued vector and j = 1 Thus Φ X (ω) = + + f X (x) exp(j ω T x) dx proportional to the inverse multi-dimensional Fourier transform of f X (x); therefore, we can find f X (x) by taking the Fourier transform (with the appropriate proportionality factor): f X (x) = 1 (2π) n + + Φ X (ω) exp( j ω T x) dx EE 527, Detection and Estimation Theory, # 0b 23

Example: The characteristic function for X N (µ, σ 2 ) is given by Φ X (ω) = exp( 1 2 ω2 σ 2 + j µ ω) (9) and for a Gaussian random vector Z N (µ, Σ), Φ Z (ω) = exp( 1 2 ωt Σω + j ω T µ) (10) Now, go back to our proof: Y = AX is the characteristic function of Φ Y (ω) = E Y [exp(j ω T Y )] = E X [exp(j ω T A X)] = exp( 1 2 ωt A Σ X A T ω + j ω T A µ X ) Thus Y = A X N (A µ X, AΣ X A T ) EE 527, Detection and Estimation Theory, # 0b 24

Property 3: Marginals of a Gaussian random vector are Gaussian, ie if X is a Gaussian random vector, then, for any {i 1, i 2,, i k } {1, 2,, n}, Y = X i1 X i2 X ik is a Gaussian random vector To show this, [ we use ] Property 2 X1 Here is an example with n = 3 and Y = We set X 3 Y = [ 1 0 0 0 0 1 ] X 1 X 2 X 3 thus Here and Y N ( [ ] [ ] ) µ 1 Σ1,1 Σ, 1,3 µ 3 Σ 3,1 Σ 3,3 E X X 1 X 2 = µ 1 µ 2 X 3 µ 3 cov X X 1 X 2 = Σ 1,1 Σ 1,2 Σ 1,3 Σ 2,1 Σ 2,2 Σ 2,3 X 3 Σ 3,1 Σ 3,2 Σ 3,3 EE 527, Detection and Estimation Theory, # 0b 25

and note that [ µ1 µ 3 ] = [ 1 0 0 0 0 1 ] µ 1 µ 2 µ 3 and [ ] Σ1,1 Σ 1,3 Σ 3,1 Σ 3,3 = [ 1 0 0 0 0 1 ] Σ 1,1 Σ 1,2 Σ 1,3 Σ 2,1 Σ 2,2 Σ 2,3 Σ 3,1 Σ 3,2 Σ 3,3 1 0 0 0 0 1 The converse of Property 3 does not hold in general; here is a counterexample: Example: Suppose X 1 N (0, 1) and X 2 = { 1, wp 1 2 1, wp 1 2 are independent RVs and consider X 3 = X 1 X 2 Observe that X 3 N (0, 1) and f X 1,X3 (x 1, x 3 ) is not a jointly Gaussian pdf EE 527, Detection and Estimation Theory, # 0b 26

Property 4: Conditionals of Gaussian random vectors are Gaussian, ie if then X = [ X1 X 2 ] {X 2 X 1 = x 1 } N and {X 1 X 2 = x 2 } N ( [ ] [ ] ) µ N 1 Σ1,1 Σ, 1,2 µ 2 Σ 2,1 Σ 2,2 ( ) Σ 2,1 Σ1,1 1 (x 1 µ 1 )+µ 2, Σ 2,2 Σ 2,1 Σ1,1 1 Σ 1,2 ( ) Σ 1,2 Σ2,2 1 (x 2 µ 2 )+µ 1, Σ 1,1 Σ 1,2 Σ2,2 1 Σ 2,1 EE 527, Detection and Estimation Theory, # 0b 27

Example: Compare this result to the case of n = 2 in (7): {X 2 X 1 = x 1 } N ( Σ2,1 (x 1 µ 1 ) + µ 2, Σ 2,2 Σ 2 ) 1,2 Σ 1,1 Σ 1,1 In particular, having X = X 2 and Y = X 1, y = x 1, this result becomes: {X Y = y} N ( σx,y σ 2 Y (y µ Y ) + µ X, σ 2 X σ2 X,Y σ 2 Y ) where σ X,Y = cov X,Y (X, Y ), σ 2 X = cov X,X (X, X) = var X (X), and σ 2 Y = cov Y,Y (Y, Y ) = var Y (Y ) Now, it is clear that ρ X,Y = σ X,Y σ X σ Y where σ X = σ 2 X > 0 and σ Y = σ 2 Y > 0 Example: Suppose EE 527, Detection and Estimation Theory, # 0b 28

Property 5: If X N (µ, Σ) then (x µ) T Σ 1 (x µ) χ 2 d (Chi-square in your distr table) EE 527, Detection and Estimation Theory, # 0b 29

Additive Gaussian Noise Channel Consider a communication channel with input X N (µ X, τ 2 X) and noise W N (0, σ 2 ) where X and W are independent and the measurement Y is Y = X + W Since X and W are independent, we have f X,W (x, w) = f X (x) f W (x) and [ X W ] ( [ µ N X 0 What is f Y X (y x)? Since ], [ τ 2 X 0 0 σ 2 ] ) {Y X = x} = x + W N (x, σ 2 ) we have f Y X (y x) = N (y x, σ 2 ) EE 527, Detection and Estimation Theory, # 0b 30

[ X Y How about f Y (y)? Construct the joint pdf f X,Y (x, y) of X and Y : since [ ] [ ] [ ] X 1 0 X = Y 1 1 W then ] N ( [ 1 0 1 1 ] [ µx 0 ], [ 1 0 1 1 ] [ ] [ τ 2 X 0 1 1 0 σ 2 0 1 ] ) yielding [ X Y ] N ( [ ] [ ] µ X τ 2 X τ, 2 ) X µ X τ 2 X τ 2 X + σ 2 Therefore, f Y (y) = N ( y µx, τ 2 X + σ 2) EE 527, Detection and Estimation Theory, # 0b 31

Complex Gaussian Distribution Consider joint pdf of real and imaginary part of an n 1 complex vector Z Z = U + j V Assume X = [ U Y ] The 2 n-variate Gaussian pdf of the (real!) vector X is f X (x) = 1 (2 π) 2 n Σ X exp [ 1 2 (z µ X) T Σ 1 X (z µ X ) ] where µ X = [ µu ], Σ µ X = V [ ] ΣUU Σ UV Σ V U Σ V V Therefore, Pr{x A} = x A f X (x) dx EE 527, Detection and Estimation Theory, # 0b 32

Complex Gaussian Distribution (cont) Assume that Σ X has a special structure: Σ UU = Σ V V and Σ UV = Σ V U [Note that Σ UV = Σ T V U has to hold as well] Then, we can define a complex Gaussian pdf of Z as f Z (z) = 1 π n Σ X exp [ (z µ Z ) H Σ 1 Z (z µ Z ) ] where µ Z = µ U + j µ V Σ Z = E Z {(Z µ Z ) (Z µ Z ) H } = 2 (Σ UU + j Σ V U ) 0 = E Z {(Z µ Z ) (Z µ Z ) T } EE 527, Detection and Estimation Theory, # 0b 33