Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Similar documents
Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)

Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver

Basics on 2-D 2 D Random Signal

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

Review of Discrete-Time System

Statistical signal processing

Part III Spectrum Estimation

Signals and Spectra - Review

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process

Stochastic Processes. A stochastic process is a function of two variables:

Advanced Digital Signal Processing -Introduction

ECE 636: Systems identification

CCNY. BME I5100: Biomedical Signal Processing. Stochastic Processes. Lucas C. Parra Biomedical Engineering Department City College of New York

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied

Communication Theory II

Practical Spectral Estimation

Random signals II. ÚPGM FIT VUT Brno,

Lesson 1. Optimal signalbehandling LTH. September Statistical Digital Signal Processing and Modeling, Hayes, M:

Chapter 6. Random Processes

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c =

Stochastic Process II Dr.-Ing. Sudchai Boonto

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University

16.584: Random (Stochastic) Processes

Random Processes Why we Care

Introduction to Probability and Stochastic Processes I

EEM 409. Random Signals. Problem Set-2: (Power Spectral Density, LTI Systems with Random Inputs) Problem 1: Problem 2:

for valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I

SRI VIDYA COLLEGE OF ENGINEERING AND TECHNOLOGY UNIT 3 RANDOM PROCESS TWO MARK QUESTIONS

14 - Gaussian Stochastic Processes

Some Time-Series Models

13. Power Spectrum. For a deterministic signal x(t), the spectrum is well defined: If represents its Fourier transform, i.e., if.

EAS 305 Random Processes Viewgraph 1 of 10. Random Processes

Stochastic Processes

Discrete time processes

Review of Probability

Massachusetts Institute of Technology

EE482: Digital Signal Processing Applications

ECE534, Spring 2018: Solutions for Problem Set #5

CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME

Linear Stochastic Models. Special Types of Random Processes: AR, MA, and ARMA. Digital Signal Processing

Probability and Statistics

Gaussian, Markov and stationary processes

Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes

Lecture - 30 Stationary Processes

5 Kalman filters. 5.1 Scalar Kalman filter. Unit delay Signal model. System model

Introduction to Stochastic processes

Probability and Statistics for Final Year Engineering Students

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

ECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else

Chapter 2 Random Processes

Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi)

If we want to analyze experimental or simulated data we might encounter the following tasks:

Module 4. Signal Representation and Baseband Processing. Version 2 ECE IIT, Kharagpur

ECON 616: Lecture 1: Time Series Basics

6. Methods for Rational Spectra It is assumed that signals have rational spectra m k= m

where r n = dn+1 x(t)

Final Exam January 31, Solutions

Chapter 7: The z-transform

Test 2 Electrical Engineering Bachelor Module 8 Signal Processing and Communications

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Class 1: Stationary Time Series Analysis

Multi-rate Signal Processing 3. The Polyphase Representation

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011

P 1.5 X 4.5 / X 2 and (iii) The smallest value of n for

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

Reliability Theory of Dynamically Loaded Structures (cont.)

Detection & Estimation Lecture 1

System Identification, Lecture 4

Multi-rate Signal Processing 7. M-channel Maximally Decmiated Filter Banks

Digital Signal Processing Lecture 3 - Discrete-Time Systems

Chapter 3 - Temporal processes

EE538 Final Exam Fall 2007 Mon, Dec 10, 8-10 am RHPH 127 Dec. 10, Cover Sheet

Econometría 2: Análisis de series de Tiempo

Statistics of stochastic processes

1 Linear Difference Equations

Time Series: Theory and Methods

Empirical Macroeconomics

Statistical and Adaptive Signal Processing

5 Analog carrier modulation with noise

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution

The Cooper Union Department of Electrical Engineering ECE111 Signal Processing & Systems Analysis Final May 4, 2012

EEG- Signal Processing

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

Lecture 7 MIMO Communica2ons

Lecture 4 - Spectral Estimation

Detection & Estimation Lecture 1

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Chapter 6: Random Processes 1

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Chapter 2 Wiener Filtering

Empirical Macroeconomics

STOCHASTIC PROBABILITY THEORY PROCESSES. Universities Press. Y Mallikarjuna Reddy EDITION

System Identification, Lecture 4

Statistics of Stochastic Processes

ECE Homework Set 3

EE303: Communication Systems

Transcription:

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted from ENEE630 slides developed by Profs. K.J. Ray Liu and Min Wu at the University of Maryland, College Park. Contact: chauwai.wong@ncsu.edu. Updated: January 16, 2018. ECE792-41 Lecture Part I 1 / 40

Outline of Part I 1. Discrete-time Stochastic Processes 2. Discrete Wiener Filtering 3. Linear Prediction ECE792-41 Lecture Part I 2 / 40

Outline of Section 1 Basic properties and characterization 1st and 2nd moment functions; ergodicity correlation matrix; power-spectrum density The Rational Transfer Function Model ARMA, AR, MA processes Wold Decomposition Theorem ARMA, AR, and MA models and properties asymptotic stationarity of AR process Readings for 1.1: Haykin 4th Ed. 1.1-1.3, 1.12, 1.14; see also Hayes 3.3, 3.4, and background reviews 2.2, 2.3, 3.2 ECE792-41 Lecture Part I 3 / 40

Stochastic/Random Processes To describe the time evolution of a statistical phenomenon according to probabilistic laws. Example random processes: speech signals, image, noise, temperature and other spatial/temporal measurements, etc. Discrete-time Stochastic Process {u[n]} Focus on the stochastic process that is defined / observed at discrete and uniformly spaced instants of time View it as a sequence of random variables (r.v.s) that are related in some statistical way:..., u[n M],..., u[n], u[n + 1],... A random process is not just a single function of time; it has an infinite number of different realizations ECE792-41 Lecture Part I 4 / 40

Parametric Signal Modeling A general way to completely characterize a random process is by joint probability density function (pdf) of all possible subsets of the r.v.s in it: Probability of {u[n 1 ], u[n 2 ],..., u[n k ]} Question: process? How to use only a few parameters to describe a Determine (1) a model, and then (2) the model parameters This part of the course studies the signal modeling (including models, applicable conditions, how to determine the parameters, etc) ECE792-41 Lecture Part I 5 / 40

(1) Partial Characterization by 1st and 2nd moments It is often difficult to determine and efficiently describe the joint pdf for a general random process. As a compromise, we consider partial characterization of the process by specifying its 1st and 2nd moments. Consider a stochastic time series {u[n]}, where u[n], u[n 1],... may be complex valued. We define the following functions: mean-value function: m[n] = E [u[n]], n Z autocorrelation function: r(n, n k) = E [u[n]u [n k]] autocovariance function: c(n, n k) = E [(u[n] m[n])(u[n k] m[n k]) ] Without loss of generality, we often consider zero-mean random process E [u[n]] = 0 n, since we can always subtract the mean in preprocessing. Now the autocorrelation and autocovariance functions become identical. ECE792-41 Lecture Part I 6 / 40

Wide-Sense Stationary (w.s.s.) Wide-Sense Stationarity If n, m[n] = m and r(n, n k) = r(k) (or c(n, n k) = c(k)), then the sequence u[n] is said to be wide-sense stationary (w.s.s.), or also called stationary to the second order. The strict stationarity requires the entire statistical property (characterized by joint probability density or mass function) to be invariant to time shifts. The partial characterization using 1st and 2nd moments offers two important advantages: 1 reflect practical measurements; 2 well suited for linear operations of random processes. ECE792-41 Lecture Part I 7 / 40

(2) Ensemble Average vs. Time Average Statistical expectation E( ) as an ensemble average: take average across (different realizations of) the process, or over the sample space. Time average: take average along the process or the time. This is what we can rather easily measure from one realization of the random process. Question: Answer: Are these two average types the same? No. (Examples from a random processes class) Consider two special cases of correlations between signal samples: 1 u[n], u[n 1], i.i.d. 2 u[n] = u[n 1] = all sample points are exact copies ECE792-41 Lecture Part I 8 / 40

Mean Ergodicity For a w.s.s. process, we may use the time average to estimate the mean m. ˆm(N) = 1 N N 1 n=0 u[n] ˆm(N) is an unbiased estimator of the mean of the process. E [ ˆm(N)] = m N. Question: How much does ˆm(N) from one observation deviate from the true mean? Mean Ergodic A w.s.s. process {u[n]} is mean ergodic if ˆm(N) m.s. m, i.e., lim N E [ ˆm(N) m 2] = 0. ECE792-41 Lecture Part I 9 / 40

Mean Ergodicity Question: Under what condition will mean ergodic be satisfied? (Details) E [ ˆm(N) m 2] = E [( ˆm(N) m)( ˆm(N) m) ] = (nece.& suff.) lim N 1 lim N 1 N N 1 l=0 N c(l) = 0. N 1 l= N+1 l (1 N )c(l) = 0, or A sufficient condition for mean ergodicity is that {u[n]} is asymptotically uncorrelated, i.e., lim l c(l) = 0. ECE792-41 Lecture Part I 10 / 40

Correlation Ergodicity Similarly, let an autocorrelation estimator be ˆr(k, N) = 1 N N 1 n=0 u[n]u [n k] The w.s.s. process {u[n]} is said to be correlation ergodic if the mean squared difference between r(k) and ˆr(k, N) approaches zero as N. ECE792-41 Lecture Part I 11 / 40

(3) Correlation Matrix Given an observation vector u[n] of a w.s.s. process, the correlation matrix R is defined as R E [ u[n]u H [n] ] where H denotes Hermitian transposition (i.e., conjugate transpose). u[n] Each entry in R is u[n 1] u[n]., [R] i,j = E [u[n i]u [n j]] = r(j i) (0 i, j M 1) u[n M + 1] r(0) r(1) r(m 1) r( 1) r(0) r(1). Thus R =......... r( M + 2) r(0) r(1) r( M + 1) r(0) ECE792-41 Lecture Part I 12 / 40

Properties of R 1 R is Hermitian, i.e., R H = R Proof (Details) 2 R is Toeplitz. A matrix is said to be Toeplitz if all elements in the main diagonal are identical, and the elements in any other diagonal parallel to the main diagonal are identical. R Toeplitz the w.s.s. property. ECE792-41 Lecture Part I 13 / 40

Properties of R 3 R is non-negative definite, i.e., x H Rx 0, x Proof (Details) eigenvalues of a Hermitian matrix are real. (similar relation in FT: real in one domain conjugate symmetric in the other) eigenvalues of a non-negative definite matrix are non-negative. Proof (Details) ECE792-41 Lecture Part I 14 / 40

Properties of R u[n M + 1] 4 u B [n]., i.e., reversely ordering u[n], u[n 1] u[n] then the corresponding correlation matrix becomes r(0) r( 1) r( M + 1) E [ u B [n](u B [n]) H] = r(1) r(0)........ = RT r(m 1) r(0) ECE792-41 Lecture Part I 15 / 40

Properties of R 5 Recursive relations: correlation matrix for (M + 1) 1 u[n]: (Details) ECE792-41 Lecture Part I 16 / 40

(4) Example-1: Complex Sinusoidal Signal x[n] = A exp [j(2πf 0 n + φ)] where A and f 0 are real constant, φ uniform distribution over [0, 2π) (i.e., random phase) E [x[n]] =? E [x[n]x [n k]] =? Is x[n] is w.s.s.? ECE792-41 Lecture Part I 17 / 40

Example-2: Complex Sinusoidal Signal with Noise Let y[n] = x[n] + w[n] where w[n] is white Gaussian noise uncorrelated to x[n], w[n] N(0, σ 2 ) { Note: for white noise, E [w[n]w σ 2 k = 0 [n k]] = 0 o.w. r y (k) = E [y[n]y [n k]] =? R y =? Rank of Correlation Matrices R x, R w, R y =? ECE792-41 Lecture Part I 18 / 40

(5) Power Spectral Density (a.k.a. Power Spectrum) Power spectral density (p.s.d.) of a w.s.s. process {x[n]} P X (ω) DTFT[r x (k)] = k= r x (k) DTFT 1 [PX (ω)] = 1 2π r x (k)e jωk π π P X (ω)e jωk dω The p.s.d. provides frequency domain description of the 2nd-order moment of the process (may also be defined as a function of f : ω = 2πf ) The power spectrum in terms of ZT: P X (z) = ZT[r x (k)] = k= r x(k)z k Physical meaning of p.s.d.: describes how the signal power of a random process is distributed as a function of frequency. ECE792-41 Lecture Part I 19 / 40

Properties of Power Spectral Density r x (k) is conjugate symmetric: r x (k) = r x ( k) P X (ω) is real valued: P X (ω) = P X (ω); P X (z) = P X (1/z ) For real-valued random process: r x (k) is real-valued and even symmetric P X (ω) is real and even symmetric, i.e., P X (ω) = P X ( ω); P X (z) = P X (z ) For w.s.s. process, P X (ω) 0 (nonnegative) The power of a zero-mean w.s.s. random process is proportional to the area under the p.s.d. curve over one period 2π, i.e., E [ x[n] 2] = r x (0) = 1 2π 2π P 0 X (ω)dω Proof: note r x (0) = IDTFT of P X (ω) at k = 0 ECE792-41 Lecture Part I 20 / 40

(6) Filtering a Random Process (Details) ECE792-41 Lecture Part I 21 / 40

Filtering a Random Process ECE792-41 Lecture Part I 22 / 40

Filtering a Random Process In terms of ZT: P Y (z) = P X (z)h(z)h (1/z ) P Y (ω) = P X (ω)h(ω)h (ω) = P X (ω) H(ω) 2 When h[n] is real, H (z ) = H(z) P Y (z) = P X (z)h(z)h(1/z) ECE792-41 Lecture Part I 23 / 40

Interpretation of p.s.d. If we choose H(z) to be an ideal bandpass filter with very narrow bandwidth around any ω 0, and measure the output power: E [ y[n] 2] = r y (0) = 1 = 1 2π 2π +π π P Y (ω)dω +π π P X (ω) H(ω) 2 dω = 1 2π. = 1 2π P X (ω 0 ) B 0 ω0 +B/2 ω 0 B/2 P X (ω) 1 dω P X (ω 0 ). = E [ y[n] 2] 2π B, and P X (ω) 0 ω i.e., p.s.d. is non-negative, and can be measured via power of {y[n]}. P X (ω) can be viewed as a density function describing how the power in x[n] varies with frequency. The above BPF operation also provides a way to measure it by BPF. ECE792-41 Lecture Part I 24 / 40

Summary: Review of Discrete-Time Random Process 1 An ensemble of sequences, where each outcome of the sample space corresponds to a discrete-time sequence 2 A general and complete way to characterize a random process: through joint p.d.f. 3 w.s.s process: can be characterized by 1st and 2nd moments (mean, autocorrelation) These moments are ensemble averages; E [x[n]], r(k) = E [x[n]x [n k]] Time average is easier to estimate (from just 1 observed sequence) Mean ergodicity and autocorrelation ergodicity: correlation function should be asymptotically decay, i.e., uncorrelated between samples that are far apart. the time average over large number of samples converges to the ensemble average in mean-square sense. ECE792-41 Lecture Part I 25 / 40

Characterization of w.s.s. Process through Correlation Matrix and p.s.d. 1 Define a vector on signal samples (note the indexing order): u[n] = [u(n), u(n 1),..., u(n M + 1)] T 2 Take expectation on the outer product: r(0) r(1) r(m 1) R E [ u[n]u H [n] ] = r( 1) r(0) r(1)......... r( M + 1) r(0) 3 Correlation function of w.s.s. process is a one-variable deterministic sequence take DTFT(r[k]) to get p.s.d. We can take DTFT on one sequence from the sample space of random process; different outcomes of the process will give different DTFT results; p.s.d. describes the statistical power distribution of the random process in spectrum domain. ECE792-41 Lecture Part I 26 / 40

Properties of Correlation Matrix and p.s.d. 4 Properties of correlation matrix: Toeplitz (by w.s.s.) Hermitian (by conjugate symmetry of r[k]); non-negative definite Note: if we reversely order the sample vector, the corresponding correlation matrix will be transposed. This is the convention used in Hayes book (i.e. the sample is ordered from n M + 1 to n), while Haykin s book uses ordering of n, n 1,... to n M + 1. 5 Properties of p.s.d.: real-valued (by conjugate symmetry of correlation function); non-negative (by non-negative definiteness of R matrix) ECE792-41 Lecture Part I 27 / 40

Filtering a Random Process 1 Each specific realization of the random process is just a discrete-time signal that can be filtered in the way we ve studied in undergrad DSP. 2 The ensemble of the filtering output is a random process. What can we say about the properties of this random process given the input process and the filter? 3 The results will help us further study such an important class of random processes that are generated by filtering a noise process by discrete-time linear filter with rational transfer function. Many discrete-time random processes encountered in practice can be well approximated by such a rational transfer function model: ARMA, AR, MA (see II.1.2) ECE792-41 Lecture Part I 28 / 40