UCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)

Size: px
Start display at page:

Download "UCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)"

Transcription

1 UCSD ECE 53 Handout #46 Prof. Young-Han Kim Thursday, June 5, 04 Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei). Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process, i.e., Z,Z,... are i.i.d. N(0,). Define the process X n, n as X 0 = 0, and X n = X n +Z n for n. (a) Is X n an independent increment process? Justify your answer. (b) Is X n a Markov process? Justify your answer. (c) Is X n a Gaussian process? Justify your answer. (d) Find the mean and autocorrelation functions of X n. (e) Specify the first and second order pdfs of X n. (f) Specify the joint pdf of X,X, and X 3. (g) Find E(X 0 X,X,...,X 0 ). Solution: (a) Yes. The increments X n, X n X n,...,x nk X nk are sums of different Z i random variables, and the Z i are IID. (b) Yes. Since the process has independent increments, it is Markov. (c) Yes. Any set of samples of X n, n are obtained by a linear transformation of IID N(0,) random variables and therefore all nth order distributions of X n are jointly Gaussian (it is not sufficient to show that the random variable X n is Gaussian for each n). (d) We have [ ] E[X n ] = E Z i = E[Z i ] = R X (n,n ) = E[X n X n ] n = E Z i n j= Z j n n = E[Z i Z j ] = j= min(n,n ) = min(n,n ). 0 = 0, E(Z i) (IID)

2 (e) As shown above, X n is Gaussian with mean zero and variance Var(X n ) = E[Xn] E [X n ] = R X (n,n) 0 = n. Thus, X n N(0,n). Cov(X n,x n ) = E(X n X n ) E(X n )E(X n ) = min(n,n ). Therefore, X n and X n are ( jointly Gaussian random variables ) with mean µ = [0 0] T n and covariance matrix Σ = min(n,n ). min(n,n ) n (f) X n, n is a zero mean Gaussian random process. Thus X E[X ] R X (,) R X (,) R X (,3) X N E[X ], R X (,) R X (,) R X (,3) X 3 E[X 3 ] R X (3,) R X (3,) R X (3,3) Substituting, we get X 0 X N 0,. X (g) Since X n is an independent increment process, E(X 0 X,X,...,X 0 ) = E(X 0 X 0 +X 0 X,X,...,X 0 ) = E(X 0 X 0 X,X,...,X 0 )+E(X 0 X,X,...,X 0 ) = E(X 0 X 0 )+X 0 = 0+X 0 = X 0.. Arrow of time. Let X 0 be a Gaussian random variable with zero mean and unit variance, and X n = αx n +Z n for n, where α is a fixed constant with α < and Z,Z,... are i.i.d. N(0, α ), independent of X 0. (a) Is the process {X n } Gaussian? (b) Is {X n } Markov? (c) Find R X (n,m). (d) Find the (nonlinear) MMSE estimate of X 00 given (X,X,...,X 99 ). (e) Find the MMSE estimate of X 00 given (X 0,X 0,...,X 99 ). (f) Find the MMSE estimate of X 00 given (X,...,X 99,X 0,...,X 99 ).

3 Solution: (a) Yes, the process is Gaussian, since it is the linear transform of white Gaussian process {Z n }. (b) Yes, the process is Markov since X n {X = x,...,x n = x n } N(αx n, α ), which depends only on x n. (c) First note that R X (n,n) = α R X (n,n )+( α ) = for all n. Since we can express X n = α k X n k +α k Z n k+ + +Z n, andx n k isindependent of(z n k+,...,z n ), wehaver X (n,n k) = E(X n X n k ) = α k. Thus, R X (n,m) = α n m. (d) Because the process is Gaussian, the MMSE estimator is linear. From Markovity, ˆX00 = E(X 00 X,...,X 99 ) = E(X 00 X 99 ) = R X(00,99) R X (99,99) X 99 = αx 99. (e) Again from the Markovity and the symmetry, (X,...,X n ) has the same distribution as (X n,...,x ) hence ˆX 00 = E(X 00 X 0,...,X 99 ) = E(X 00 X 0 ) = αx 0. (f) First note that X 00 is conditionally independent of (X,...,X 98,X 0,...,X 99 ) given (X 99,X 0 ). To see this, consider f(x 00 x,...,x 99,x 0,...,x 99 ) = Thus, f(x,...,x 99 ) f(x,...,x 99,x 0,...,x 99 ) = f(x )f(x x ) f ( x 99 x 98 )f(x 00 x 99 )f(x 0 x 00 )f(x 0 x 0 ) f(x 99 x 98 ) f(x )f(x x ) f(x 99 x 98 )f(x 0 x 99 )f(x 0 x 0 ) f(x 99 x 98 ) = f(x 00 x 99 )f(x 0 x 00 ) f(x 0 x 99 ) = f(x 00,x 0 x 99 ) f(x 0 x 99 ) = f(x 00 x 99,x 0 ). ˆX 00 = E(X 00 X 99,X 0 ) = [ R X (00,99) R X (00,0) ][ ] [ ] R X (99,99) R X (99,0) X99 R X (0,99) R X (0,0) X 0 = [ α α ][ α ] [ ] X99 α X 0 = α +α (X 99 +X 0 ). 3. AM modulation. Consider the AM modulated random process X(t) = A(t)cos(πt+Θ), 3

4 where the amplitude A(t) is a zero-mean WSS process with autocorrelation function R A (τ) = e τ, the phase Θ is a Unif[0,π) random variable, and A(t) and Θ are independent. Is X(t) a WSS process? Justify your answer. Solution: X(t)iswide-sensestationaryifEX(t)isindependentoftandifR X (t,t )depends only on t t. Consider and E[X(t)] = E[A(t)cos(ωt+Θ)] = E[A(t)]E[cos(ωt + Θ)] by independence = 0, R X (t,t ) = E[X(t )X(t )] = E[A(t )cos(ωt +Θ)A(t )cos(ωt +Θ)] = E[A(t )A(t )cos(ωt +Θ)cos(ωt +Θ)] = E[A(t )A(t ) Ecos(ωt +Θ)cos(ωt +Θ)] by independence = R A (t t )E[cos(ωt +Θ)cos(ωt +Θ)] [ ] = R A (t t )E (cos(ω(t +t )+Θ)+cos(ω(t t ))) = cos(ω(t +t ))cos(θ) R A(t t )E sin(ω(t +t ))sin(θ) +cos(ω(t t )) = Ecos(ω(t +t )) Ecos(Θ) R A(t t ) Esin(ω(t +t )) Esin(Θ) +Ecos(ω(t t )) = R A(t t )cos(ω(t t )), which is a function of t t only. Hence X(t) is wide-sense stationary. 4

5 4. LTI system with WSS process input. Let Y(t) = h(t) X(t) and Z(t) = X(t) Y(t) as shown in the Figure. (a) Find S Z (f). (b) Find E(Z (t)). Your answers should be in terms of S X (f) and the transfer function H(f) = F[h(t)]. X(t) h(t) Y(t) + - Z(t) Figure : LTI system. Solution: (a) To find S Z (f), we first find the autocorrelation function R Z (τ) = E(Z(t)Z(t+τ)) Now, taking the Fourier Transform, we get = E((X(t) Y(t))(X(t+τ) Y(t+τ))) = R X (τ)+r Y (τ) R YX ( τ) R XY ( τ) = R X (τ)+r Y (τ) R XY (τ) R XY ( τ). S Z (f) = S X (f)+s Y (f) S XY (f) S XY ( f) = S X (f)+ H(f) S X (f) H( f)s X (f) H(f)S X (f) = S X (f) ( + H(f) Re[H(f)] ) = S X (f) H(f). (b) To find the average power of Z(t), we find the area under S Z (f) E(Z (t)) = H(f) S X (f)df. 5. Echo filtering. A signal X(t) and its echo arrive at the receiver as Y(t) = X(t)+X(t )+ Z(t). Here the signal X(t) is a zero-mean WSS process with power spectral density S X (f) and the noise Z(t) is a zero-mean WSS with power spectral density S Z (f) = N 0 /, uncorrelated with X(t). (a) Find S Y (f) in terms of S X (f),, and N 0. (b) Find the best linear filter to estimate X(t) from {Y(s)} <s<. Solution: 5

6 (a) We can write Y(t) = g(t) X(t)+Z(t) where g(t) = δ(t)+δ(t ). Thus, S Y (f) = G(f) S X (f)+s Z (f) = +e jπ f S X (f)+ N 0. (b) Since S YX (f) = (+e jπ f )S X (f), ˆX(t) = h(t) Y(t), where the linear filter h(t) has the transfer function H(f) = S XY(f) S Y (f) = S YX( f) = S Y (f) (+e jπ f )S X (f) +e jπ f S X (f)+ N Discrete-time LTI system with white noise input. Let {X n : < n < } be a discrete-time white noise process, i.e., E(X n ) = 0, < n <, and { n = 0, R X (n) = 0 otherwise. The process is filtered using a linear time invariant system with impulse response α n = 0, h(n) = β n =, 0 otherwise. Find α and β such that the output process Y n has n = 0, R Y (n) = n =, 0 otherwise. Solution: We are given that R X (n) is a discrete-time unit impulse. Therefore R Y (n) = h(n) R X (n) h( n) = h(n) h( n). The impulse response h(n) is the sequence (α,β,0,0,...). The convolution with h( n) has only finitely many nonzero terms. R Y (0) = = h(0) h(0) = α +β R Y (+) = = h() h( ) = αβ R Y ( ) = = R Y () This pair of equations has two solutions: α = + and β = + or α = and β =. 7. Finding time of flight. Finding the distance to an object is often done by sending a signal and measuring the time of flight, the time it takes for the signal to return (assuming speed of signal, e.g., light, is known). Let X(t) be the signal sent and Y(t) = X(t δ)+z(t) be the 6

7 R Y X(t) Figure : Crosscorrelation function. t signal received, where δ is the unknown time of flight. Assume that X(t) and Z(t) (the sensor noise) are uncorrelated zero mean WSS processes. The estimated crosscorrelation function of Y(t) and X(t), R YX (t) is shown in Figure. Find the time of flight δ. Solution: Consider R YX (τ) = E(Y(t+τ)X(t)) = E((X(t +τ)+z(t+τ))x(t)) = R X (τ ). Now since the maximum of R X (α) is achieved for α = 0, by inspection of the given R YX we get that 5 = 0. Thus = 5. 7

8 Additional Exercises Do not turn in solutions to these problems.. Moving average process. Let Z 0,Z,Z,... be i.i.d. N(0,). (a) Let X n = Z n +Z n for n. Find the mean and autocorrelation function of X n. (b) Is {X n } wide-sense stationary? Justify your answer. (c) Is {X n } Gaussian? Justify your answer. (d) Is {X n } strict-sense stationary? Justify your answer. (e) Find E(X 3 X,X ). (f) Find E(X 3 X ). (g) Is {X n } Markov? Justify your answer. (h) Is {X n } independent increment? Justify your answer. (i) Let Y n = Z n +Z n for n. Find the mean and autocorrelation function of Y n. (j) Is {Y n } wide-sense stationary? Justify your answer. (k) Is {Y n } Gaussian? Justify your answer. (l) Is {Y n } strict-sense stationary? Justify your answer. (m) Find E(Y 3 Y,Y ). (n) Find E(Y 3 Y ). (o) Is {Y n } Markov? Justify your answer. (p) Is {Y n } independent increment? Justify your answer. Solution: (a) E(X n ) = E(Z n )+E(Z n ) = 0. R X (m,n) = E(X m X n ) [( )( )] = E Z n +Z n Z m +Z m E[Z n ], n m = 4 = E[Z n ]+E[Z n], n = m E[Z n], m n = 0, otherwise = 5 4, n = m, n m = 0, otherwise. 8

9 (b) Since the mean and autocorrelation functions are time-invariant, the process is WSS. (c) Since (X,...,X n ) is a linear transform of a GRV (Z 0,Z,...,Z n ), the process is Gaussian. (d) Since the process is WSS and Gaussian, it is SSS. (e) Since the process is Gaussian, the conditional expectation (MMSE estimate) is linear. Hence, E(X 3 X,X ) = [ ] [ ] 5 [ ] 0 4 X = (0X 4X ). (f) Similarly, E(X 3 X ) = (/5)X. (g) Since E(X 3 X,X ) E(X 3 X ), the process is not Markov. (h) Since the process is not Markov, it is not independent increment. (i) The mean function µ n = E(Y n ) = 0. The autocorrelation function 5 4 X n = m, R Y (n,m) = n = m or n = m+, 0 otherwise. (j) Since the mean and autocorrelation functions are time-invariant, the process is WSS. (k) Since (Y,...,Y n ) is a linear transform of a GRV (Z 0,Z,...,Z n ), the process is Gaussian. (l) Since the process is WSS and Gaussian, it is SSS. (m) Since the process is Gaussian, the conditional expectation (MMSE estimate) is linear. Hence, E(Y 3 Y,Y ) = [ 0 ][ ] [ ] Y = 3 (Y Y ). (n) Similarly, E(Y 3 Y ) = (/)Y. (o) Since E(Y 3 Y,Y ) E(Y 3 Y ), the process is not Markov. (p) Since the process is not Markov, it is not independent increment.. Gauss-Markov process. Let X 0 = 0 and X n = X n + Z n for n, where Z,Z,... are i.i.d. N(0,). Find the mean and autocorrelation function of X n. Solution: Using the method of lecture notes 8, we can easily verify that E(X n ) = 0 for every n and that R X (n,m) = E(X n X m ) = n m 4 3 [ ( 4 )min{n,m} ]. 3. Random binary waveform. In a digital communication channel the symbol is represented by the fixed duration rectangular pulse { for 0 t < g(t) = 0 otherwise, Y 9

10 and the symbol 0 is represented by g(t). The data transmitted over the channel is represented by the random process X(t) = A k g(t k), for t 0, k=0 where A 0,A,... are i.i.d random variables with { + w.p. A i = w.p.. (a) Find its first and second order pmfs. (b) Find the mean and the autocorrelation function of the process X(t). Solution: (a) The first order pmf is p X(t) (x) = P(X(t) = x) ( ) = P A k g(t k) = x k=0 = P ( A t = x ) = P(A 0 = x) IID { =, x = ± 0, otherwise. Now note that X(t ) and X(t ) are dependent only if t and t fall within the same time interval. Otherwise, they are independent. Thus, the second order pmf is (b) For t 0, p X(t )X(t )(x,y) = P(X(t ) = x,x(t ) = y) ( ) = P A k g(t k) = x, A k g(t k) = y k=0 k=0 = P ( A t = x,a t = y ) { P(A0 = x,a = 0 = y), t = t P(A 0 = x,a = y), otherwise, t = t & (x,y) = (,),(, ) = 4, t t & (x,y) = (,),(, ),(,),(, ) 0, otherwise. [ ] E[X(t)] = E A k g(t k) = k=0 g(t k)e[a k ] k=0 = 0. 0

11 For the autocorrelation R X (t,t ), we note once again that only if t and t fall within the same interval, will X(t ) be dependent on X(t ); if they do not fall in the same interval then they are independent from one another. Then, R X (t,t ) = E[X(t )X(t )] = g(t k)g(t k)e[a k ] = k=0 {, t = t 0, otherwise. 4. Absolute-value random walk. Let X n be a random walk defined by X 0 = 0, X n = Z i, n, where {Z i } is an i.i.d. process with P(Z = ) = P(Z = +) =. Define the absolute value random process Y n = X n. (a) Find P{Y n = k}. (b) Find P{max i<0 Y i = 0 Y 0 = 0}.

12 Solution: (a) This is a straightforward calculation and we can use results from lecture notes. If k 0 then P{Y n = k} = P{X n = +k or X n = k}. If k > 0 then P{Y n = k} = P{X n = k}, while P{Y n = 0} = P{X n = 0}. Thus ( n n (n+k)/)( ) k > 0, n k is even, n k 0 ( P{Y n = k} = n n n/)( ) k = 0, n is even, n 0 0 otherwise. (b) If Y 0 = X 0 = 0 then there are only two sample paths with max i<0 X i = 0 that is, Z = Z = = Z 0 = +,Z = = Z 0 = or Z = Z = = Z 0 =,Z = = Z 0 = +. Since the total number of sample paths is ( 0 0) and all paths are equally likely, P { max Y i = 0 Y 0 = 0 } = ( i<0 0 ) = = Mixture of two WSS processes. Let X(t) and Y(t) be two zero-mean WSS processes with autocorrelation functions R X (τ) and R Y (τ), respectively. Define the process { X(t), with probability Z(t) = Y(t), with probability. Find the mean and autocorrelation functions for Z(t). Is Z(t) a WSS process? Justify your answer. Solution: To show that Z(t) is WSS, we show that its mean and autocorrelation functions are time invariant. Consider and similarly µ Z (t) = E[Z(t)] = E(Z Z = X)P{Z = X}+E(Z Z = Y)P{Z = Y} = (µ X +µ Y ) = 0, R Z (t+τ,t) = E[Z(t+τ)Z(t)] = (R X(τ)+R Y (τ)). Since µ Z (t) is independent of time and R Z (t+τ,t) depends only on τ, Z(t) is WSS.

13 6. Stationary Gauss-Markov process. Consider the following variation on the Gauss-Markov process discussed in Lecture Notes 8: X 0 N(0,a) X n = X n +Z n, n, where Z,Z,Z 3,... are i.i.d. N(0,) independent of X 0. (a) Find a such that X n is stationary. Find the mean and autocorrelation functions of X n. (b) (Optional.) Consider the sample mean S n = n n X i, n. Show that S n converges totheprocessmeaninprobabilityeventhoughthesequencex n isnoti.i.d. (Astationary process for which the sample mean converges to the process mean is called mean ergodic.) Solution: (a) We are asked to find a such that E(X n ) is independent of n and R X (n,n ) depends only on n n. For X n to be stationary, E(X n) must be independent of n. Thus E(X n) = 4 E(X n )+E(Z n)+e(x n Z n ) = 4 E(X n)+. Therefore, a = E(X0 ) = E(X n) = 4 3. Using the method of lecture notes 8, we can easily verify that E(X n ) = 0 for every n and that R X (n,n ) = E(X n X n ) = 4 3 n n. (b) To prove convergence in probability, we first prove convergence in mean square and then use the fact that mean square convergence implies convergence in probability. ( E(S n ) = E n ) X i = n E(X i ) = n 0 = 0. To show convergence in mean square we show that Var(S n ) 0 as n. ( Var(S n ) = Var n = n 4 3n j= ) (( X i = E n R X (i,j) = 4 3n ( n ) + i ) ) X i ( n ) n+ (n i) i 4 ( + 3n ) i = 4 n. (since E(X i ) = 0) Thus S n converges to the process mean, even though the sequence is not i.i.d. 3

14 7. QAM random process. Consider the random process X(t) = Z cosωt+z sinωt, < t <, where Z and Z are i.i.d. discrete random variables such that p Zi (+) = p Zi ( ) =. (a) Is X(t) wide-sense stationary? Justify your answer. (b) Is X(t) strict-sense stationary? Justify your answer. Solution: (a) We first check the mean. E(X(t)) = E(Z )cosωt+e(z )sinωt = 0 cos(ωt) + 0 sin(ωt) = 0. The mean is independent of t. Next we consider the autocorrelation function. E(X(t+τ)X(t)) = E((Z cos(ω(t+τ))+z sin(ω(t+τ)))(z cos(ωt)+z sin(ωt))) = E(Z )cos(ω(t+τ))cos(ωt)+e(z )sin(ω(t+τ))sin(ωt) = cos(ω(t+τ))cos(ωt)+sin(ω(t+τ))sin(ωt) = cos(ω(t+τ) ωt)) = cosωτ. The autocorrelation function is also time invariant. Therefore X(t) is WSS. (b) Note that X(0) = Z cos0 + Z sin0 = Z, so X(0) has the same pmf as Z. On the other hand, X(π/4ω) = Z cos(π/4)+z (sinπ/4) = (Z +Z ) 4 x = ± = x = 0 0 otherwise This shows that X(π/4ω) does not have the same pdf or even same range as X(0). Therefore X(t) is not first-order stationary and as a result is not SSS. 8. Finding impulse response of LTI system. To find the impulse response h(t) of an LTI system (e.g., a concert hall), i.e., to identify the system, white noise X(t), < t <, is applied to its input and the output Y(t) is measured. Given the input and output sample functions, the crosscorrelation R YX (τ) is estimated. Show how R YX (τ) can be used to find h(t). Solution: Since white noise has a flat psd, the crosspower spectral density of the input X(t) and the output Y(t) is just the transfer function of the system scaled by the psd of the white noise. S YX (f) = H(f)S X (f) = H(f) N 0 R YX (τ) = F (S YX (f)) = N 0 h(τ) 4

15 Thus to estimate the impulse response of a linear time invariant system, we apply white noise to its input, estimate the crosscorrelation function of its input and output, and scale it by /N 0. 5

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011 UCSD ECE53 Handout #40 Prof. Young-Han Kim Thursday, May 9, 04 Homework Set #8 Due: Thursday, June 5, 0. Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process,

More information

UCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7

UCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7 UCSD ECE50 Handout #4 Prof Young-Han Kim Wednesday, June 6, 08 Solutions to Exercise Set #7 Polya s urn An urn initially has one red ball and one white ball Let X denote the name of the first ball drawn

More information

UCSD ECE250 Handout #20 Prof. Young-Han Kim Monday, February 26, Solutions to Exercise Set #7

UCSD ECE250 Handout #20 Prof. Young-Han Kim Monday, February 26, Solutions to Exercise Set #7 UCSD ECE50 Handout #0 Prof. Young-Han Kim Monday, February 6, 07 Solutions to Exercise Set #7. Minimum waiting time. Let X,X,... be i.i.d. exponentially distributed random variables with parameter λ, i.e.,

More information

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017) UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable

More information

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011 UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and

More information

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE53 Handout #34 Prof Young-Han Kim Tuesday, May 7, 04 Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) Linear estimator Consider a channel with the observation Y XZ, where the

More information

SRI VIDYA COLLEGE OF ENGINEERING AND TECHNOLOGY UNIT 3 RANDOM PROCESS TWO MARK QUESTIONS

SRI VIDYA COLLEGE OF ENGINEERING AND TECHNOLOGY UNIT 3 RANDOM PROCESS TWO MARK QUESTIONS UNIT 3 RANDOM PROCESS TWO MARK QUESTIONS 1. Define random process? The sample space composed of functions of time is called a random process. 2. Define Stationary process? If a random process is divided

More information

Chapter 6: Random Processes 1

Chapter 6: Random Processes 1 Chapter 6: Random Processes 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.

More information

16.584: Random (Stochastic) Processes

16.584: Random (Stochastic) Processes 1 16.584: Random (Stochastic) Processes X(t): X : RV : Continuous function of the independent variable t (time, space etc.) Random process : Collection of X(t, ζ) : Indexed on another independent variable

More information

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process Lecture Notes 7 Stationary Random Processes Strict-Sense and Wide-Sense Stationarity Autocorrelation Function of a Stationary Process Power Spectral Density Continuity and Integration of Random Processes

More information

Introduction to Probability and Stochastic Processes I

Introduction to Probability and Stochastic Processes I Introduction to Probability and Stochastic Processes I Lecture 3 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark Slides

More information

EAS 305 Random Processes Viewgraph 1 of 10. Random Processes

EAS 305 Random Processes Viewgraph 1 of 10. Random Processes EAS 305 Random Processes Viewgraph 1 of 10 Definitions: Random Processes A random process is a family of random variables indexed by a parameter t T, where T is called the index set λ i Experiment outcome

More information

Stochastic Processes

Stochastic Processes Elements of Lecture II Hamid R. Rabiee with thanks to Ali Jalali Overview Reading Assignment Chapter 9 of textbook Further Resources MIT Open Course Ware S. Karlin and H. M. Taylor, A First Course in Stochastic

More information

ECE Homework Set 3

ECE Homework Set 3 ECE 450 1 Homework Set 3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3

More information

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno Stochastic Processes M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno 1 Outline Stochastic (random) processes. Autocorrelation. Crosscorrelation. Spectral density function.

More information

Name of the Student: Problems on Discrete & Continuous R.Vs

Name of the Student: Problems on Discrete & Continuous R.Vs Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan

More information

P 1.5 X 4.5 / X 2 and (iii) The smallest value of n for

P 1.5 X 4.5 / X 2 and (iii) The smallest value of n for DHANALAKSHMI COLLEGE OF ENEINEERING, CHENNAI DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING MA645 PROBABILITY AND RANDOM PROCESS UNIT I : RANDOM VARIABLES PART B (6 MARKS). A random variable X

More information

ECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else

ECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else ECE 450 Homework #3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3 4 5

More information

Properties of the Autocorrelation Function

Properties of the Autocorrelation Function Properties of the Autocorrelation Function I The autocorrelation function of a (real-valued) random process satisfies the following properties: 1. R X (t, t) 0 2. R X (t, u) =R X (u, t) (symmetry) 3. R

More information

EEM 409. Random Signals. Problem Set-2: (Power Spectral Density, LTI Systems with Random Inputs) Problem 1: Problem 2:

EEM 409. Random Signals. Problem Set-2: (Power Spectral Density, LTI Systems with Random Inputs) Problem 1: Problem 2: EEM 409 Random Signals Problem Set-2: (Power Spectral Density, LTI Systems with Random Inputs) Problem 1: Consider a random process of the form = + Problem 2: X(t) = b cos(2π t + ), where b is a constant,

More information

Chapter 2 Random Processes

Chapter 2 Random Processes Chapter 2 Random Processes 21 Introduction We saw in Section 111 on page 10 that many systems are best studied using the concept of random variables where the outcome of a random experiment was associated

More information

Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts

Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts White Gaussian Noise I Definition: A (real-valued) random process X t is called white Gaussian Noise if I X t is Gaussian for each time instance t I Mean: m X (t) =0 for all t I Autocorrelation function:

More information

for valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I

for valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I Code: 15A04304 R15 B.Tech II Year I Semester (R15) Regular Examinations November/December 016 PROBABILITY THEY & STOCHASTIC PROCESSES (Electronics and Communication Engineering) Time: 3 hours Max. Marks:

More information

Fundamentals of Noise

Fundamentals of Noise Fundamentals of Noise V.Vasudevan, Department of Electrical Engineering, Indian Institute of Technology Madras Noise in resistors Random voltage fluctuations across a resistor Mean square value in a frequency

More information

ECE534, Spring 2018: Solutions for Problem Set #5

ECE534, Spring 2018: Solutions for Problem Set #5 ECE534, Spring 08: s for Problem Set #5 Mean Value and Autocorrelation Functions Consider a random process X(t) such that (i) X(t) ± (ii) The number of zero crossings, N(t), in the interval (0, t) is described

More information

= 4. e t/a dt (2) = 4ae t/a. = 4a a = 1 4. (4) + a 2 e +j2πft 2

= 4. e t/a dt (2) = 4ae t/a. = 4a a = 1 4. (4) + a 2 e +j2πft 2 ECE 341: Probability and Random Processes for Engineers, Spring 2012 Homework 13 - Last homework Name: Assigned: 04.18.2012 Due: 04.25.2012 Problem 1. Let X(t) be the input to a linear time-invariant filter.

More information

Random Processes Handout IV

Random Processes Handout IV RP-IV.1 Random Processes Handout IV CALCULATION OF MEAN AND AUTOCORRELATION FUNCTIONS FOR WSS RPS IN LTI SYSTEMS In the last classes, we calculated R Y (τ) using an intermediate function f(τ) (h h)(τ)

More information

Solutions to Homework Set #6 (Prepared by Lele Wang)

Solutions to Homework Set #6 (Prepared by Lele Wang) Solutions to Homework Set #6 (Prepared by Lele Wang) Gaussian random vector Given a Gaussian random vector X N (µ, Σ), where µ ( 5 ) T and 0 Σ 4 0 0 0 9 (a) Find the pdfs of i X, ii X + X 3, iii X + X

More information

UCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE53 Handout #7 Prof. Young-Han Kim Tuesday, May 6, 4 Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei). Neural net. Let Y = X + Z, where the signal X U[,] and noise Z N(,) are independent.

More information

Final Examination Solutions (Total: 100 points)

Final Examination Solutions (Total: 100 points) Final Examination Solutions (Total: points) There are 4 problems, each problem with multiple parts, each worth 5 points. Make sure you answer all questions. Your answer should be as clear and readable

More information

Communication Systems Lecture 21, 22. Dong In Kim School of Information & Comm. Eng. Sungkyunkwan University

Communication Systems Lecture 21, 22. Dong In Kim School of Information & Comm. Eng. Sungkyunkwan University Communication Systems Lecture 1, Dong In Kim School of Information & Comm. Eng. Sungkyunkwan University 1 Outline Linear Systems with WSS Inputs Noise White noise, Gaussian noise, White Gaussian noise

More information

Signals and Spectra (1A) Young Won Lim 11/26/12

Signals and Spectra (1A) Young Won Lim 11/26/12 Signals and Spectra (A) Copyright (c) 202 Young W. Lim. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version.2 or any later

More information

Name of the Student: Problems on Discrete & Continuous R.Vs

Name of the Student: Problems on Discrete & Continuous R.Vs Engineering Mathematics 08 SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : University Questions REGULATION : R03 UPDATED ON : November 07 (Upto N/D 07 Q.P) (Scan the

More information

E X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.

E X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl. E X A M Course code: Course name: Number of pages incl. front page: 6 MA430-G Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours Resources allowed: Notes: Pocket calculator,

More information

ECE6604 PERSONAL & MOBILE COMMUNICATIONS. Week 3. Flat Fading Channels Envelope Distribution Autocorrelation of a Random Process

ECE6604 PERSONAL & MOBILE COMMUNICATIONS. Week 3. Flat Fading Channels Envelope Distribution Autocorrelation of a Random Process 1 ECE6604 PERSONAL & MOBILE COMMUNICATIONS Week 3 Flat Fading Channels Envelope Distribution Autocorrelation of a Random Process 2 Multipath-Fading Mechanism local scatterers mobile subscriber base station

More information

3F1 Random Processes Examples Paper (for all 6 lectures)

3F1 Random Processes Examples Paper (for all 6 lectures) 3F Random Processes Examples Paper (for all 6 lectures). Three factories make the same electrical component. Factory A supplies half of the total number of components to the central depot, while factories

More information

Chapter 4 Random process. 4.1 Random process

Chapter 4 Random process. 4.1 Random process Random processes - Chapter 4 Random process 1 Random processes Chapter 4 Random process 4.1 Random process 4.1 Random process Random processes - Chapter 4 Random process 2 Random process Random process,

More information

ECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen. D. van Alphen 1

ECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen. D. van Alphen 1 ECE 650 Lecture #10 (was Part 1 & 2) D. van Alphen D. van Alphen 1 Lecture 10 Overview Part 1 Review of Lecture 9 Continuing: Systems with Random Inputs More about Poisson RV s Intro. to Poisson Processes

More information

Random Processes Why we Care

Random Processes Why we Care Random Processes Why we Care I Random processes describe signals that change randomly over time. I Compare: deterministic signals can be described by a mathematical expression that describes the signal

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your

More information

Problem Sheet 1 Examples of Random Processes

Problem Sheet 1 Examples of Random Processes RANDOM'PROCESSES'AND'TIME'SERIES'ANALYSIS.'PART'II:'RANDOM'PROCESSES' '''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''Problem'Sheets' Problem Sheet 1 Examples of Random Processes 1. Give

More information

Probability and Statistics for Final Year Engineering Students

Probability and Statistics for Final Year Engineering Students Probability and Statistics for Final Year Engineering Students By Yoni Nazarathy, Last Updated: May 24, 2011. Lecture 6p: Spectral Density, Passing Random Processes through LTI Systems, Filtering Terms

More information

7 The Waveform Channel

7 The Waveform Channel 7 The Waveform Channel The waveform transmitted by the digital demodulator will be corrupted by the channel before it reaches the digital demodulator in the receiver. One important part of the channel

More information

Problems on Discrete & Continuous R.Vs

Problems on Discrete & Continuous R.Vs 013 SUBJECT NAME SUBJECT CODE MATERIAL NAME MATERIAL CODE : Probability & Random Process : MA 61 : University Questions : SKMA1004 Name of the Student: Branch: Unit I (Random Variables) Problems on Discrete

More information

EE538 Final Exam Fall :20 pm -5:20 pm PHYS 223 Dec. 17, Cover Sheet

EE538 Final Exam Fall :20 pm -5:20 pm PHYS 223 Dec. 17, Cover Sheet EE538 Final Exam Fall 005 3:0 pm -5:0 pm PHYS 3 Dec. 17, 005 Cover Sheet Test Duration: 10 minutes. Open Book but Closed Notes. Calculators ARE allowed!! This test contains five problems. Each of the five

More information

The distribution inherited by Y is called the Cauchy distribution. Using that. d dy ln(1 + y2 ) = 1 arctan(y)

The distribution inherited by Y is called the Cauchy distribution. Using that. d dy ln(1 + y2 ) = 1 arctan(y) Stochastic Processes - MM3 - Solutions MM3 - Review Exercise Let X N (0, ), i.e. X is a standard Gaussian/normal random variable, and denote by f X the pdf of X. Consider also a continuous random variable

More information

Fig 1: Stationary and Non Stationary Time Series

Fig 1: Stationary and Non Stationary Time Series Module 23 Independence and Stationarity Objective: To introduce the concepts of Statistical Independence, Stationarity and its types w.r.to random processes. This module also presents the concept of Ergodicity.

More information

ECE353: Probability and Random Processes. Lecture 18 - Stochastic Processes

ECE353: Probability and Random Processes. Lecture 18 - Stochastic Processes ECE353: Probability and Random Processes Lecture 18 - Stochastic Processes Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu From RV

More information

ECE-340, Spring 2015 Review Questions

ECE-340, Spring 2015 Review Questions ECE-340, Spring 2015 Review Questions 1. Suppose that there are two categories of eggs: large eggs and small eggs, occurring with probabilities 0.7 and 0.3, respectively. For a large egg, the probabilities

More information

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random

More information

ECE 636: Systems identification

ECE 636: Systems identification ECE 636: Systems identification Lectures 3 4 Random variables/signals (continued) Random/stochastic vectors Random signals and linear systems Random signals in the frequency domain υ ε x S z + y Experimental

More information

EE4601 Communication Systems

EE4601 Communication Systems EE4601 Communication Systems Week 4 Ergodic Random Processes, Power Spectrum Linear Systems 0 c 2011, Georgia Institute of Technology (lect4 1) Ergodic Random Processes An ergodic random process is one

More information

UCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE 53 Handout #0 Prof. Young-Han Kim Thursday, April 4, 04 Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei). Time until the n-th arrival. Let the random variable N(t) be the number

More information

This examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS

This examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of

More information

This examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS

This examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 6 December 2006 This examination consists of

More information

ECE 630: Statistical Communication Theory

ECE 630: Statistical Communication Theory ECE 630: Statistical Communication Theory Dr. B.-P. Paris Dept. Electrical and Comp. Engineering George Mason University Last updated: April 6, 2017 2017, B.-P. Paris ECE 630: Statistical Communication

More information

Module 9: Stationary Processes

Module 9: Stationary Processes Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.

More information

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π Solutions to Homework Set #5 (Prepared by Lele Wang). Neural net. Let Y X + Z, where the signal X U[,] and noise Z N(,) are independent. (a) Find the function g(y) that minimizes MSE E [ (sgn(x) g(y))

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3.0 INTRODUCTION The purpose of this chapter is to introduce estimators shortly. More elaborated courses on System Identification, which are given

More information

Gaussian, Markov and stationary processes

Gaussian, Markov and stationary processes Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November

More information

Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes

Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes Fundamentals of Digital Commun. Ch. 4: Random Variables and Random Processes Klaus Witrisal witrisal@tugraz.at Signal Processing and Speech Communication Laboratory www.spsc.tugraz.at Graz University of

More information

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172.

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172. ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172. 1. Enter your name, student ID number, e-mail address, and signature in the space provided on this page, NOW! 2. This is a closed book exam.

More information

2A1H Time-Frequency Analysis II Bugs/queries to HT 2011 For hints and answers visit dwm/courses/2tf

2A1H Time-Frequency Analysis II Bugs/queries to HT 2011 For hints and answers visit   dwm/courses/2tf Time-Frequency Analysis II (HT 20) 2AH 2AH Time-Frequency Analysis II Bugs/queries to david.murray@eng.ox.ac.uk HT 20 For hints and answers visit www.robots.ox.ac.uk/ dwm/courses/2tf David Murray. A periodic

More information

4 Derivations of the Discrete-Time Kalman Filter

4 Derivations of the Discrete-Time Kalman Filter Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time

More information

Final. Fall 2016 (Dec 16, 2016) Please copy and write the following statement:

Final. Fall 2016 (Dec 16, 2016) Please copy and write the following statement: ECE 30: Probabilistic Methods in Electrical and Computer Engineering Fall 06 Instructor: Prof. Stanley H. Chan Final Fall 06 (Dec 6, 06) Name: PUID: Please copy and write the following statement: I certify

More information

Name of the Student: Problems on Discrete & Continuous R.Vs

Name of the Student: Problems on Discrete & Continuous R.Vs SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : Additional Problems MATERIAL CODE : JM08AM004 REGULATION : R03 UPDATED ON : March 05 (Scan the above QR code for the direct

More information

Stochastic Processes. Chapter Definitions

Stochastic Processes. Chapter Definitions Chapter 4 Stochastic Processes Clearly data assimilation schemes such as Optimal Interpolation are crucially dependent on the estimates of background and observation error statistics. Yet, we don t know

More information

VALLIAMMAI ENGINEERING COLLEGE SRM Nagar, Kattankulathur 603 203. DEPARTMENT OF ELECTRONICS & COMMUNICATION ENGINEERING SUBJECT QUESTION BANK : MA6451 PROBABILITY AND RANDOM PROCESSES SEM / YEAR:IV / II

More information

ECE 673-Random signal analysis I Final

ECE 673-Random signal analysis I Final ECE 673-Random signal analysis I Final Q ( point) Comment on the following statement "If cov(x ; X 2 ) 0; then the best predictor (without limitation on the type of predictor) of the random variable X

More information

Chapter 6. Random Processes

Chapter 6. Random Processes Chapter 6 Random Processes Random Process A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). For a fixed (sample path): a random process

More information

2A1H Time-Frequency Analysis II

2A1H Time-Frequency Analysis II 2AH Time-Frequency Analysis II Bugs/queries to david.murray@eng.ox.ac.uk HT 209 For any corrections see the course page DW Murray at www.robots.ox.ac.uk/ dwm/courses/2tf. (a) A signal g(t) with period

More information

Probability and Statistics

Probability and Statistics Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph

More information

PROBABILITY AND RANDOM PROCESSESS

PROBABILITY AND RANDOM PROCESSESS PROBABILITY AND RANDOM PROCESSESS SOLUTIONS TO UNIVERSITY QUESTION PAPER YEAR : JUNE 2014 CODE NO : 6074 /M PREPARED BY: D.B.V.RAVISANKAR ASSOCIATE PROFESSOR IT DEPARTMENT MVSR ENGINEERING COLLEGE, NADERGUL

More information

Statistical signal processing

Statistical signal processing Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable

More information

Chapter 5 Random Variables and Processes

Chapter 5 Random Variables and Processes Chapter 5 Random Variables and Processes Wireless Information Transmission System Lab. Institute of Communications Engineering National Sun Yat-sen University Table of Contents 5.1 Introduction 5. Probability

More information

Stationary independent increments. 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process.

Stationary independent increments. 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process. Stationary independent increments 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process. 2. If each set of increments, corresponding to non-overlapping collection of

More information

2. (a) What is gaussian random variable? Develop an equation for guassian distribution

2. (a) What is gaussian random variable? Develop an equation for guassian distribution Code No: R059210401 Set No. 1 II B.Tech I Semester Supplementary Examinations, February 2007 PROBABILITY THEORY AND STOCHASTIC PROCESS ( Common to Electronics & Communication Engineering, Electronics &

More information

ENGG2430A-Homework 2

ENGG2430A-Homework 2 ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,

More information

Empirical Mean and Variance!

Empirical Mean and Variance! Global Image Properties! Global image properties refer to an image as a whole rather than components. Computation of global image properties is often required for image enhancement, preceding image analysis.!

More information

GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems

GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems Aids Allowed: 2 8 1/2 X11 crib sheets, calculator DATE: Tuesday

More information

Name of the Student: Problems on Discrete & Continuous R.Vs

Name of the Student: Problems on Discrete & Continuous R.Vs Engineering Mathematics 03 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA 6 MATERIAL NAME : Problem Material MATERIAL CODE : JM08AM008 (Scan the above QR code for the direct download of

More information

FINAL EXAM: 3:30-5:30pm

FINAL EXAM: 3:30-5:30pm ECE 30: Probabilistic Methods in Electrical and Computer Engineering Spring 016 Instructor: Prof. A. R. Reibman FINAL EXAM: 3:30-5:30pm Spring 016, MWF 1:30-1:0pm (May 6, 016) This is a closed book exam.

More information

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution ECE 54 Stochastic Signals and Systems Problem Set Solution Problem Solutions : Yates and Goodman,..4..7.3.3.4.3.8.3 and.8.0 Problem..4 Solution Since E[Y (t] R Y (0, we use Theorem.(a to evaluate R Y (τ

More information

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver Stochastic Signals Overview Definitions Second order statistics Stationarity and ergodicity Random signal variability Power spectral density Linear systems with stationary inputs Random signal memory Correlation

More information

COURSE OUTLINE. Introduction Signals and Noise Filtering Noise Sensors and associated electronics. Sensors, Signals and Noise 1

COURSE OUTLINE. Introduction Signals and Noise Filtering Noise Sensors and associated electronics. Sensors, Signals and Noise 1 Sensors, Signals and Noise 1 COURSE OUTLINE Introduction Signals and Noise Filtering Noise Sensors and associated electronics Processing Noise with Linear Filters 2 Mathematical Foundations Filtering Stationary

More information

Question Paper Code : AEC11T03

Question Paper Code : AEC11T03 Hall Ticket No Question Paper Code : AEC11T03 VARDHAMAN COLLEGE OF ENGINEERING (AUTONOMOUS) Affiliated to JNTUH, Hyderabad Four Year B Tech III Semester Tutorial Question Bank 2013-14 (Regulations: VCE-R11)

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

UNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS

UNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS UNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS In many practical situations, multiple random variables are required for analysis than a single random variable. The analysis of two random variables especially

More information

Digital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10

Digital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10 Digital Band-pass Modulation PROF. MICHAEL TSAI 211/11/1 Band-pass Signal Representation a t g t General form: 2πf c t + φ t g t = a t cos 2πf c t + φ t Envelope Phase Envelope is always non-negative,

More information

5 Operations on Multiple Random Variables

5 Operations on Multiple Random Variables EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y

More information

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering Stochastic Processes and Linear Algebra Recap Slides Stochastic processes and variables XX tt 0 = XX xx nn (tt) xx 2 (tt) XX tt XX

More information

TSKS01 Digital Communication Lecture 1

TSKS01 Digital Communication Lecture 1 TSKS01 Digital Communication Lecture 1 Introduction, Repetition, and Noise Modeling Emil Björnson Department of Electrical Engineering (ISY) Division of Communication Systems Emil Björnson Course Director

More information

Signals and Systems Chapter 2

Signals and Systems Chapter 2 Signals and Systems Chapter 2 Continuous-Time Systems Prof. Yasser Mostafa Kadah Overview of Chapter 2 Systems and their classification Linear time-invariant systems System Concept Mathematical transformation

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

GATE EE Topic wise Questions SIGNALS & SYSTEMS

GATE EE Topic wise Questions SIGNALS & SYSTEMS www.gatehelp.com GATE EE Topic wise Questions YEAR 010 ONE MARK Question. 1 For the system /( s + 1), the approximate time taken for a step response to reach 98% of the final value is (A) 1 s (B) s (C)

More information

Chapter 6 - Random Processes

Chapter 6 - Random Processes EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process

More information

MA6451 PROBABILITY AND RANDOM PROCESSES

MA6451 PROBABILITY AND RANDOM PROCESSES MA6451 PROBABILITY AND RANDOM PROCESSES UNIT I RANDOM VARIABLES 1.1 Discrete and continuous random variables 1. Show that the function is a probability density function of a random variable X. (Apr/May

More information

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2}, ECE32 Spring 25 HW Solutions April 6, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where

More information

Lecture 4: Least Squares (LS) Estimation

Lecture 4: Least Squares (LS) Estimation ME 233, UC Berkeley, Spring 2014 Xu Chen Lecture 4: Least Squares (LS) Estimation Background and general solution Solution in the Gaussian case Properties Example Big picture general least squares estimation:

More information