UCSD ECE250 Handout #20 Prof. Young-Han Kim Monday, February 26, Solutions to Exercise Set #7
|
|
- Corey Harper
- 6 years ago
- Views:
Transcription
1 UCSD ECE50 Handout #0 Prof. Young-Han Kim Monday, February 6, 07 Solutions to Exercise Set #7. Minimum waiting time. Let X,X,... be i.i.d. exponentially distributed random variables with parameter λ, i.e., f Xi (x) λe λx, for x 0. (a) Does Y n min{x,x,...,x n } converge in probability as n approaches infinity? (b) If it converges what is the limit? (c) What about Z n ny n? (a) For any set of values X i s, the sequence Y n is monotonically decreasing in n. Since the random variables are nonnegative, it is reasonable to guess that Y n converges to 0. Now Y n will converge in probability to 0 if and only if for any ǫ > 0, lim n P{ Y n > ǫ} 0. P{ Y n > ǫ} P{Y n > ǫ} P{X > ǫ,x > ǫ,..,x n > ǫ} P{X > ǫ}p{x > ǫ}...p{x n > ǫ} ( F X (ǫ))( F X (ǫ))...( F X (ǫ)) ( F X (ǫ)) n ( ( e λǫ )) n e λǫn. As n goes to infinity (for any finite ǫ > 0) this converges to zero. Therefore Y n converges to 0 in probability. (b) The limit to which Y n converges in probability is 0. (c) Does Z n ny n converges to 0 in probability? No, it does not. In fact, P{ Z n > ǫ} P{nY n > ǫ} P{Y n > ǫ n } e λ ǫ n n e λǫ which does not depend on n. So Z n does not converge to 0 in probability. Note that the distribution of Z n is exponential with parameter λ, the same as the distribution of X i. F Zn (z) P{Z n < z} e λz.
2 In conclusion, if X i s are i.i.d. exp(λ), then Y n min i n {X i} exp(nλ) and Thus but Z n ny n exp(λ). P{Y n > ǫ} e λǫn 0, so Y n 0 in probability, P{Z n > ǫ} e λǫ 0, so Z n 0 in probability.. Roundoff errors. The sum of a list of 00 real numbers is to be computed. Suppose that these numbers are rounded off to the nearest integer so that each number has an error that is uniformly distributed in the interval ( 0.5, 0.5). Use the central limit theorem to estimate the probability that the total error in the sum of the 00 numbers exceeds 6. Errorsareindependentanduniformlydistributedin( 0.5,0.5), i.e. e i Unif( 0.5,0.5). Usingthe Central Limit Theorem, thetotal error e 00 i e i can beapproximated as a Gaussian random Variable with mean and Variance E(e) σ e 0 00 E(e i ) i 00 Var(e i ) i Therefore the probability of the given event is { e E(e) P{e > 6} P σ e ( ) 6 Q 8.33 > 6 0 } If the error is interpreted as an absolute difference, then P{ e > 6} P{e > 6}
3 3. The signal received over a wireless communication channel can be represented by two sums X n n Z j cosθ j, and n j X n n Z j sinθ j, n j where Z,Z,... are i.i.d. with mean µ and variance [ σ and Θ,Θ,... are i.i.d. U[0,π Xn independent of Z,Z,... Find the distribution of as n approaches. Since Z,Z,... are i.i.d. and Θ,Θ,... are also i.i.d. and independent of Z,Z,..., [ [ Z cosθ Z cosθ, Z sinθ Z sinθ [ Zj cosθ Let Y j j Z j sinθ j,... are i.i.d. random vectors.. We have Similarly, E[Z j sinθ j 0. Thus, E[Y j 0. Moreover, we have X n E[Z j cosθ j E[Z j E[cosΘ j µ 0. π 0 π cosθdθ E[(Z j cosθ j ) E[Z je[cos Θ j Similarly, E[(Z j sinθ j ) µ +σ. Also, (µ +σ ) µ +σ. π 0 π cos θdθ E[(Z j cosθ j ) (Z j sinθ j ) E[Z j E[cosΘ j sinθ j (µ +σ ) 0. π 0 π cosθsinθdθ 3
4 Thus, Now, [ Xn X n Cov(Y j ) E[Y j Y T j E[Y j E[Y j T E[Y j Yj T [ E[(Z j cosθ j ) E[(Z j cosθ j ) (Z j sinθ j ) E[(Z j cosθ j ) (Z j sinθ j ) E[(Z j sinθ j ) µ +σ I. n Y j, and thus, by the central limit theorem, as n, n j [ Xn X n N (0, µ +σ ) I in distribution. 4. Polya urn. An urn initially has one red ball and one white ball. Let X denote the name of the first ball drawn from the urn. Replace that ball and one like it. Let X denote the name of the next ball drawn. Replace it and one like it. Continue, drawing and replacing. (a) Argue that the probability of drawing k reds follwed by n k whites is 3... k k+ (n k)... (k +) (n+) k!(n k)! (n+)! ( (n+) nk ). (b) Let P n be the proportion of red balls in the urn after the n th drawing. Argue that Pr{P n k n+ } n+, for k,,...,n+. Thus all proportions are equally probable. This shows that P n tends to a uniformly distributed random Variable in distribution, i.e., lim Pr{P n t} t, 0 t. n (c) What can you say about the behavior of the proportion P n if you started initially with one red ball in the urn and two white balls? Specifically, what is the limiting distribution of P n? Can you show Pr{P n k n+3 } k n+, for k,,...,n+? (a) Let r be the number of red balls, w the number of white balls and t the total number of balls r +w in the urn. The picture of the balls in the urn for each drawing is given below. The ratio r/t or w/t gives the probability for the next drawing. 4
5 r/t / st drawing r/t /3 d drawing r/t 3/4... (k ) th drawing r/t k/k + k th drawing w/t /k + (k+) th drawing w/t /k +3 (k+) th drawing w/t 3/k +4.. (n ) th drawing n th drawing The probability of drawing k reds followed by n k whites is. w/t n k/n k (k +) (n k)... (k +) (n+) k!(n k)! (n+)! ( n. k) (n+) In fact, regardless of the ordering, the probability of drawing k reds in n draws is given by this expression as well. Note that after the n th drawing, there are k+ red balls and n k+ white balls for a total of n+ balls in the urn. (b) After the n th drawing, where k red balls have been drawn, the proportion P n of red balls is number of red balls in the urn P n total number of balls in the urn k+ n+. There are ( n k) orderings of outcomes with k red balls and n k white balls, and each ordering has the same probability. In fact, in the expression of the probability of drawing k reds followed by n k whites, there will be just permutations of the numerators in the case of a different sequence of drawings. Therefore { Pr P n k+ } n+ ( ) n k n+ ( ( n k ) ) (n+) for k 0,...,n. All the proportions are equally probable (the probability does not depend on k) and P n tends to a uniformly distributed random Variable Unif[0, in distribution. (c) If there are one red ball and two white balls to start with: 5
6 r/t /3 st drawing r/t /4 d drawing r/t 3/5... (k ) th drawing r/t k/k + k th drawing w/t /k +3 (k +) th drawing w/t 3/k +4 (k +) th drawing w/t 4/k +5.. (n ) th drawing n th drawing The probability of drawing k reds followed by n k whites is After the n th drawing,. w/t n k+/n k (k +) (n k+) k!(n k +)!... (k +3) (n+) (n+)! Similarly as in part (b), { Pr P n k + } n+3 number of red balls in the urn P n total number of balls in the urn k+ n+3. The distribution is linearly decreasing in p k+ Pr{P n p} ( ) ( ) n k!(n k+)! k (n+)! ( ) n! k!(n k +)! k!(n k)! (n+)! (n k+) (n+)(n+). n+3. ( ) ( ) (n+3) p+. (n+)(n+) n+ Therefore the limiting distribution of P n as n goes to infinity is a triangular distribution with density f Pn (p) ( p). Note: the Polya urn for this problem generates a process that is identical to the following mixture of Bernoulli s Θ f Θ (θ) ( θ), 0 θ <. Let X i Bernoulli(Θ), then P n converges to Θ as n tends to infinity. Considering the general case of λ red balls and λ white balls to start with, the limiting distribution is where C(λ,λ ) is a function of λ and λ. f Θ (θ) C(λ,λ )θ λ ( θ) λ 6
7 5. Symmetric random walk. Let X n be a random walk defined by X 0 0, n X n Z i, where Z,Z,... are i.i.d. with P{Z } P{Z }. (a) Find P{X 0 0}. (b) Approximate P{ 0 X 00 0} using the central limit theorem. (c) Find P{X n k}. (a) Since the event {X 0 0} is equivalent to {Z Z 0 }, we have P{X 0 0} 0. (b) Since E(Z j ) 0 and E(Zj ), by the central limit theorem, { ( ) } 00 P{ 0 X 00 0} P Z i 00 (c) i i Q() Φ() P{X n k} P{(n+k)/ heads in n independent coin tosses} ( ) n n+k n for n k n with n+k even. 6. Absolute-value random walk. Consider the symmetric random walk X n in the previous problem. Define the absolute value random process Y n X n. (a) Find P{Y n k}. (b) Find P{max i<0 Y i 0 Y 0 0}. (a) If k 0 then P{Y n k} P{X n +k or X n k}. If k > 0 then P{Y n k} P{X n k}, while P{Y n 0} P{X n 0}. Thus ( n n (n+k)/)( ) k > 0, n k is even, n k 0 ( P{Y n k} n n n/)( ) k 0, n is even, n 0 0 otherwise. 7
8 (b) If Y 0 X 0 0 then there are only two sample paths with max i<0 X i 0 that is, Z Z Z 0 +,Z Z 0 or Z Z Z 0,Z Z 0 +. Since the total number of sample paths is ( 0 0) and all paths are equally likely, P { max Y i 0 Y 0 0 } ( i<0 0 ) Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process, i.e., Z,Z,... are i.i.d. N(0,). Define the process X n, n as X 0 0, and X n X n +Z n for n. (a) Is X n an independent increment process? Justify your answer. (b) Is X n a Gaussian process? Justify your answer. (c) Find the mean and autocorrelation functions of X n. (d) Specify the first order pdf of X n. (e) Specify the joint pdf of X 3,X 5, and X 8. (f) Find E(X 0 X,X,...,X 0 ). (g) Given X 4, X, and 0 X 3 4, find the minimum MSE estimate of X 4. (a) Yes. The increments X n, X n X n,...,x nk X nk are sums of different Z i random variables, and the Z i are IID. (b) Yes. Any set of samples of X n, n are obtained by a linear transformation of IID N(0,) random variables and therefore all nth order distributions of X n are jointly Gaussian (it is not sufficient to show that the random variable X n is Gaussian for each n). (c) We have [ n E[X n E Z i i n E[Z i i R X (n,n ) E[X n X n n n E Z i i j Z j n n E[Z i Z j i j min(n,n ) i min(n,n ). n 0 0, i E(Z i) (IID) 8
9 (d) As shown above, X n is Gaussian with mean zero and variance Var(X n ) E[Xn E [X n R X (n,n) 0 n. Thus, X n N(0,n). Cov(X n,x n ) E(X n X n ) E(X n )E(X n ) min(n,n ). Therefore, X n and X n are ( jointly Gaussian random variables ) with mean µ [0 0 T n and covariance matrix Σ min(n,n ). min(n,n ) n (e) X n, n is a zero mean Gaussian random process. Thus X 3 E[X 3 R X (3,3) R X (3,5) R X (3,8) X 5 N E[X 5, R X (5,3) R X (5,5) R X (5,8) X 8 E[X 8 R X (8,3) R X (8,5) R X (8,8) Substituting, we get X X 5 N 0, X (f) Since X n is an independent increment process, E(X 0 X,X,...,X 0 ) E(X 0 X 0 +X 0 X,X,...,X 0 ) E(X 0 X 0 X,X,...,X 0 )+E(X 0 X,X,...,X 0 ) E(X 0 X 0 )+X 0 0+X 0 X 0. (g) The MMSE estimate of X 4 given X x, X x and a X 3 b equals E[X 4 X x,x x,a X 3 b. Thus, the MMSE estimate is given by [ {X E X 4 x,x x,a X 3 b} [ {X E X +Z 3 +Z 4 x,x x,a X +Z 3 b} [ {a x x +E Z 3 Z 3 b x } +E[Z 4 ( since Z 3 is independent of (X,X ), and Z 4 is independent of (X,X,Z 3 ) ) [ {a x x +E Z 3 Z 3 b x }. 9
10 Plugging in the values, the required MMSE estimate is ˆX [ {Z3 4 +E Z 3 [,}. We have f Z3 (z 3 ) f Z3 Z 3 [,(z 3 ) P(Z 3 [,), z 3 [,, 0, otherwise [ {Z3 which is symmetric about z 3 0. Thus, E Z 3 [,} 0 and we have ˆX A random process. Let X n Z n +Z n for n, where Z 0,Z,Z,... are i.i.d. N(0,). (a) Find the mean and autocorrelation function of {X n }. (b) Is {X n } Gaussian? Justify your answer. (c) Find E(X 3 X,X ). (d) Find E(X 3 X ). (e) Is {X n } Markov? Justify your answer. (f) Is {X n } independent increment? Justify your answer. (a) E(X n ) E(Z n )+E(Z n ) 0. R X (m,n) E(X m X n ) E[(Z m +Z m )(Z n +Z n ) E[Zn, n m E[Zn +E[Z n, n m E[Zn, m n 0, otherwise, n m, n m 0, otherwise. (b) Since (X,...,X n ) is a linear transform of a GRV (Z 0,Z,...,Z n ), the process is Gaussian. (c) Since the process is Gaussian, the conditional expectation (MMSE estimate) is linear. Hence, E(X 3 X,X ) [ 0 [ [ X 3 (X X ). (d) Similarly, E(X 3 X ) (/)X. (e) Since E(X 3 X,X ) E(X 3 X ), the process is not Markov. (f) Since the process is not Markov, it is not independent increment. 0 X
11 9. Moving average process. Let X n Z n + Z n for n, where Z 0,Z,Z,... are i.i.d. N(0,). Find the mean and autocorrelation function of X n. E(X n ) E(Z n )+E(Z n ) 0. R X (m,n) E(X m X n ) [( )( ) E Z m +Z m Z n +Z n E[Z n, n m 4 E[Z n +E[Z n, n m E[Z n, m n 0, otherwise 5 4, n m, n m 0, otherwise. 0. Autoregressive process. Let X 0 0 and X n X n +Z n for n, where Z,Z,... are i.i.d. N(0,). Find the mean and autocorrelation function of X n. E[X n E[X n +E[Z n 4 E[X n + E[Z n... n E[X n E[Z 0. For n > m we can write Therefore, X n n mx m + n m Z m Z n +Z n n m n mx m + R X (n,m) E(X n X m ) (n m) E[X m, i0 iz n i. since X m and Z n i, i 0,...,n m, are independent and E[X m E[Z n i 0. To find E[X m consider Thus in general, E[X, E[X 4 E[X +E[Z 4 +,. E[X n 4 n ( 4 n). R X (n,m) E(X n X m ) n m 4 3 [ ( 4 )min{n,m}.
12 . Random binary waveform. In a digital communication channel the symbol is represented by the fixed duration rectangular pulse { for 0 t < g(t) 0 otherwise, and the symbol 0 is represented by g(t). The data transmitted over the channel is represented by the random process X(t) A k g(t k), for t 0, k0 where A 0,A,... are i.i.d random variables with { + w.p. A i w.p.. (a) Find its first and second order pmfs. (b) Find the mean and the autocorrelation function of the process X(t). (a) The first order pmf is p X(t) (x) P(X (t) x) ( ) P A k g(t k) x k0 P ( A t x ) P(A 0 x) IID {, x ± 0, otherwise. Now note that X(t ) and X(t ) are dependent only if t and t fall within the same time interval. Otherwise, they are independent. Thus, the second order pmf is p X(t )X(t )(x,y) P(X(t ) x,x (t ) y) ( ) P A k g(t k) x, A k g(t k) y k0 k0 P ( A t x,a t y ) { P(A0 x,a 0 y), t t P(A 0 x,a y), otherwise, t t & (x,y) (,),(, ) 4, t t & (x,y) (,),(, ),(,),(, ) 0, otherwise.
13 (b) For t 0, [ E[X(t) E A k g(t k) k0 g(t k)e[a k k0 0. For the autocorrelation R X (t,t ), we note once again that only if t and t fall within the same interval, will X(t ) be dependent on X(t ); if they do not fall in the same interval then they are independent from one another. Then, R X (t,t ) E[X(t )X(t ) g(t k)g(t k)e[a k k0 {, t t 0, otherwise. 3
UCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7
UCSD ECE50 Handout #4 Prof Young-Han Kim Wednesday, June 6, 08 Solutions to Exercise Set #7 Polya s urn An urn initially has one red ball and one white ball Let X denote the name of the first ball drawn
More informationUCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE 53 Handout #46 Prof. Young-Han Kim Thursday, June 5, 04 Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei). Discrete-time Wiener process. Let Z n, n 0 be a discrete time white
More informationUCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011
UCSD ECE53 Handout #40 Prof. Young-Han Kim Thursday, May 9, 04 Homework Set #8 Due: Thursday, June 5, 0. Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process,
More informationUCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)
UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable
More informationUCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011
UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, 2014 Homework Set #6 Due: Thursday, May 22, 2011 1. Linear estimator. Consider a channel with the observation Y = XZ, where the signal X and
More informationUCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE53 Handout #34 Prof Young-Han Kim Tuesday, May 7, 04 Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei) Linear estimator Consider a channel with the observation Y XZ, where the
More informationSolutions to Homework Set #6 (Prepared by Lele Wang)
Solutions to Homework Set #6 (Prepared by Lele Wang) Gaussian random vector Given a Gaussian random vector X N (µ, Σ), where µ ( 5 ) T and 0 Σ 4 0 0 0 9 (a) Find the pdfs of i X, ii X + X 3, iii X + X
More informationLecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs
Lecture Notes 3 Multiple Random Variables Joint, Marginal, and Conditional pmfs Bayes Rule and Independence for pmfs Joint, Marginal, and Conditional pdfs Bayes Rule and Independence for pdfs Functions
More informationChapter 6: Random Processes 1
Chapter 6: Random Processes 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More information6.041/6.431 Fall 2010 Quiz 2 Solutions
6.04/6.43: Probabilistic Systems Analysis (Fall 200) 6.04/6.43 Fall 200 Quiz 2 Solutions Problem. (80 points) In this problem: (i) X is a (continuous) uniform random variable on [0, 4]. (ii) Y is an exponential
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationUCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE 53 Handout #0 Prof. Young-Han Kim Thursday, April 4, 04 Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei). Time until the n-th arrival. Let the random variable N(t) be the number
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationIntroduction to Probability and Stochastic Processes I
Introduction to Probability and Stochastic Processes I Lecture 3 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark Slides
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationCourse: ESO-209 Home Work: 1 Instructor: Debasis Kundu
Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear
More informationECE Homework Set 3
ECE 450 1 Homework Set 3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3
More informationE X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.
E X A M Course code: Course name: Number of pages incl. front page: 6 MA430-G Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours Resources allowed: Notes: Pocket calculator,
More information2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).
Name M362K Final Exam Instructions: Show all of your work. You do not have to simplify your answers. No calculators allowed. There is a table of formulae on the last page. 1. Suppose X 1,..., X 1 are independent
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationStationary independent increments. 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process.
Stationary independent increments 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process. 2. If each set of increments, corresponding to non-overlapping collection of
More informationRandom Processes Why we Care
Random Processes Why we Care I Random processes describe signals that change randomly over time. I Compare: deterministic signals can be described by a mathematical expression that describes the signal
More informationECE534, Spring 2018: Solutions for Problem Set #5
ECE534, Spring 08: s for Problem Set #5 Mean Value and Autocorrelation Functions Consider a random process X(t) such that (i) X(t) ± (ii) The number of zero crossings, N(t), in the interval (0, t) is described
More informationUCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE53 Handout #7 Prof. Young-Han Kim Tuesday, May 6, 4 Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei). Neural net. Let Y = X + Z, where the signal X U[,] and noise Z N(,) are independent.
More informationChapter 4 : Expectation and Moments
ECE5: Analysis of Random Signals Fall 06 Chapter 4 : Expectation and Moments Dr. Salim El Rouayheb Scribe: Serge Kas Hanna, Lu Liu Expected Value of a Random Variable Definition. The expected or average
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationECE 450 Homework #3. 1. Given the joint density function f XY (x,y) = 0.5 1<x<2, 2<y< <x<4, 2<y<3 0 else
ECE 450 Homework #3 0. Consider the random variables X and Y, whose values are a function of the number showing when a single die is tossed, as show below: Exp. Outcome 1 3 4 5 6 X 3 3 4 4 Y 0 1 3 4 5
More information1.1 Review of Probability Theory
1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationTHE QUEEN S UNIVERSITY OF BELFAST
THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M
More informationLecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process
Lecture Notes 7 Stationary Random Processes Strict-Sense and Wide-Sense Stationarity Autocorrelation Function of a Stationary Process Power Spectral Density Continuity and Integration of Random Processes
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is
More informationMidterm Exam 1 Solution
EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2015 Kannan Ramchandran September 22, 2015 Midterm Exam 1 Solution Last name First name SID Name of student on your left:
More information2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.
CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook
More informationChapter 4 Random process. 4.1 Random process
Random processes - Chapter 4 Random process 1 Random processes Chapter 4 Random process 4.1 Random process 4.1 Random process Random processes - Chapter 4 Random process 2 Random process Random process,
More informationMultivariate distributions
CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density
More informationECE534, Spring 2018: Solutions for Problem Set #3
ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their
More informationMathematical Methods for Computer Science
Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge
More informationECE353: Probability and Random Processes. Lecture 18 - Stochastic Processes
ECE353: Probability and Random Processes Lecture 18 - Stochastic Processes Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu From RV
More informationUniversity of Illinois ECE 313: Final Exam Fall 2014
University of Illinois ECE 313: Final Exam Fall 2014 Monday, December 15, 2014, 7:00 p.m. 10:00 p.m. Sect. B, names A-O, 1013 ECE, names P-Z, 1015 ECE; Section C, names A-L, 1015 ECE; all others 112 Gregory
More informationFinal Examination Solutions (Total: 100 points)
Final Examination Solutions (Total: points) There are 4 problems, each problem with multiple parts, each worth 5 points. Make sure you answer all questions. Your answer should be as clear and readable
More informationExercises with solutions (Set D)
Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationECE Lecture #10 Overview
ECE 450 - Lecture #0 Overview Introduction to Random Vectors CDF, PDF Mean Vector, Covariance Matrix Jointly Gaussian RV s: vector form of pdf Introduction to Random (or Stochastic) Processes Definitions
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationfor valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I
Code: 15A04304 R15 B.Tech II Year I Semester (R15) Regular Examinations November/December 016 PROBABILITY THEY & STOCHASTIC PROCESSES (Electronics and Communication Engineering) Time: 3 hours Max. Marks:
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationUNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS
UNIT-2: MULTIPLE RANDOM VARIABLES & OPERATIONS In many practical situations, multiple random variables are required for analysis than a single random variable. The analysis of two random variables especially
More informationChapter 6 - Random Processes
EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process
More informationSolutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π
Solutions to Homework Set #5 (Prepared by Lele Wang). Neural net. Let Y X + Z, where the signal X U[,] and noise Z N(,) are independent. (a) Find the function g(y) that minimizes MSE E [ (sgn(x) g(y))
More informationEE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.
EE376A - Information Theory Final, Monday March 14th 216 Solutions Instructions: You have three hours, 3.3PM - 6.3PM The exam has 4 questions, totaling 12 points. Please start answering each question on
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4.1 (*. Let independent variables X 1,..., X n have U(0, 1 distribution. Show that for every x (0, 1, we have P ( X (1 < x 1 and P ( X (n > x 1 as n. Ex. 4.2 (**. By using induction
More informationRandom Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.
Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationSolutions to Set #2 Data Compression, Huffman code and AEP
Solutions to Set #2 Data Compression, Huffman code and AEP. Huffman coding. Consider the random variable ( ) x x X = 2 x 3 x 4 x 5 x 6 x 7 0.50 0.26 0. 0.04 0.04 0.03 0.02 (a) Find a binary Huffman code
More informationWe introduce methods that are useful in:
Instructor: Shengyu Zhang Content Derived Distributions Covariance and Correlation Conditional Expectation and Variance Revisited Transforms Sum of a Random Number of Independent Random Variables more
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationChapter 2: Random Variables
ECE54: Stochastic Signals and Systems Fall 28 Lecture 2 - September 3, 28 Dr. Salim El Rouayheb Scribe: Peiwen Tian, Lu Liu, Ghadir Ayache Chapter 2: Random Variables Example. Tossing a fair coin twice:
More informationt x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.
Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae
More informationAppendix A : Introduction to Probability and stochastic processes
A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of
More informationSolutions to Homework Set #3 (Prepared by Yu Xiang) Let the random variable Y be the time to get the n-th packet. Find the pdf of Y.
Solutions to Homework Set #3 (Prepared by Yu Xiang). Time until the n-th arrival. Let the random variable N(t) be the number of packets arriving during time (0,t]. Suppose N(t) is Poisson with pmf p N
More informationStochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno
Stochastic Processes M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno 1 Outline Stochastic (random) processes. Autocorrelation. Crosscorrelation. Spectral density function.
More informationECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172.
ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172. 1. Enter your name, student ID number, e-mail address, and signature in the space provided on this page, NOW! 2. This is a closed book exam.
More informationTest Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics
Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests
More informationConditional distributions (discrete case)
Conditional distributions (discrete case) The basic idea behind conditional distributions is simple: Suppose (XY) is a jointly-distributed random vector with a discrete joint distribution. Then we can
More informationFinal. Fall 2016 (Dec 16, 2016) Please copy and write the following statement:
ECE 30: Probabilistic Methods in Electrical and Computer Engineering Fall 06 Instructor: Prof. Stanley H. Chan Final Fall 06 (Dec 6, 06) Name: PUID: Please copy and write the following statement: I certify
More informationMATH 151, FINAL EXAM Winter Quarter, 21 March, 2014
Time: 3 hours, 8:3-11:3 Instructions: MATH 151, FINAL EXAM Winter Quarter, 21 March, 214 (1) Write your name in blue-book provided and sign that you agree to abide by the honor code. (2) The exam consists
More informationDiscrete Random Variables
CPSC 53 Systems Modeling and Simulation Discrete Random Variables Dr. Anirban Mahanti Department of Computer Science University of Calgary mahanti@cpsc.ucalgary.ca Random Variables A random variable is
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationi=1 k i=1 g i (Y )] = k f(t)dt and f(y) = F (y) except at possibly countably many points, E[g(Y )] = f(y)dy = 1, F(y) = y
Math 480 Exam 2 is Wed. Oct. 31. You are allowed 7 sheets of notes and a calculator. The exam emphasizes HW5-8, and Q5-8. From the 1st exam: The conditional probability of A given B is P(A B) = P(A B)
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More information[POLS 8500] Review of Linear Algebra, Probability and Information Theory
[POLS 8500] Review of Linear Algebra, Probability and Information Theory Professor Jason Anastasopoulos ljanastas@uga.edu January 12, 2017 For today... Basic linear algebra. Basic probability. Programming
More informationMultivariate probability distributions and linear regression
Multivariate probability distributions and linear regression Patrik Hoyer 1 Contents: Random variable, probability distribution Joint distribution Marginal distribution Conditional distribution Independence,
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationHomework 4 Solution, due July 23
Homework 4 Solution, due July 23 Random Variables Problem 1. Let X be the random number on a die: from 1 to. (i) What is the distribution of X? (ii) Calculate EX. (iii) Calculate EX 2. (iv) Calculate Var
More informationDiscrete Probability Refresher
ECE 1502 Information Theory Discrete Probability Refresher F. R. Kschischang Dept. of Electrical and Computer Engineering University of Toronto January 13, 1999 revised January 11, 2006 Probability theory
More informationECE 313: Conflict Final Exam Tuesday, May 13, 2014, 7:00 p.m. 10:00 p.m. Room 241 Everitt Lab
University of Illinois Spring 1 ECE 313: Conflict Final Exam Tuesday, May 13, 1, 7: p.m. 1: p.m. Room 1 Everitt Lab 1. [18 points] Consider an experiment in which a fair coin is repeatedly tossed every
More informationP 1.5 X 4.5 / X 2 and (iii) The smallest value of n for
DHANALAKSHMI COLLEGE OF ENEINEERING, CHENNAI DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING MA645 PROBABILITY AND RANDOM PROCESS UNIT I : RANDOM VARIABLES PART B (6 MARKS). A random variable X
More informationProbability reminders
CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so
More informationECE-340, Spring 2015 Review Questions
ECE-340, Spring 2015 Review Questions 1. Suppose that there are two categories of eggs: large eggs and small eggs, occurring with probabilities 0.7 and 0.3, respectively. For a large egg, the probabilities
More information4 Sums of Independent Random Variables
4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables
More informationDiscrete Distributions
A simplest example of random experiment is a coin-tossing, formally called Bernoulli trial. It happens to be the case that many useful distributions are built upon this simplest form of experiment, whose
More informationCDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random
CDA6530: Performance Models of Computers and Networks Chapter 2: Review of Practical Random Variables Definition Random variable (RV)X (R.V.) X: A function on sample space X: S R Cumulative distribution
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example
More informationIE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.
Closed book and notes. 60 minutes. A summary table of some univariate continuous distributions is provided. Four Pages. In this version of the Key, I try to be more complete than necessary to receive full
More informationTwelfth Problem Assignment
EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationM378K In-Class Assignment #1
The following problems are a review of M6K. M7K In-Class Assignment # Problem.. Complete the definition of mutual exclusivity of events below: Events A, B Ω are said to be mutually exclusive if A B =.
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationLecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157
Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationProbability and Statistics
Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph
More information