UCSD ECE153 Handout #27 Prof. Young-Han Kim Tuesday, May 6, Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei)

Similar documents
Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π

UCSD ECE153 Handout #34 Prof. Young-Han Kim Tuesday, May 27, Solutions to Homework Set #6 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE153 Handout #30 Prof. Young-Han Kim Thursday, May 15, Homework Set #6 Due: Thursday, May 22, 2011

ENGG2430A-Homework 2

Solutions to Homework Set #6 (Prepared by Lele Wang)

UCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei)

More than one variable

Final Examination Solutions (Total: 100 points)

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

2 (Statistics) Random variables

18.440: Lecture 26 Conditional expectation

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx

Lecture Notes Special: Using Neural Networks for Mean-Squared Estimation. So: How can we evaluate E(X Y )? What is a neural network? How is it useful?

Let X and Y denote two random variables. The joint distribution of these random

ECE302 Exam 2 Version A April 21, You must show ALL of your work for full credit. Please leave fractions as fractions, but simplify them, etc.

UCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},

Bivariate Distributions

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011

STA 256: Statistics and Probability I

UCSD ECE250 Handout #24 Prof. Young-Han Kim Wednesday, June 6, Solutions to Exercise Set #7

Lecture 4: Least Squares (LS) Estimation

Chapter 4 Multiple Random Variables

UCSD ECE250 Handout #20 Prof. Young-Han Kim Monday, February 26, Solutions to Exercise Set #7

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

Homework 5 Solutions

ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

Lecture 16: Hierarchical models and miscellanea

SDS 321: Introduction to Probability and Statistics

Appendix A : Introduction to Probability and stochastic processes

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

conditional cdf, conditional pdf, total probability theorem?

Chapter 5 Joint Probability Distributions

ESTIMATION THEORY. Chapter Estimation of Random Variables

UC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Problem Set 8 Fall 2007

1 Random Variable: Topics

Solutions to Homework Set #3 (Prepared by Yu Xiang) Let the random variable Y be the time to get the n-th packet. Find the pdf of Y.

Random Signals and Systems. Chapter 3. Jitendra K Tugnait. Department of Electrical & Computer Engineering. Auburn University.

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

STT 441 Final Exam Fall 2013

MAS113 Introduction to Probability and Statistics. Proofs of theorems

Bivariate Distributions. Discrete Bivariate Distribution Example

MATH 38061/MATH48061/MATH68061: MULTIVARIATE STATISTICS Solutions to Problems on Random Vectors and Random Sampling. 1+ x2 +y 2 ) (n+2)/2

Continuous r.v practice problems

Lecture 11. Probability Theory: an Overveiw

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

18.440: Lecture 28 Lectures Review

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

A DARK GREY P O N T, with a Switch Tail, and a small Star on the Forehead. Any

STAT/MATH 395 PROBABILITY II

Review: mostly probability and some statistics

Formulas for probability theory and linear models SF2941

ECE Lecture #9 Part 2 Overview

MAS113 Introduction to Probability and Statistics. Proofs of theorems

Bivariate distributions

Problem Set 7 Due March, 22

Lecture 4 : Random variable and expectation

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

Math Spring Practice for the final Exam.

1 Solution to Problem 2.1

General Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

ECE 4400:693 - Information Theory

Algorithms for Uncertainty Quantification

Functions of two random variables. Conditional pairs

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Chapter 5,6 Multiple RandomVariables

STAT 430/510: Lecture 16

Continuous Random Variables

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

Bivariate Paired Numerical Data

EE4601 Communication Systems

Probability review. September 11, Stoch. Systems Analysis Introduction 1

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

ECE 302: Probabilistic Methods in Engineering

Introduction to Probability and Stocastic Processes - Part I

Bivariate Distributions

2 Functions of random variables

CMPSCI 240: Reasoning Under Uncertainty

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

HW5 Solutions. (a) (8 pts.) Show that if two random variables X and Y are independent, then E[XY ] = E[X]E[Y ] xy p X,Y (x, y)

Probability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.

p. 6-1 Continuous Random Variables p. 6-2

Statistics 351 Probability I Fall 2006 (200630) Final Exam Solutions. θ α β Γ(α)Γ(β) (uv)α 1 (v uv) β 1 exp v }

4 Derivations of the Discrete-Time Kalman Filter

Joint Distribution of Two or More Random Variables

A Course Material on. Probability and Random Processes

MATH/STAT 3360, Probability Sample Final Examination Model Solutions

CHAPTER 4 MATHEMATICAL EXPECTATION. 4.1 Mean of a Random Variable

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

STAT 515 MIDTERM 2 EXAM November 14, 2018

ECSE B Solutions to Assignment 8 Fall 2008

Lecture 2: Repetition of probability theory and statistics

A Probability Review

Transcription:

UCSD ECE53 Handout #7 Prof. Young-Han Kim Tuesday, May 6, 4 Solutions to Homework Set #5 (Prepared by TA Fatemeh Arbabjolfaei). Neural net. Let Y = X + Z, where the signal X U[,] and noise Z N(,) are independent. (a) Find the function g(y) that minimizes MSE = E [ (sgn(x) g(y)) ], where sgn(x) = { x + x >. (b) Plot g(y) vs. y. Solution: The minimum MSE is achieved when g(y) = E(sgn(X) Y). We have g(y) = E(sgn(X) Y = y) = To find the conditional pdf of X given Y, we use f X Y (x y) = f Y X(y x)f X (x) f Y (y) Since X and Z are independent,, where f X (x) = sgn(x)f X Y (x y)dx. { X otherwise. f Y X (y x) = f Z (y x) Y {X = x} N(x,). To find f Y (y) we integrate f Y X (y x)f X (x) over x: f Y (y) = = f Y X (y x)f X (x)dx = π e (x y) dx = = (Q( y ) Q( y +)). Combining the above results, we get π e ( (y x) dx π e (x y) dx ) e (x y) dx π

g(y) = sgn(x)f X Y (x y)dx = = = f Y (y) ( π e (x y) dx+ sgn(x) Q( y )+Q( y) Q( y +) Q( y ) Q( y +). e (y x) π f Y (y) dx ) e (x y) dx π The plot is shown below. Note the sigmoidal shape corresponding to the common neural network activation function..5.5 g(y).5.5 8 6 4 4 6 8 y. Additive shot noise channel. Consider an additive noise channel Y = X + Z, where the signal X N(,), and the noise Z {X = x} N(,x ), i.e., the noise power of increases linearly with the signal squared. (a) Find E(Z ). (b) Find the best linear MSE estimate of X given Y. Solution: (a) Since Z {X = x} N(,x ), Therefore, E(Z X) = Var(Z X = x) = E(Z X = x) E(Z X = x) = x. E(Z ) = E(E(Z X)) = E(X ) =.

(b) From the the best linear estimate formula, Here we have E(X) =, ˆX = Cov(X,Y) σy (Y E(Y))+E(X). E(Y) = E(X +Z) = E(X)+E(Z) = E(X)+E(E(Z X)) = +E() =, E(XZ) = E(E(XZ X)) = E(XE(Z X)) = E(X ) = E() =, σ Y = E(Y ) (E(Y)) = E((X +Z) ) = E(X )+E(Z )+E(XZ) = ++ =, Cov(X,Y) = E((X E(X))(Y E(Y))) = E(XY) = E(X(X +Z)) = E(X )+E(XZ) = + =. Using all of the above, we get ˆX = Y. 3. Estimation vs. detection. Let the signal { +, with probability X =, with probability, and the noise Z Unif[,] be independent random variables. Their sum Y = X + Z is observed. (a) Find the best MSE estimate of X given Y and its MSE. (b) Now suppose we use a decoder to decide whether X = + or X = so that the probability of error is minimized. Find the optimal decoder and its probability of error. Compare the optimal decoder s MSE to the minimum MSE. Solution: (a) We can easily find the piecewise constant density of Y 4 y f Y (y) = 8 < y 3 otherwise 3

The conditional probabilities of X given Y are 3 y < P{X = + Y = y} = y + + < y +3 3 y < P{X = Y = y} = y + + < y +3 Thus the best MSE estimate is 3 Y < g(y) = E(X Y) = Y + + + < Y +3 The minimum mean square error is E Y (Var(X Y)) = E Y (E(X Y) (E(X Y)) ) = E( g(y) ) = E(g(Y) ) = ( = = 3 3 8 dy + 3 8 dy g(y) f Y (y)dy +3 4 dy + + 8 dy = 4 4 =. 8 dy ) (b) The optimal decoder is given by the MAP rule. The a posteriori pmf of X was found in part (a). Thus the MAP rule reduces to 3 y D(y) = ± < y < + + + < y +3 Since either value can be chosen for D(y) in the center range of Y, a symmetrical decoder is sufficient, i.e., { y < D(y) = + y The probability of decoding error is P{d(Y) X} = P{X =, Y } + P{X =, Y < } = P{X = Y }P{Y } + P{X = Y < }P{Y < } = 4 + 4 = 4. 4

If we use the decoder (detector) as an estimator, its MSE is E ( (d(y) X) ) = 3 4 + 4 =. This MSE is twice that of the minimum mean square error estimator. 5

Solutions to Additional Exercises. Orthogonality. Let ˆX be the minimum MSE estimate of X given Y. (a) Show that for any function g(y), E((X ˆX)g(Y)) =, i.e., the error (X ˆX) and g(y) are orthogonal. (b) Show that Var(X) = E(Var(X Y))+Var( ˆX). Provide a geometric interpretation for this result. Solution: (a) We use iterated expectation and the fact that E(g(Y) Y) = g(y): ( E (X ˆX)g(Y) ) [ ( = E E (X ˆX)g(Y) Y )] (b) First we write and = E[E((X E(X Y))g(Y)) Y)] = E(g(Y)E((X E(X Y)) Y)) = E(g(Y)(E(X Y) E(X Y))) =. E(Var(X Y)) = E(X ) E((E(X Y)) ), Var(E(X Y)) = E((E(X Y)) ) (E(E(X Y))) Adding the two terms completes the proof. = E((E(X Y)) ) (E(X)). Interpretation: IfweviewX,E(X Y),andX E(X Y)asvectorswith norm Var(X), Var(E(X Y)), and E(Var(X Y)), respectively, then this result provides a Pythagorastheorem, wherethesignal, theerror, andtheestimatearethesidesofarighttriangle (estimate and error being orthogonal).. Let X and Y be two random variables. Let Z = X +Y and let W = X Y. Find the best linear estimate of W given Z as a function of E(X),E(Y),σ X,σ Y,ρ XY and Z. Solution: By the theorem of MMSE linear estimate, we have Ŵ = Cov(W,Z) σz (Z E(Z))+E(W). 6

Here we have E(W) = E(X) E(Y), E(Z) = E(X)+E(Y), σ Z = σ X +σ Y +ρ XY σ X σ Y, Cov(W,Z) = E(WZ) E(W)E(Z) = E((X Y)(X +Y)) (E(X) E(Y))(E(X)+E(Y)) = E(X ) E(Y ) (E(X)) +(E(Y)) = σ X σ Y. So the best linear estimate of W given Z is Ŵ = σ X σ Y σ X +σ Y +ρ XYσ X σ Y (Z E(X) E(Y))+E(X) E(Y). 3. Let X and Y be two random variables with joint pdf { x+y, x, y, f(x,y) =, otherwise. (a) Find the MMSE estimator of X given Y. (b) Find the corresponding MSE. (c) Find the pdf of Z = E(X Y). (d) Find the linear MMSE estimator of X given Y. (e) Find the corresponding MSE. Solution: (a) We first calculate the marginal pdf of Y, by a direct integration. For y < or y >, we integrate over. For y, f Y (y) = (x+y)dx = y +. Note that the limits of the integration are derived from the definition of the joint pdf. Thus, { y + f Y (y) =, y, otherwise. Now we can calculate the conditional pdf f X Y (x y) = f XY(x,y) f Y (y) for x,y. Therefore, for y hence E[X Y = y] = E[X Y] = xf X Y (x y)dx = 3 + Y +Y. 7 = x+y +y x +yx y + dx = 3 + y +y,

(b) The MSE is given by E(Var(X Y)) = EX E((E(X Y)) ) ( = x x+ ) dx = 5 ( 3 + ln(3) 44.757. ) ( 3 + y +y = ln(3) 44 ) ( y + ) dy (c) Since E[X Y = y] = 3 +y = +y + 6(+y), we have 5 9 E[X Y = y] 3 for y. From the part (a), we know Z = E[X Y] = /3+Y/ /+Y. Thus, 5 9 Z 3. We first find the cdf F Z (z) of Z and then differentiate it to get the pdf. Consider { 3 F Z (z) = P{Z z} = P + Y { = P Y < 3z 3 6z } +Y z }. For 3z 3 6z, i.e., 5 9 z 3, we have F Z (z) = 3z 3 6z By differentiating with respect to z, we get f Z (z) = ( 3z 3 6z + ) = 8(z ) 3 for 5 9 z 3. Otherwise f Z(z) =. (d) The best linear MMSE estimate is Here we have E(X) = E(Y) = { = P{+3Y 3z +6zY} = P Y f Y (y)dx = 3z 3 6z d dz ( ) y + dx (3z ) = 3 6z 6( z) ˆX = Cov(X,Y) σy (Y E(Y))+E(X). x(x+y)dydx = y(y + )dy = 7, Cov(X,Y) = E(XY) E(X)E(Y) = σ Y = E(Y ) (E(Y)) = x(x+ )dydx = 7, y (y + )dy ( 7 3( z) 3z } 3 6z xy(x+y)dydx 7 7 = 44, ) = 44. 8

So the best linear MMSE estimate is ˆX = ( Y 7 ) + 7. (e) The MSE of linear estimate is Here by symmetry, σ X = σ Y = 44. Thus, MSE = σx Cov (X,Y) σy. MSE = σ X Cov (X,Y) σ Y = 44 +.67. We can check that LMSE > MSE. 4. Additive-noise channel with signal dependent noise. Consider the channel with correlated signal X and noise Z and observation Y = X +Z, where µ X =, µ Z =, σ X = 4, σ Z = 9, ρ X,Z = 3 8. Find the best MSE linear estimate of X given Y. Solution: The best linear MMSE estimate is given by the formula Here we have ˆX = Cov(X,Y) σy (Y E(Y))+E(X). E(Y) = E(X)+E(Z) =, σ Y = 4σ X +σ Z +4ρ XZ σ X σ Z = 6, Cov(X,Y) = E(XY) E(X)E(Y) = E(X +XZ) So the best linear MMSE estimate is = (σ X +µ X)+(ρ XZ σ X σ Z +µ X µ Z ) = 3 4. ˆX = 3 3 (Y )+ = 64 64 Y + 9 3. 9