EE 4TM4: Digital Communications II Information Measures

Size: px
Start display at page:

Download "EE 4TM4: Digital Communications II Information Measures"

Transcription

1 EE 4TM4: Digital Commuicatios II Iformatio Measures Defiitio : The etropy H(X) of a discrete radom variable X is defied by We also write H(p) for the above quatity. Lemma : H(X) 0. H(X) = x X Proof: 0 p(x) implies log(/p(x)) 0. p(x) log p(x). Defiitio 2: The relative etropy or Kullback Leibler distace betwee two probability mass fuctios p(x) ad q(x) is defied as D(p q) = x X p(x)log p(x) q(x). Theorem : Jese s iequality: If f is a covex fuctio ad X is a radom variable, the Ef(X) f(ex). () Moreover, if f is strictly covex, the equality i () implies that X = EX with probability, i.e., X is a costat. Theorem 2: Let p(x), q(x), x X, be two probability mass fuctios. The with equality if ad oly if p(x) = q(x) for all x. D(p q) 0 Proof: Let A = {x : p(x) > 0} be the support set of p(x). The D(p q) = x A x A p(x)log p(x) q(x) p(x)log q(x) p(x) log x A = log x Aq(x) p(x) q(x) p(x) log x X q(x) = log = 0,

2 2 where the first iequality follows from Jeses iequality. Sice log t is a strictly cocave fuctio of t, we have the first iequality becomes equality if ad oly if q(x)/p(x) = everywhere, i.e., p(x) = q(x). Hece we have D(p q) = 0 if ad oly if p(x) = q(x) for all x. Defiitio 3: The type P x (or empirical probability distributio) of a sequece x,x 2,,x is the relative proportio of occurreces of each symbol of X, i.e., P x (a) = N(a x )/ for all a X, where N(a x ) is the umber of times the symbol a occurs i the sequece x X. Defiitio 4: Let P deote the set of types with deomiator. Defiitio 5: If P P, the the set of sequeces of legth ad type P is called the type class of P, deoted T(P), i.e., T(P) = {x X : P x = P}. Theorem 3: P = ( + X ) X (+) X. Theorem 4: If X,X 2,,X are draw i.i.d. accordig to Q(x), the the probability of x depeds oly o its type ad is give by Q (x ) = 2 (H(P x )+D(P x Q)). Proof: Q (x ) = Q(x i ) i= = a X Q(a) N(a x ) = a X Q(a) P x (a) = a X 2 P x (a)logq(a) = a X = 2 2 (P x (a)logq(a) P x (a)logp x (a)+p x (a)logp x (a)) a X ( P x (a)log P x (a) Q(a) +P x (a)logp x (a)) = 2 ( D(P x Q) H(P x )). Corollary : If x is i the type class of Q, the Theorem 5: For ay type P P, Q (x ) = 2 H(Q). (+) X 2H(P) T(P) 2 H(P). Proof: Note that the exact size of T(P) is give by ( )! T(P) = = P(a ),P(a 2 ),,P(a X ) (P(a ))!(P(a 2 ))! (P(a X ))!

3 3 We first prove the upper boud. Sice a type class must have probability, we have P (T(P)) = x T(P) P (x ) = x T(P) 2 H(P) = T(P) 2 H(P). Thus T(P) 2 H(P). Now for the lower boud. We first prove that the type class T(P) has the highest probability amog all type classes uder the probability distributio P, i.e., P (T(P)) P (T(ˆP)), for all ˆP P. We lower boud the ratio of probabilities, P (T(P)) P (T(ˆP)) = T(P) T(ˆP) a X = a X P(a)P(a) P(a) ˆP(a) ( P(a ),P(a 2),,P(a X ) ( ˆP(a ),ˆP(a 2),,ˆP(a X ) = a X ) ) (ˆP(a))! (P(a))! P(a)(P(a) ˆP(a)). a X P(a)P(a) a X P(a) ˆP(a) Now usig the simple boud (easy to prove by separately cosiderig the cases m ad m < ) m!! m, we obtai P (T(P)) P (T(ˆP)) a X = a X (P(a)) ˆP(a) P(a) P(a) (P(a) ˆP(a)) ( ˆP(a) P(a)) = ( ˆP(a) a X a X P(a)) = ( ) =. Hece P (T(P)) P (T(ˆP)). The lower boud ow follows easily from this result, sice = P (T(Q)) Q P Q P max Q P (T(Q)) = P (T(P)) Q P (+) X P (T(P)) = (+) X P (x ) x T(P) = (+) X x T(P) 2 H(P) = (+) X T(P) 2 H(P).

4 4 Theorem 6: For ay P P ad ay distributio Q, the probability of the type class T(P) uder Q is 2 D(P Q) to first order i the expoet. More precisely, Proof: We have Usig the bouds o T(P), we have + X 2 D(P Q) Q (T(P)) 2 D(P Q). Q (T(P)) = x T(P) = x T(P) Q (x ) 2 (D(P Q)+H(P)) = T(P) 2 (D(P Q)+H(P)). (+) X 2 D(P Q) Q (T(P)) 2 D(P Q). Theorem 7: Let X,X 2,,X be i.id. Q(x). The Pr{D(P x Q) > } 2 ( X log(+) ). Proof: Pr{D(P x Q) > } = Q (T(P)) P:D(P Q)> P:D(P Q)> P:D(P Q)> 2 D(P Q) 2 (+) X 2 = 2 ( X log(+) ). Theorem 8: Saov s theorem: Let X,X 2,,X be i.i.d. Q(x). Let E P be a set of probability distributios. The Q (E) = Q (E P ) (+) X 2 D(P Q), where P = argmid(p Q) is the distributio i E that is closest to Q i relative etropy. If, i additio, the P E set E is the closure of its iterior, the logq (E) D(P Q).

5 5 Proof: We first prove the upper boud: Q (E) = Q (T(P)) 2 D(P Q) max 2 D(P Q) = 2 mip E P D(P Q) 2 mip E D(P Q) 2 D(P Q) = (+) X 2 D(P Q). Note that P eed ot be a member of P. We ow come to the lower boud, for which we eed a ice set E, so that for all large, we ca fid a distributio i E P which is close to P. If we ow assume that E is the closure of its iterior (thus the iterior must be o-empty), the sice P is dese i the set of all distributios, it follows that E P is o-empty for all 0 for some 0. We ca the fid a sequece of distributios P such that P E P ad D(P Q) D(P Q). For each 0, Q (E) = Q (T(P)) Cosequetly, Q (T(P )) (+) X 2 D(P Q). limif logq (E) limif( X log(+) Combiig this with the upper boud establishes the theorem. We ca summarize the basic theorems cocerig types i four equatios: P (+) X, Q (x ) = 2 (D(P x Q)+H(P x )), T(P) 2 H(P), Q (T(P)) 2 D(P Q). D(P Q)) = D(P Q). Theorem 9: Weak Law of Large Numbers: Let X,X 2, be a sequece of idepedet radom variables havig a commo distributio Q, ad let E[X i ] = µ. The for ay > 0 { P X +X 2 + +X } µ 0 as

6 6 Proof: Let Y = X+X2+ +X by Chebyshev s iequality µ. Note that E[Y ] = 0 ad Var(Y ) = σ2, where σ2 = Var(X i ). Therefore, P{ Y } σ2 2 0 as Remark: This proof shows that the probability goes to zero at least as fast as. I fact, it is possible to derive a tighter boud. Without loss of geerality, we assume E[X i ] = 0. Note that P{ Y } = P{ Y 4 4 } E[Y 4 ] 4. Sice E(X +X 2 + +X ) 4 = E(X 4 )+3( )(E(X 2 )) 2, it follows that the probability goes to zero at least as fast as 2. by Remark: Accordig to Saov s theorem, the probability goes to zero expoetially fast with the expoet give mi P: x xp(x) µ D(P Q). Theorem 0: Cetral Limit Theorem: Let X,X 2, be a sequece of idepedet, idetically distributed radom variables, each with mea µ ad variace σ 2. The the distributio of X +X 2 + +X µ σ teds to the stadard ormal as. That is, { X +X 2 + +X µ P σ as. } a a e x2 /2 dx 2π Coectio betwee the cetral limit theorem ad the divergece: LetX,X 2, be a sequece of i.i.d. Beroulli radom variables with parameter p, i.e., P(X i = ) = p ad P(X i = 0) = p. Note that E[X i ] = p ad Var(X i ) = E[Xi 2] E[X i] 2 = p p 2 = p( p). Accordig to the cetral limit theorem, { X +X 2 + +X p } P [a,b] b e x2 /2 dx. p( p) 2π Note that here we are essetially coutig the total probability of type classes such that X +X 2 + +X [p+a p( p),p+b p( p) ], i.e., the empirical p x [p+a p( p)/,p+b p( p)/ ]. Now cosider the Taylor expasio of D(q p) at the eighborhood of p. Note that So which equals 0 whe q = p, ad D(q p) = qlog q p +( q)log q p. d dq D(q p) = log q q log p p, d 2 dq 2D(q p) = q( q), a

7 7 which equals p( p) whe q = p. Therefore, for q = p+x p( p)/, we have D(q p) 2p( p) (x p( p)/ ) 2 = x2 2. Cosequetly, the probability of the type class with empirical p x p+x p( p)/ is approximately e x2 2 = e x 2 /2, which gives the right expoet i the cetral limit theorem. Note that i the limit, oe ca replace the sum of the probabilities of type classes by itegral to recover the cetral limit theorem (the costat factor / 2π requires a more accurate approximatio usig Stirlig s formula). For the o-biary case, oe ca also establish a coectio betwee divergece ad the multi-dimesioal cetral limit theorem. Theorem : H(X) log X, where X deotes the umber of elemets i the rage of X, with equality if ad oly if X has a uiform distributio over X. Proof: Let u(x) = X be the uiform probability mass fuctio over X, ad let p(x) be the probability mass fuctio for X. The Hece by the o-egativity of relative etropy, D(p u) = p(x)log p(x) u(x) = log X H(X). 0 D(p u) = log X H(X). This result ca also be proved usig Lagragia multiplier. Coditioal etropy: H(Y X) = x X,y Yp(x,y)logp(y x) = x Xp(x) y Yp(y x)logp(y x) = x X Coditioal etropy: residual ucertaity Joit etropy: H(X,Y) = x X,y Y p(x,y)logp(x,y). We have H(X,Y) = H(X)+H(Y X) = H(Y)+H(X Y). More geerally, H(X ) = H(X )+H(X 2 X )+ +H(X X,,X ). p(x)h(y X = x). Mutual iformatio: I(X;Y) = (x,y) X Y p(x,y)log p(x,y) p(x)p(y) = H(X) H(X Y) = H(Y) H(Y X) = H(X)+H(Y) H(X,Y).

8 8 Note that I(X;Y) = D(p(x,y) p(x)p(y)). Therefore, I(X;Y) 0 with equality if ad oly if X ad Y are idepedet. Sice I(X;Y) 0, it follows that H(Y X) H(Y) H(X,Y) H(X)+H(Y). Moreover,H(g(X)) H(X) with equality ifg is oe-to-oe over the support ofx, i.e., the set{x X : p(x) > 0}. This is because H(X,g(X)) = H(g(X) X)+ H(X) = H(X) ad H(X,g(X)) = H(g(X)) + H(X g(x)) H(g(X)). Theorem 2: Fao s iequality: Suppose we wish to estimate a radom variable X with a distributio p(x). We observe a radom variable Y which is related to X by the coditioal distributio p(x y). From Y, we calculate a fuctio g(y) = ˆX, which is a estimate of X. We wish to boud the probability that ˆX X. We observe that X Y ˆX forms a Markov chai. Defie the probability of error P e = P{ ˆX X}. The H(P e )+P e log( X ) H(X Y), where H(P e ) = P e logp e ( P e )log( P e ). This iequality ca be weakeed to +P e log X H(X Y). Proof: Let E = if ˆX X ad E = 0 if ˆX = X. Note that H(E,X Y) = H(X Y)+H(E X,Y) = H(X Y). O the other had, H(E,X Y) = H(E Y)+H(X E,Y) H(E)+H(X E,Y) = H(P e )+H(X E,Y) = H(P e )+P(E = 0)H(X Y,E = 0)+P(E = )H(X Y,E = ) H(P e )+( P e )0+P e log( X ). Coditioal mutual iformatio: I(X;Y Z) = z Z p(z)i(x;y Z = z) = z Zp(z) x X,y Y p(x, y z) log p(x, y z) p(x z)p(y z).

9 9 Note that I(X;Y Z) = z Zp(z)D(p(x,y z) p(x z)p(y z)) = D(p(x,y z) p(x z)p(y z) p(z)). Also ote that I(X;Y Z) = H(X Z) H(X Y,Z) = H(Y Z) H(Y X,Z) = H(X Z)+H(Y Z) H(X,Y Z). The coditioal mutual iformatio I(X;Y Z) is oegative ad is equal to zero if ad oly if X ad Y are coditioally idepedet give Z, i.e., X Z Y form a Markov chai. Theorem 3: Data processig iequality: If X Y Z form a Markov chai, the I(X;Z) I(X;Y). Cosequetly, for ay fuctio g, I(X;g(Y)) I(X;Y). Proof: To prove the data processig iequality, we use the chai rule to expad I(X;Y,Z) i two ways as I(X;Y,Z) = I(X;Y)+I(X;Z Y) = I(X;Y) = I(X;Z)+I(X;Y Z) I(X;Z). Note that ulike etropy, o geeral iequality relatioship exists betwee the coditioal mutual iformatio I(X;Y Z) ad the mutual iformatio I(X;Y). There are, however two importat special cases. If X ad Z are idepedet, the This is because I(X;Y Z) = I(X;Y,Z) I(X;Y). If Z X Y form a Markov chai, the This is because I(X;Y Z) I(Z,X;Y) = I(X;Y). Typical Sequeces I(X;Y Z) I(X;Y). I(X;Y Z) I(X;Y). Let X,X 2, be a sequece of idepedet ad idetically distributed radom variables. The by the (weak) law of large umbers, for each x X, N(x x ) p(x) i probability. Thus, with high probability, the radom empirical pmf N(x X ) does ot deviate much from the true pmf p(x). For X p(x) ad (0,), defie the set of -typical -sequeces x (or the typical set i short) as T () (X) = {x : N(x x ) p(x) p(x) for all x X}.

10 0 Note that the typical set ca be viewed as the uio of type classes whose type is close to p(x). Also ote that p(x) = 0 implies N(x x ) = 0 because otherwise D(p x p) =. Lemma 2: Typical Average Lemma: Let x T () (X). The for ay oegative fuctio g(x) o X, ( )E(g(X)) g(x i ) (+)E(g(X)). Typical sequeces satisfy the followig properties: i= ) Let p(x ) = i= p X(x i ). The for each x T (X), 2 (H(X)+δ()) p(x ) 2 (H(X) δ()), where δ() = H(X). This follows by the typical average lemma with g(x) = logp(x). 2) The cardiality of the typical set is upper bouded as T () 2 (H(X)+δ()). This ca be show by summig the lower boud i property over the typical set. 3) If X,X 2, are i.i.d. with X i p X (x i ), the by the LLN, lim P{X T () } =. This result ca also be proved usig Theorem 7 i Lecture. Note that the size of the typical set is i geeral egligible compared with X. However, it captures almost all probability. 4) The cardiality of the typical set is lower bouded as T () ( )2 (H(X) δ()) for sufficietly large. This follows by property 3 ad the upper boud i property. Explai these properties from the perspective of method of types. Joitly Typical Sequeces Let (X,Y) p X,Y (x,y). The set of joitly -typical -sequeces is defied as { T () (X,Y) = (x,y ) : N(x,y x,y ) p X,Y (x,y) p X,Y(x,y) for all (x,y) X Y { Also defie the set of coditioally -typical sequeces as T () (Y x ) = y : (x,y ) T () properties of typical sequeces ca be exteded to joitly typical sequeces as follows. ) Let (x,y ) T () (X,Y) ad p(x,y ) = i= p X,Y(x i,y i ). The (a) x T () (X) ad y T () (Y), (b) 2 (+)H(X) p(x ) 2 ( )H(X) ad 2 (+)H(Y) p(y ) 2 ( )H(Y), }. } (X,Y). The (c) 2 (+)H(X Y) p(x y ) 2 ( )H(X Y) ad 2 (+)H(Y X) p(y x ) 2 ( )H(Y X), (d) 2 (+)H(X,Y) p(x,y ) 2 ( )H(X,Y ). 2) T () (X,Y) 2 (+)H(X,Y). 3) If p(x,y ) = i= p X,Y(x i,y i ), the lim P{(X,Y ) T () (X,Y)} =.

11 4) T () (X,Y) ( )2 ( )H(X,Y ) for sufficietly large. 5) For every x X, we have T () (Y x ) 2 (+)H(Y X). 6) Let X p X (x) ad Y = g(x). Let x T () i [ : ]. The followig property deserves special attetio. (X). The y T () (Y x ) if ad oly if y i = g(x i ) for Lemma 3: Coditioal Typicality Lemma: Let (X,Y) p(x,y). Suppose that x T () (X) ad Y p(y x ) = i= p Y X(y i x i ). The, for every >, The proof of this lemma follows by the LLN. lim P{(x,Y ) T () (X,Y)} =. The coditioal typicality lemma implies the followig additioal property of joitly typical sequeces. 5) If x T () (X) ad <, the for sufficietly large, Joit Typicality for a Triple of Radom Variables T () (Y x ) ( )2 ( )H(X Y). Let (X,Y,Z) p(x,y,z). The set of joitly -typical (x,y,z ) sequeces is defied as T () (X,Y,Z) = {(x,y,z ) : N(x,y,z x,y,z ) p(x,y,z) p(x,y,z) for all (x,y,z) X Y Z}. Suppose that (x,y,z ) T () (X,Y,Z) ad p(x,y,z ) = i= p X,Y,Z(x i,y i,z i ). The ) x T () ad (y,z ) T () (Y,Z), 2) p(x,y,z ) 2 H(X,Y,Z), 3) p(x,y z ) 2 H(X,Y Z), 4) T () (X y,z ) 2 (H(X Y,Z)+δ()), ad 5) if (y,z ) T () (Y,Z) ad <, the for sufficietly large, T () (X y,z ) 2 (H(X Y,Z) δ()). The followig two-part lemma will be used i may achievability proofs of codig theorem. Lemma 4: Joit Typicality Lemma: Let (X,Y,Z) p(x,y,z) ad <. The there exists δ() > 0 that teds to zero as 0 such that the followig statemets hold: ) If ( x,ỹ ) is a pair of arbitrary sequeces ad Z i= p Z X( z i x i ), the 2) If (x,y ) T () P{( x,ỹ, Z ) T () (X,Y,Z)} 2 (I(Y;Z X) δ()). ad Z i= p Z X( z i x i ), the for sufficietly large, P{(x,y, Z ) T () (X,Y,Z)} 2 (I(Y;Z X)+δ()).

12 2 Proof: To prove the first statemet, cosider P{( x,ỹ, Z T () (X,Y,Z))} = Similarly, for every sufficietly large, P{(x,y, Z ) T () which proves the secod statemet. z T () (Z x,ỹ ) p( z x ) T () (Z x,ỹ ) 2 (H(Z X) H(Z X)) 2 (H(Z X,Y )+H(Z X,Y )) 2 (H(Z X) H(Z X)) = 2 (I(Y;Z X) δ()). (X,Y,Z)} T () (Z x,y ) 2 (H(Z X)+H(Z X)) ( )2 (H(Z X,Y ) H(Z X,Y)) 2 (H(Z X)+H(Z X)) = 2 (I(Y;Z X)+δ()), REFERENCES [] T. Cover ad J. A. Thomas, Elemets of Iformatio Theory, 2d Editio. Hoboke, NJ: Wiley, [2] A. El Gamal ad Y.-H. Kim, Network Iformatio Theory. Cambridge, U.K.: Cambridge Uiversity Press, 20.

The method of types. PhD short course Information Theory and Statistics Siena, September, Mauro Barni University of Siena

The method of types. PhD short course Information Theory and Statistics Siena, September, Mauro Barni University of Siena PhD short course Iformatio Theory ad Statistics Siea, 15-19 September, 2014 The method of types Mauro Bari Uiversity of Siea Outlie of the course Part 1: Iformatio theory i a utshell Part 2: The method

More information

Entropy and Ergodic Theory Lecture 5: Joint typicality and conditional AEP

Entropy and Ergodic Theory Lecture 5: Joint typicality and conditional AEP Etropy ad Ergodic Theory Lecture 5: Joit typicality ad coditioal AEP 1 Notatio: from RVs back to distributios Let (Ω, F, P) be a probability space, ad let X ad Y be A- ad B-valued discrete RVs, respectively.

More information

Lecture 7: October 18, 2017

Lecture 7: October 18, 2017 Iformatio ad Codig Theory Autum 207 Lecturer: Madhur Tulsiai Lecture 7: October 8, 207 Biary hypothesis testig I this lecture, we apply the tools developed i the past few lectures to uderstad the problem

More information

Distribution of Random Samples & Limit theorems

Distribution of Random Samples & Limit theorems STAT/MATH 395 A - PROBABILITY II UW Witer Quarter 2017 Néhémy Lim Distributio of Radom Samples & Limit theorems 1 Distributio of i.i.d. Samples Motivatig example. Assume that the goal of a study is to

More information

EE 4TM4: Digital Communications II Probability Theory

EE 4TM4: Digital Communications II Probability Theory 1 EE 4TM4: Digital Commuicatios II Probability Theory I. RANDOM VARIABLES A radom variable is a real-valued fuctio defied o the sample space. Example: Suppose that our experimet cosists of tossig two fair

More information

This section is optional.

This section is optional. 4 Momet Geeratig Fuctios* This sectio is optioal. The momet geeratig fuctio g : R R of a radom variable X is defied as g(t) = E[e tx ]. Propositio 1. We have g () (0) = E[X ] for = 1, 2,... Proof. Therefore

More information

Lecture 15: Strong, Conditional, & Joint Typicality

Lecture 15: Strong, Conditional, & Joint Typicality EE376A/STATS376A Iformatio Theory Lecture 15-02/27/2018 Lecture 15: Strog, Coditioal, & Joit Typicality Lecturer: Tsachy Weissma Scribe: Nimit Sohoi, William McCloskey, Halwest Mohammad I this lecture,

More information

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)].

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)]. Probability 2 - Notes 0 Some Useful Iequalities. Lemma. If X is a radom variable ad g(x 0 for all x i the support of f X, the P(g(X E[g(X]. Proof. (cotiuous case P(g(X Corollaries x:g(x f X (xdx x:g(x

More information

Spring Information Theory Midterm (take home) Due: Tue, Mar 29, 2016 (in class) Prof. Y. Polyanskiy. P XY (i, j) = α 2 i 2j

Spring Information Theory Midterm (take home) Due: Tue, Mar 29, 2016 (in class) Prof. Y. Polyanskiy. P XY (i, j) = α 2 i 2j Sprig 206 6.44 - Iformatio Theory Midterm (take home) Due: Tue, Mar 29, 206 (i class) Prof. Y. Polyaskiy Rules. Collaboratio strictly prohibited. 2. Write rigorously, prove all claims. 3. You ca use otes

More information

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

This exam contains 19 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam. Probability ad Statistics FS 07 Secod Sessio Exam 09.0.08 Time Limit: 80 Miutes Name: Studet ID: This exam cotais 9 pages (icludig this cover page) ad 0 questios. A Formulae sheet is provided with the

More information

Lecture 19: Convergence

Lecture 19: Convergence Lecture 19: Covergece Asymptotic approach I statistical aalysis or iferece, a key to the success of fidig a good procedure is beig able to fid some momets ad/or distributios of various statistics. I may

More information

Chapter 5. Inequalities. 5.1 The Markov and Chebyshev inequalities

Chapter 5. Inequalities. 5.1 The Markov and Chebyshev inequalities Chapter 5 Iequalities 5.1 The Markov ad Chebyshev iequalities As you have probably see o today s frot page: every perso i the upper teth percetile ears at least 1 times more tha the average salary. I other

More information

Information Theory and Statistics Lecture 4: Lempel-Ziv code

Information Theory and Statistics Lecture 4: Lempel-Ziv code Iformatio Theory ad Statistics Lecture 4: Lempel-Ziv code Łukasz Dębowski ldebowsk@ipipa.waw.pl Ph. D. Programme 203/204 Etropy rate is the limitig compressio rate Theorem For a statioary process (X i)

More information

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f. Lecture 5 Let us give oe more example of MLE. Example 3. The uiform distributio U[0, ] o the iterval [0, ] has p.d.f. { 1 f(x =, 0 x, 0, otherwise The likelihood fuctio ϕ( = f(x i = 1 I(X 1,..., X [0,

More information

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1. Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio

More information

Refinement of Two Fundamental Tools in Information Theory

Refinement of Two Fundamental Tools in Information Theory Refiemet of Two Fudametal Tools i Iformatio Theory Raymod W. Yeug Istitute of Network Codig The Chiese Uiversity of Hog Kog Joit work with Siu Wai Ho ad Sergio Verdu Discotiuity of Shao s Iformatio Measures

More information

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering CEE 5 Autum 005 Ucertaity Cocepts for Geotechical Egieerig Basic Termiology Set A set is a collectio of (mutually exclusive) objects or evets. The sample space is the (collectively exhaustive) collectio

More information

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1 8. The cetral limit theorems 8.1. The cetral limit theorem for i.i.d. sequeces. ecall that C ( is N -separatig. Theorem 8.1. Let X 1, X,... be i.i.d. radom variables with EX 1 = ad EX 1 = σ (,. Suppose

More information

LECTURE 8: ASYMPTOTICS I

LECTURE 8: ASYMPTOTICS I LECTURE 8: ASYMPTOTICS I We are iterested i the properties of estimators as. Cosider a sequece of radom variables {, X 1}. N. M. Kiefer, Corell Uiversity, Ecoomics 60 1 Defiitio: (Weak covergece) A sequece

More information

1 Convergence in Probability and the Weak Law of Large Numbers

1 Convergence in Probability and the Weak Law of Large Numbers 36-752 Advaced Probability Overview Sprig 2018 8. Covergece Cocepts: i Probability, i L p ad Almost Surely Istructor: Alessadro Rialdo Associated readig: Sec 2.4, 2.5, ad 4.11 of Ash ad Doléas-Dade; Sec

More information

Shannon s noiseless coding theorem

Shannon s noiseless coding theorem 18.310 lecture otes May 4, 2015 Shao s oiseless codig theorem Lecturer: Michel Goemas I these otes we discuss Shao s oiseless codig theorem, which is oe of the foudig results of the field of iformatio

More information

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1 EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum

More information

Lecture 6: Source coding, Typicality, and Noisy channels and capacity

Lecture 6: Source coding, Typicality, and Noisy channels and capacity 15-859: Iformatio Theory ad Applicatios i TCS CMU: Sprig 2013 Lecture 6: Source codig, Typicality, ad Noisy chaels ad capacity Jauary 31, 2013 Lecturer: Mahdi Cheraghchi Scribe: Togbo Huag 1 Recap Uiversal

More information

STAT Homework 1 - Solutions

STAT Homework 1 - Solutions STAT-36700 Homework 1 - Solutios Fall 018 September 11, 018 This cotais solutios for Homework 1. Please ote that we have icluded several additioal commets ad approaches to the problems to give you better

More information

Lecture 8: Convergence of transformations and law of large numbers

Lecture 8: Convergence of transformations and law of large numbers Lecture 8: Covergece of trasformatios ad law of large umbers Trasformatio ad covergece Trasformatio is a importat tool i statistics. If X coverges to X i some sese, we ofte eed to check whether g(x ) coverges

More information

Lecture 11: Channel Coding Theorem: Converse Part

Lecture 11: Channel Coding Theorem: Converse Part EE376A/STATS376A Iformatio Theory Lecture - 02/3/208 Lecture : Chael Codig Theorem: Coverse Part Lecturer: Tsachy Weissma Scribe: Erdem Bıyık I this lecture, we will cotiue our discussio o chael codig

More information

Information Theory Tutorial Communication over Channels with memory. Chi Zhang Department of Electrical Engineering University of Notre Dame

Information Theory Tutorial Communication over Channels with memory. Chi Zhang Department of Electrical Engineering University of Notre Dame Iformatio Theory Tutorial Commuicatio over Chaels with memory Chi Zhag Departmet of Electrical Egieerig Uiversity of Notre Dame Abstract A geeral capacity formula C = sup I(; Y ), which is correct for

More information

Lecture 14: Graph Entropy

Lecture 14: Graph Entropy 15-859: Iformatio Theory ad Applicatios i TCS Sprig 2013 Lecture 14: Graph Etropy March 19, 2013 Lecturer: Mahdi Cheraghchi Scribe: Euiwoog Lee 1 Recap Bergma s boud o the permaet Shearer s Lemma Number

More information

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Convergence of random variables. (telegram style notes) P.J.C. Spreij Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space

More information

Lecture 7: Properties of Random Samples

Lecture 7: Properties of Random Samples Lecture 7: Properties of Radom Samples 1 Cotiued From Last Class Theorem 1.1. Let X 1, X,...X be a radom sample from a populatio with mea µ ad variace σ

More information

ECE 6980 An Algorithmic and Information-Theoretic Toolbox for Massive Data

ECE 6980 An Algorithmic and Information-Theoretic Toolbox for Massive Data ECE 6980 A Algorithmic ad Iformatio-Theoretic Toolbo for Massive Data Istructor: Jayadev Acharya Lecture # Scribe: Huayu Zhag 8th August, 017 1 Recap X =, ε is a accuracy parameter, ad δ is a error parameter.

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

Entropy Rates and Asymptotic Equipartition

Entropy Rates and Asymptotic Equipartition Chapter 29 Etropy Rates ad Asymptotic Equipartitio Sectio 29. itroduces the etropy rate the asymptotic etropy per time-step of a stochastic process ad shows that it is well-defied; ad similarly for iformatio,

More information

Lecture 20: Multivariate convergence and the Central Limit Theorem

Lecture 20: Multivariate convergence and the Central Limit Theorem Lecture 20: Multivariate covergece ad the Cetral Limit Theorem Covergece i distributio for radom vectors Let Z,Z 1,Z 2,... be radom vectors o R k. If the cdf of Z is cotiuous, the we ca defie covergece

More information

4.1 Data processing inequality

4.1 Data processing inequality ECE598: Iformatio-theoretic methods i high-dimesioal statistics Sprig 206 Lecture 4: Total variatio/iequalities betwee f-divergeces Lecturer: Yihog Wu Scribe: Matthew Tsao, Feb 8, 206 [Ed. Mar 22] Recall

More information

Mathematics 170B Selected HW Solutions.

Mathematics 170B Selected HW Solutions. Mathematics 17B Selected HW Solutios. F 4. Suppose X is B(,p). (a)fidthemometgeeratigfuctiom (s)of(x p)/ p(1 p). Write q = 1 p. The MGF of X is (pe s + q), sice X ca be writte as the sum of idepedet Beroulli

More information

Supplementary Materials for Statistical-Computational Phase Transitions in Planted Models: The High-Dimensional Setting

Supplementary Materials for Statistical-Computational Phase Transitions in Planted Models: The High-Dimensional Setting Supplemetary Materials for Statistical-Computatioal Phase Trasitios i Plated Models: The High-Dimesioal Settig Yudog Che The Uiversity of Califoria, Berkeley yudog.che@eecs.berkeley.edu Jiamig Xu Uiversity

More information

INFORMATION THEORY AND STATISTICS. Jüri Lember

INFORMATION THEORY AND STATISTICS. Jüri Lember INFORMATION THEORY AND STATISTICS Lecture otes ad exercises Sprig 203 Jüri Lember Literature:. T.M. Cover, J.A. Thomas "Elemets of iformatio theory", Wiley, 99 ja 2006; 2. Yeug, Raymod W. "A first course

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS MASSACHUSTTS INSTITUT OF TCHNOLOGY 6.436J/5.085J Fall 2008 Lecture 9 /7/2008 LAWS OF LARG NUMBRS II Cotets. The strog law of large umbers 2. The Cheroff boud TH STRONG LAW OF LARG NUMBRS While the weak

More information

Sequences and Series of Functions

Sequences and Series of Functions Chapter 6 Sequeces ad Series of Fuctios 6.1. Covergece of a Sequece of Fuctios Poitwise Covergece. Defiitio 6.1. Let, for each N, fuctio f : A R be defied. If, for each x A, the sequece (f (x)) coverges

More information

Approximations and more PMFs and PDFs

Approximations and more PMFs and PDFs Approximatios ad more PMFs ad PDFs Saad Meimeh 1 Approximatio of biomial with Poisso Cosider the biomial distributio ( b(k,,p = p k (1 p k, k λ: k Assume that is large, ad p is small, but p λ at the limit.

More information

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization ECE 90 Lecture 4: Maximum Likelihood Estimatio ad Complexity Regularizatio R Nowak 5/7/009 Review : Maximum Likelihood Estimatio We have iid observatios draw from a ukow distributio Y i iid p θ, i,, where

More information

Lecture 4. We also define the set of possible values for the random walk as the set of all x R d such that P(S n = x) > 0 for some n.

Lecture 4. We also define the set of possible values for the random walk as the set of all x R d such that P(S n = x) > 0 for some n. Radom Walks ad Browia Motio Tel Aviv Uiversity Sprig 20 Lecture date: Mar 2, 20 Lecture 4 Istructor: Ro Peled Scribe: Lira Rotem This lecture deals primarily with recurrece for geeral radom walks. We preset

More information

Lecture 2: Concentration Bounds

Lecture 2: Concentration Bounds CSE 52: Desig ad Aalysis of Algorithms I Sprig 206 Lecture 2: Cocetratio Bouds Lecturer: Shaya Oveis Ghara March 30th Scribe: Syuzaa Sargsya Disclaimer: These otes have ot bee subjected to the usual scrutiy

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Itroductio to Probability ad Statistics Lecture 23: Cotiuous radom variables- Iequalities, CLT Puramrita Sarkar Departmet of Statistics ad Data Sciece The Uiversity of Texas at Austi www.cs.cmu.edu/

More information

Learning Theory: Lecture Notes

Learning Theory: Lecture Notes Learig Theory: Lecture Notes Kamalika Chaudhuri October 4, 0 Cocetratio of Averages Cocetratio of measure is very useful i showig bouds o the errors of machie-learig algorithms. We will begi with a basic

More information

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2. SAMPLE STATISTICS A radom sample x 1,x,,x from a distributio f(x) is a set of idepedetly ad idetically variables with x i f(x) for all i Their joit pdf is f(x 1,x,,x )=f(x 1 )f(x ) f(x )= f(x i ) The sample

More information

Quick Review of Probability

Quick Review of Probability Quick Review of Probability Berli Che Departmet of Computer Sciece & Iformatio Egieerig Natioal Taiwa Normal Uiversity Refereces: 1. W. Navidi. Statistics for Egieerig ad Scietists. Chapter & Teachig Material.

More information

Quick Review of Probability

Quick Review of Probability Quick Review of Probability Berli Che Departmet of Computer Sciece & Iformatio Egieerig Natioal Taiwa Normal Uiversity Refereces: 1. W. Navidi. Statistics for Egieerig ad Scietists. Chapter 2 & Teachig

More information

Advanced Stochastic Processes.

Advanced Stochastic Processes. Advaced Stochastic Processes. David Gamarik LECTURE 2 Radom variables ad measurable fuctios. Strog Law of Large Numbers (SLLN). Scary stuff cotiued... Outlie of Lecture Radom variables ad measurable fuctios.

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Basics of Probability Theory (for Theory of Computation courses)

Basics of Probability Theory (for Theory of Computation courses) Basics of Probability Theory (for Theory of Computatio courses) Oded Goldreich Departmet of Computer Sciece Weizma Istitute of Sciece Rehovot, Israel. oded.goldreich@weizma.ac.il November 24, 2008 Preface.

More information

Unbiased Estimation. February 7-12, 2008

Unbiased Estimation. February 7-12, 2008 Ubiased Estimatio February 7-2, 2008 We begi with a sample X = (X,..., X ) of radom variables chose accordig to oe of a family of probabilities P θ where θ is elemet from the parameter space Θ. For radom

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 6 9/24/2008 DISCRETE RANDOM VARIABLES AND THEIR EXPECTATIONS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 6 9/24/2008 DISCRETE RANDOM VARIABLES AND THEIR EXPECTATIONS MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 6 9/24/2008 DISCRETE RANDOM VARIABLES AND THEIR EXPECTATIONS Cotets 1. A few useful discrete radom variables 2. Joit, margial, ad

More information

An Introduction to Randomized Algorithms

An Introduction to Randomized Algorithms A Itroductio to Radomized Algorithms The focus of this lecture is to study a radomized algorithm for quick sort, aalyze it usig probabilistic recurrece relatios, ad also provide more geeral tools for aalysis

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit theorems Throughout this sectio we will assume a probability space (Ω, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

Rademacher Complexity

Rademacher Complexity EECS 598: Statistical Learig Theory, Witer 204 Topic 0 Rademacher Complexity Lecturer: Clayto Scott Scribe: Ya Deg, Kevi Moo Disclaimer: These otes have ot bee subjected to the usual scrutiy reserved for

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/5.070J Fall 203 Lecture 3 9//203 Large deviatios Theory. Cramér s Theorem Cotet.. Cramér s Theorem. 2. Rate fuctio ad properties. 3. Chage of measure techique.

More information

Lecture 13: Maximum Likelihood Estimation

Lecture 13: Maximum Likelihood Estimation ECE90 Sprig 007 Statistical Learig Theory Istructor: R. Nowak Lecture 3: Maximum Likelihood Estimatio Summary of Lecture I the last lecture we derived a risk (MSE) boud for regressio problems; i.e., select

More information

The standard deviation of the mean

The standard deviation of the mean Physics 6C Fall 20 The stadard deviatio of the mea These otes provide some clarificatio o the distictio betwee the stadard deviatio ad the stadard deviatio of the mea.. The sample mea ad variace Cosider

More information

MAT1026 Calculus II Basic Convergence Tests for Series

MAT1026 Calculus II Basic Convergence Tests for Series MAT026 Calculus II Basic Covergece Tests for Series Egi MERMUT 202.03.08 Dokuz Eylül Uiversity Faculty of Sciece Departmet of Mathematics İzmir/TURKEY Cotets Mootoe Covergece Theorem 2 2 Series of Real

More information

Asymptotic Coupling and Its Applications in Information Theory

Asymptotic Coupling and Its Applications in Information Theory Asymptotic Couplig ad Its Applicatios i Iformatio Theory Vicet Y. F. Ta Joit Work with Lei Yu Departmet of Electrical ad Computer Egieerig, Departmet of Mathematics, Natioal Uiversity of Sigapore IMS-APRM

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals

More information

1 Introduction to reducing variance in Monte Carlo simulations

1 Introduction to reducing variance in Monte Carlo simulations Copyright c 010 by Karl Sigma 1 Itroductio to reducig variace i Mote Carlo simulatios 11 Review of cofidece itervals for estimatig a mea I statistics, we estimate a ukow mea µ = E(X) of a distributio by

More information

Lecture 3 : Random variables and their distributions

Lecture 3 : Random variables and their distributions Lecture 3 : Radom variables ad their distributios 3.1 Radom variables Let (Ω, F) ad (S, S) be two measurable spaces. A map X : Ω S is measurable or a radom variable (deoted r.v.) if X 1 (A) {ω : X(ω) A}

More information

Kernel density estimator

Kernel density estimator Jauary, 07 NONPARAMETRIC ERNEL DENSITY ESTIMATION I this lecture, we discuss kerel estimatio of probability desity fuctios PDF Noparametric desity estimatio is oe of the cetral problems i statistics I

More information

Maximum Likelihood Estimation and Complexity Regularization

Maximum Likelihood Estimation and Complexity Regularization ECE90 Sprig 004 Statistical Regularizatio ad Learig Theory Lecture: 4 Maximum Likelihood Estimatio ad Complexity Regularizatio Lecturer: Rob Nowak Scribe: Pam Limpiti Review : Maximum Likelihood Estimatio

More information

Lecture 7: Channel coding theorem for discrete-time continuous memoryless channel

Lecture 7: Channel coding theorem for discrete-time continuous memoryless channel Lecture 7: Chael codig theorem for discrete-time cotiuous memoryless chael Lectured by Dr. Saif K. Mohammed Scribed by Mirsad Čirkić Iformatio Theory for Wireless Commuicatio ITWC Sprig 202 Let us first

More information

Ada Boost, Risk Bounds, Concentration Inequalities. 1 AdaBoost and Estimates of Conditional Probabilities

Ada Boost, Risk Bounds, Concentration Inequalities. 1 AdaBoost and Estimates of Conditional Probabilities CS8B/Stat4B Sprig 008) Statistical Learig Theory Lecture: Ada Boost, Risk Bouds, Cocetratio Iequalities Lecturer: Peter Bartlett Scribe: Subhrasu Maji AdaBoost ad Estimates of Coditioal Probabilities We

More information

Math 451: Euclidean and Non-Euclidean Geometry MWF 3pm, Gasson 204 Homework 3 Solutions

Math 451: Euclidean and Non-Euclidean Geometry MWF 3pm, Gasson 204 Homework 3 Solutions Math 451: Euclidea ad No-Euclidea Geometry MWF 3pm, Gasso 204 Homework 3 Solutios Exercises from 1.4 ad 1.5 of the otes: 4.3, 4.10, 4.12, 4.14, 4.15, 5.3, 5.4, 5.5 Exercise 4.3. Explai why Hp, q) = {x

More information

Exercise 4.3 Use the Continuity Theorem to prove the Cramér-Wold Theorem, Theorem. (1) φ a X(1).

Exercise 4.3 Use the Continuity Theorem to prove the Cramér-Wold Theorem, Theorem. (1) φ a X(1). Assigmet 7 Exercise 4.3 Use the Cotiuity Theorem to prove the Cramér-Wold Theorem, Theorem 4.12. Hit: a X d a X implies that φ a X (1) φ a X(1). Sketch of solutio: As we poited out i class, the oly tricky

More information

Lecture 3: August 31

Lecture 3: August 31 36-705: Itermediate Statistics Fall 018 Lecturer: Siva Balakrisha Lecture 3: August 31 This lecture will be mostly a summary of other useful expoetial tail bouds We will ot prove ay of these i lecture,

More information

Lecture Notes for Analysis Class

Lecture Notes for Analysis Class Lecture Notes for Aalysis Class Topological Spaces A topology for a set X is a collectio T of subsets of X such that: (a) X ad the empty set are i T (b) Uios of elemets of T are i T (c) Fiite itersectios

More information

IIT JAM Mathematical Statistics (MS) 2006 SECTION A

IIT JAM Mathematical Statistics (MS) 2006 SECTION A IIT JAM Mathematical Statistics (MS) 6 SECTION A. If a > for ad lim a / L >, the which of the followig series is ot coverget? (a) (b) (c) (d) (d) = = a = a = a a + / a lim a a / + = lim a / a / + = lim

More information

Joint Probability Distributions and Random Samples. Jointly Distributed Random Variables. Chapter { }

Joint Probability Distributions and Random Samples. Jointly Distributed Random Variables. Chapter { } UCLA STAT A Applied Probability & Statistics for Egieers Istructor: Ivo Diov, Asst. Prof. I Statistics ad Neurology Teachig Assistat: Neda Farziia, UCLA Statistics Uiversity of Califoria, Los Ageles, Sprig

More information

ECE 901 Lecture 13: Maximum Likelihood Estimation

ECE 901 Lecture 13: Maximum Likelihood Estimation ECE 90 Lecture 3: Maximum Likelihood Estimatio R. Nowak 5/7/009 The focus of this lecture is to cosider aother approach to learig based o maximum likelihood estimatio. Ulike earlier approaches cosidered

More information

Asymptotic distribution of products of sums of independent random variables

Asymptotic distribution of products of sums of independent random variables Proc. Idia Acad. Sci. Math. Sci. Vol. 3, No., May 03, pp. 83 9. c Idia Academy of Scieces Asymptotic distributio of products of sums of idepedet radom variables YANLING WANG, SUXIA YAO ad HONGXIA DU ollege

More information

Stat 421-SP2012 Interval Estimation Section

Stat 421-SP2012 Interval Estimation Section Stat 41-SP01 Iterval Estimatio Sectio 11.1-11. We ow uderstad (Chapter 10) how to fid poit estimators of a ukow parameter. o However, a poit estimate does ot provide ay iformatio about the ucertaity (possible

More information

Lecture 23: Minimal sufficiency

Lecture 23: Minimal sufficiency Lecture 23: Miimal sufficiecy Maximal reductio without loss of iformatio There are may sufficiet statistics for a give problem. I fact, X (the whole data set) is sufficiet. If T is a sufficiet statistic

More information

Stat410 Probability and Statistics II (F16)

Stat410 Probability and Statistics II (F16) Some Basic Cocepts of Statistical Iferece (Sec 5.) Suppose we have a rv X that has a pdf/pmf deoted by f(x; θ) or p(x; θ), where θ is called the parameter. I previous lectures, we focus o probability problems

More information

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i

More information

ON POINTWISE BINOMIAL APPROXIMATION

ON POINTWISE BINOMIAL APPROXIMATION Iteratioal Joural of Pure ad Applied Mathematics Volume 71 No. 1 2011, 57-66 ON POINTWISE BINOMIAL APPROXIMATION BY w-functions K. Teerapabolar 1, P. Wogkasem 2 Departmet of Mathematics Faculty of Sciece

More information

Solutions of Homework 2.

Solutions of Homework 2. 1 Solutios of Homework 2. 1. Suppose X Y with E(Y )

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013 Large Deviatios for i.i.d. Radom Variables Cotet. Cheroff boud usig expoetial momet geeratig fuctios. Properties of a momet

More information

Lecture 2. The Lovász Local Lemma

Lecture 2. The Lovász Local Lemma Staford Uiversity Sprig 208 Math 233A: No-costructive methods i combiatorics Istructor: Ja Vodrák Lecture date: Jauary 0, 208 Origial scribe: Apoorva Khare Lecture 2. The Lovász Local Lemma 2. Itroductio

More information

ECE 330:541, Stochastic Signals and Systems Lecture Notes on Limit Theorems from Probability Fall 2002

ECE 330:541, Stochastic Signals and Systems Lecture Notes on Limit Theorems from Probability Fall 2002 ECE 330:541, Stochastic Sigals ad Systems Lecture Notes o Limit Theorems from robability Fall 00 I practice, there are two ways we ca costruct a ew sequece of radom variables from a old sequece of radom

More information

The Central Limit Theorem

The Central Limit Theorem Chapter The Cetral Limit Theorem Deote by Z the stadard ormal radom variable with desity 2π e x2 /2. Lemma.. Ee itz = e t2 /2 Proof. We use the same calculatio as for the momet geeratig fuctio: exp(itx

More information

Agnostic Learning and Concentration Inequalities

Agnostic Learning and Concentration Inequalities ECE901 Sprig 2004 Statistical Regularizatio ad Learig Theory Lecture: 7 Agostic Learig ad Cocetratio Iequalities Lecturer: Rob Nowak Scribe: Aravid Kailas 1 Itroductio 1.1 Motivatio I the last lecture

More information

Lecture 18: Sampling distributions

Lecture 18: Sampling distributions Lecture 18: Samplig distributios I may applicatios, the populatio is oe or several ormal distributios (or approximately). We ow study properties of some importat statistics based o a radom sample from

More information

Singular Continuous Measures by Michael Pejic 5/14/10

Singular Continuous Measures by Michael Pejic 5/14/10 Sigular Cotiuous Measures by Michael Peic 5/4/0 Prelimiaries Give a set X, a σ-algebra o X is a collectio of subsets of X that cotais X ad ad is closed uder complemetatio ad coutable uios hece, coutable

More information

It is always the case that unions, intersections, complements, and set differences are preserved by the inverse image of a function.

It is always the case that unions, intersections, complements, and set differences are preserved by the inverse image of a function. MATH 532 Measurable Fuctios Dr. Neal, WKU Throughout, let ( X, F, µ) be a measure space ad let (!, F, P ) deote the special case of a probability space. We shall ow begi to study real-valued fuctios defied

More information

Exponential Families and Bayesian Inference

Exponential Families and Bayesian Inference Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where

More information

PAPER : IIT-JAM 2010

PAPER : IIT-JAM 2010 MATHEMATICS-MA (CODE A) Q.-Q.5: Oly oe optio is correct for each questio. Each questio carries (+6) marks for correct aswer ad ( ) marks for icorrect aswer.. Which of the followig coditios does NOT esure

More information

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality

More information

5 Birkhoff s Ergodic Theorem

5 Birkhoff s Ergodic Theorem 5 Birkhoff s Ergodic Theorem Amog the most useful of the various geeralizatios of KolmogorovâĂŹs strog law of large umbers are the ergodic theorems of Birkhoff ad Kigma, which exted the validity of the

More information

L = n i, i=1. dp p n 1

L = n i, i=1. dp p n 1 Exchageable sequeces ad probabilities for probabilities 1996; modified 98 5 21 to add material o mutual iformatio; modified 98 7 21 to add Heath-Sudderth proof of de Fietti represetatio; modified 99 11

More information

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3 MATH 337 Sequeces Dr. Neal, WKU Let X be a metric space with distace fuctio d. We shall defie the geeral cocept of sequece ad limit i a metric space, the apply the results i particular to some special

More information

The natural exponential function

The natural exponential function The atural expoetial fuctio Attila Máté Brookly College of the City Uiversity of New York December, 205 Cotets The atural expoetial fuctio for real x. Beroulli s iequality.....................................2

More information