S1 Notation and Assumptions

Size: px
Start display at page:

Download "S1 Notation and Assumptions"

Transcription

1 Statistica Siica: Supplemet Robust-BD Estimatio ad Iferece for Varyig-Dimesioal Geeral Liear Models Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag Uiversity of Wiscosi-Madiso Supplemetary Material S Notatio ad Assumptios For a matrix M, its eigevalues, miimum eigevalue, maximum eigevalue ad trace are labeledbyλ j (M),λ mi (M),λ max (M)adtr(M)respectively. Let M = sup x = Mx = {λ max (M T M)} /2 be the matrix L 2 orm; let M F = {tr(m T M)} /2 be the Frobeius orm. See Golub ad Va Loa (996) for details. Throughout the proof, C is used as a geeric fiite costat. We first impose some regularity coditios, which are ot the weakest possible but facilitate the techical derivatios. Coditio A: A. sup β <. A. X = max j p X j is bouded almost surely. A2. E( X XT ) exists ad is osigular. A4. There isalargeeoughopesubset ofr p + which cotaisthe trueparameter poit β, such that F ( X T β) is bouded almost surely for all β i the subset. A5. w( ) is a bouded fuctio. Assume that ψ(r) is a bouded, odd fuctio, ad twice differetiable, such that ψ (r), ψ (r)r, ψ (r), ψ (r)r ad ψ (r)r 2 arebouded; V( ) >, V (2) ( ) is cotiuous. The matrix H is positive defiite, with eigevalues uiformly bouded away from. A6. q (4) ( ) is cotiuous, ad q (2) ( ) <. G (3) ( ) is cotiuous. A7. F( ) is mootoe ad a bijectio, F (3) ( ) is cotiuous, ad F () ( ). Coditio B:

2 S2 Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag B5. The matrices Ω ad H are positive defiite, with eigevalues uiformly bouded away from. Also, H Ω is bouded away from. Coditio C: C4. Thereis a largeeoughope subset ofr p + whichcotaisthe trueparameter poit β, such that A β = g, ad F ( X T β) is bouded almost surely for all β i the subset. Coditio D: D5. The eigevalues of H are uiformly bouded away from. Also, H /2 Ω /2 is bouded away from. S2 Proofs of Mai Results Proof of Theorem We follow the idea of the proof i Fa ad Peg (24). Let r = p / ad ũ = (u,u,...,u p ) T R p +. It suffices to show that for ay give ǫ >, there exists a sufficietly large costat C ǫ such that, for large we have { } P if l ( β +r ũ ) > l ( β ) ǫ. (S2.) ũ =C ǫ This implies that with probability at least ǫ, there exists a local miimizer β of l ( β) i the ball { β + r ũ : ũ C ǫ } such that β β = O P (r ). To show (S2.), cosider l ( β +r ũ ) l ( β ) = where ũ = C ǫ. {ρ q (Y i,f ( X T i ( β +r ũ )))w(x i ) ρ q (Y i,f ( X T i β ))w(x i )} I, (S2.2) By Taylor s expasio, I = I, +I,2 +I,3, where I, = r / p (Y i ; X T i β )w(x i ) X T iũ, I,2 = r 2 /(2) p 2 (Y i ; X T i β ; )w(x i )( X T iũ) 2, (S2.3)

3 Robust-BD Estimatio ad Iferece S3 I,3 = r/(6) 3 p 3 (Y i ; X T i β )w(x i )( X T iũ) 3 for β located betwee β ; ad β ; +r ũ. Hece I, r p (Y i ; X T i β ; )w(x i ) X i ũ = O P (r p /) ũ. (S2.4) For I,2 i (S2.3), I,2 = r2 2 E{p 2 (Y i ; X T i β ; )w(x i )( X T iũ) 2 } + r2 [p 2 2 (Y i ; X T i β ; )w(x i )( X T iũ) 2 E{p 2 (Y i ; X T i β ; )w(x i )( X T iũ) 2 }] I,2, +I,2,2, where I,2, = 2 r 2 ũt H ũ. Meawhile, we have I,2,2 r 2 Thus, = r 2 O P (p / ) ũ 2. For I,3 i (S2.3), we observe that I,3 r 3 [p 2 (Y i ; X T i β ; )w(x i ) X i XT i E{p 2 (Y i; X T i β ; )w(x i ) X i XT i } ] F ũ 2 I,2 = r2 2 ũt H ũ +O P (r 2 p / ) ũ 2. (S2.5) p 3 (Y i ; X T i β ) w(x i) X T iũ 3 = O P (r 3 p3/2 ) ũ 3, which follows from Coditios A, A, A4 ad A5. By (S2.4) ad p 4 /, we ca choose some large C ǫ such that I, ad I,3 are all domiated by the first term of I,2 i (S2.5), which is positive by the eigevalue assumptio o H. This implies (S2.). Proof of Theorem 2 Notice the estimatig equatios l( β) β β= β =, sice β is a local miimizer of l ( β). Taylor s expasio applied to the left side of the estimatio equatios yields { } = p (Y i ; X T i β ; )w(x i ) X i

4 S4 Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag { { p 2 (Y i ; X T i β ; )w(x i ) X i XT i } ( β β; ) p 3 (Y i ; X T i β )w(x i){ X T i ( β β; )} 2 Xi p (Y i ; X T i β ; )w(x i ) X i }+K 2 ( β β; )+K 3, (S2.6) where β lies betwee β ; ad β. Below, we will show K 2 H = O P (p / ), (S2.7) K 3 = O P (p 5/2 /). (S2.8) First, to show (S2.7), ote that K 2 H = K 2 E(K 2 ) L. Similar argumets for the proof of I,2,2 i Theorem give L = O P (p / ). Secod, a similar proof used for I,3 i (S2.3) completes (S2.8). Third, by (S2.6) (S2.8) ad β β; = O P ( p /), we see that H ( β β; ) = p (Y i ; X T i β ; )w(x i ) X i +u, (S2.9) where u = O P (p 5/2 /). Note that by Coditio B5, Thus A Ω /2 u A F λ max (Ω /2 ) u = {tr(a A T )}/2 /λ /2 mi (Ω ) u = O P (p 5/2 / ) = o P (). A Ω /2 {H ( β β; )} = A Ω /2 p (Y i ; X T i β ; )w(x i ) X i +o P (). To complete provig Theorem 2, we apply the Lideberg-Feller cetral limit theorem(va dervaart,998)to Z i,wherez i = /2 A Ω /2 p (Y i ; X T i β ; )w(x i ) X i. Itsufficestocheck cov(z i) G; (II) E( Z i 2+δ ) = o()forsomeδ >. Coditio follows from the fact that var{p (Y; X T β ; )w(x ) X } = Ω. To verify coditio (II), otice that usig Coditios B5 ad A5, { [ E( Z i 2+δ ) (2+δ)/2 E A 2+δ F Ω /2 X {ψ(r(y,m(x ))) G (m(x ))} {q (m(x )) V(m(X ))} ] 2+δ } F w(x ) (m(x )) C (2+δ)/2 E[{λ /2 mi (Ω ) X } 2+δ {ψ(r(y,m(x ))) G (m(x ))}

5 Robust-BD Estimatio ad Iferece S5 {q (m(x )) V(m(X ))}/F (m(x )) 2+δ ] (2+δ)/2 E[ {ψ(r(y,m(x ))) G (m(x ))} Cp (2+δ)/2 {q (m(x )) V(m(X ))}/F (m(x )) 2+δ ] O((p /) (2+δ)/2 ). Thus, we get E( Z i 2+δ ) O((p /) (2+δ)/2 ) = O(p (2+δ)/2 / δ/2 ), which is o(). This verifies Coditio (II). Propositio (covariace matrix estimatio) Assume A, A, A2, A4, A5, B5, A6, ad A7 i the Appedix. Let V = H Ω H ad V = Ĥ Ω Ĥ. If p4/ as, the for ay /p -cosistet estimator β of β;, we have A ( V V )A T P for ay k (p +) matrix A satisfyig A A T G, where G is a k k matrix ad k is ay fixed iteger. Proof: Note A ( V V )A T V V A 2 F. Sice A 2 F to prove that V V = o P (). tr(g), it suffices First, we prove Ĥ H = o P (). Note that Ĥ H = {p 2 (Y i ; X T i { + I +I 2. β) p 2 (Y i ; X T i β ; )}w(x i ) X i XT i p 2 (Y i ; X T i β ; )w(x i ) X i XT i H From the proof of (S2.7) i Theorem 2, we kow that I 2 = O P (p / ) = o P (). We oly eed to cosider the term I. Let m i = m(x i ), m i = m(x i ), r i = r(y i, m i ) ad r i = r(y i,m i ). The I = [A (Y i, m i )+{ψ( r i ) G ( m i )}A ( m i ) A (Y i,m i ) {ψ(r i ) G (m i )}A (m i )]w(x i ) X i XT i = {G ( m i )A ( m i ) G (m i )A (m i )}w(x i ) X i XT i + {A (Y i, m i )+ψ( r i )A ( m i ) A (Y i,m i ) ψ(r i )A (m i )}w(x i ) X i XT i I, +I,2. Let g( ) = G ( )A ( ). By the assumptios, g( ) is differetiable. Thus g( m i ) g(m i ) = (g F ) ( X T i β )X T i( β β; ) = O P ()O P ( p )O P ( p /) = O P (p / ), }

6 S6 Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag where β is betwee β ad β;. Thus g( m(x i )) g(m(x i )) w(x i ) X i XT i = O P (p / )O P (p ) = O P (p 2 / ). F Similar argumets give I, = O P (p 2 / ) ad I,2 = O P (p 2 / ). Thus I = O P (p 2 / ) = o P (). Secod, we show Ω Ω = o P (). It is easy to see that Ω Ω = {p 2 (Y i; X T i { + =, +,2, β) p 2 (Y i; X T i β ; )}w 2 (X i ) X i XT i p 2 (Y i; X T i β ; )w 2 (X i ) X i XT i Ω where, = O P (p 2 / ) ad,2 = O P (p / ). We observe that Ω Ω = O P (p 2 / ) = o P (). Third, we show V V = o P (). Note V V = L + L 2 + L 3, where L = Ĥ ( Ω Ω )Ĥ, L 2 = Ĥ (H Ĥ)H Ω Ĥ ad L 3 = H Ω Ĥ (H Ĥ )H. By Assumptio B5, it is straightforward to verify that H O(), Ĥ O P() ad H Ω O(). Sice L Ĥ Ω Ω Ĥ, we coclude L = o P (), ad similarly L 2 = o P () ad L 3 = o P (). Hece V V = o P (). } Proof of Theorem 3 For the matrix A i (4.3), there exists a (p + k) (p +) matrix B satisfyig B B T = I p + k ad A B T =. Therefore, A β = g is equivalet to β = B Tγ + c, where γ is a (p + k) vector ad c = A T G g. Thus uder H i (4.3), we have β ; = B Tγ ; + c. The miimizig l ( β ) subject to A β = g is equivalet to miimizig l (Bγ T + c ) with respect to γ, ad we deote by γ the miimizer. Note that uder (4.4), β is the uique miimizer of l ( β ). Hece Λ = 2{l (B γ T +c ) l ( β)}. Before showig Theorem 3, we eed Lemma. Lemma Assume coditios of Theorem 3. The uder H i (4.3), we have that B T( γ γ ; ) = B T(B H B T ) B p (Y i; X T i β ; )w(x i ) X i +o P ( /2 ), ad 2{l (B T γ +c ) l ( β)} = (B T γ +c β) T H (B T γ +c β)+op ().

7 Robust-BD Estimatio ad Iferece S7 Proof: To obtai the first part, followig the proof of (S2.9) i Theorem 2, we have a similar expressio for γ, B H B( γ T γ ; ) = B p (Y i ; X T i β ; )w(x i ) X i +w, with w = o P ( /2 ). As a result, B T ( γ γ ; ) = BT (B H B T ) B p (Y i ; X T i β ; )w(x i ) X i +B T (B H B T ) w. We otice that B(B T H B) T w (B H B) T w w /λ mi (H ) = o P ( /2 ), i which the fact λ mi (B H B T) λ mi(h ) is used. The proof of the secod part proceeds i three steps. I Step, we use the followig Taylor expasio for l (B γ T +c ) l ( β), l (B γ T +c ) l ( β) = p 2 2 (Y i ; X T i β)w(x i ){ X T i(b γ T +c β)} 2 + p 6 3 (Y i ; X T i β )w(x i ){ X T i(b γ T +c β)} 3 I +I 2, where β lies betwee β ad B T γ +c. I Step 2, we aalyze the stochastic order of B γ T + c β. For a matrix X whose colum vectors are liearly idepedet, set P X = X(X T X) X T. Defie H = I p + P /2 H = P B T /2 H. The H A T BT (B H B T) B = H /2 H H /2. By (S2.9) ad the first part of Lemma, we see immediately that B γ T +c β = B T ( γ γ ; ) ( β β; ) { = H /2 H H /2 p,i w(x i ) X i }+o P ( /2 ), (S2.) wherep,i = p (Y i ; X T i β ; ). Notethat H /2 H H /2 O P (/ ). This gives B T γ +c β = OP (/ ). { p,iw(x i ) X i } = (S2.) I Step 3, we coclude from (S2.) that I 2 = O P {(p /) 3/2 } = o P (/). The 2{l (B γ T +c ) l ( β)} = 2I +o P (). Similar to the proof of Propositio, it is straightforward to see that 2I = (B γ T +c β) { T } p 2 (Y i ; X T i β)w(x i ) X i XT i (B γ T +c β)

8 S8 Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag = (B γ T +c β) { T } p 2 (Y i ; X T i β ; )w(x i ) X i XT i (B γ T +c β)+op () = (B γ T +c β) T E{p 2 (Y ; X T β ; )w(x i ) X XT }(BT γ +c β)+op () = (B γ T +c β) T H (B γ T +c β)+op (). The the secod part of Lemma is proved. to We ow show Theorem 3. For part (i), a direct use of Lemma ad (S2.) leads 2{l (B γ T +c ) l ( β)} { } T { = p,i w(x i ) X i H /2 H H /2 p,i w(x i ) X i }+o P (). Sice H is idempotet of rak k, it ca be writte as H = C T C, where C is a k (p +) matrix satisfyig C C T = I k. The 2{l (B γ T +c ) l ( β)} { } T { = C H /2 p,i w(x i ) X i C H /2 p,i w(x i ) X i }+o P (). Now cosider part (ii). If ψ(r) = r ad the q-fuctio satisfies (4.5), the p (y;θ) = q (y;θ), p 2 (y;θ) = q 2 (y;θ) ad H = Ω /C, where q j (y;θ) = j θ Q j q (y,f (θ)). I this case, similar argumets for Theorem 2 yield C H /2 which completes the proof. q (Y i ; X T i β ; )w(x i ) X i L N(,CIk ), Proof of Theorem 4 Before showig Theorem 4, Lemma 2 is eeded. Lemma 2 Assume coditios of Theorem 4. The β β ; = H p (Y i ; X T i β ; )w(x i ) X i +o P ( /2 ), (A Ĥ Ω Ĥ AT ) /2 L A ( β β; ) N(,I k ). Proof:Followig(S2.9)itheproofofTheorem2,weobservethat u = O P (p 5/2 /) = o P ( /2 ). Coditio B5 completes the proof for the first part.

9 Robust-BD Estimatio ad Iferece S9 Toshowthesecodpart,deoteU = A H Ω H AT adû = A Ĥ Ω Ĥ Notice that the eigevalues of H Ω H are uiformly bouded away from. So are the eigevalues of U. From the first part, we see that It follows that A ( β β; ) = A H p (Y i ; X T i β ; )w(x i ) X i +o P ( /2 ). U /2 A ( β β; ) = Z i +o P (), AT. where Z i = /2 U /2 A H p (Y i ; X T i β ; )w(x i ) X i. To show Z L i N(,I k ), similar to the proof for Theorem 2, we check (III) cov(z i) I k ; (IV) E( Z i 2+δ ) = o() for some δ >. Coditio (III) is straightforward sice cov(z i) = U /2 U U /2 = I k. To check coditio (IV), similar argumets used i the proof of Theorem 2 give that E( Z i 2+δ ) = O((p /) (2+δ)/2 ). This ad the boudedess of ψ yield E( Z i 2+δ ) O(p (2+δ)/2 / δ/2 ) = o(). Hece U /2 L A ( β β; ) N(,I k ). (S2.2) From the proof of Propositio, it ca be cocluded that Û U = o P () ad that the eigevalues of Û are uiformly bouded away from ad with probability tedig to oe. Cosequetly, Û /2 U /2 I k = o P (). (S2.3) Combiig(S2.2),(S2.3)adSlutsky stheoremcompletestheproofthat Û /2 A ( β L β ; ) N(,I k ). We ow show Theorem 4, which follows directly from H i (4.3) ad the secod part of Lemma 2. This completes the proof. Proof of Theorem 5 Note that W ca be decomposed ito three additive terms, I = {A ( β β; )} T (A V A T ) {A ( β β; )}, I 2 = 2(A β; g ) T (A V A T ) {A ( β β; )}, I 3 = (A β; g ) T (A V A T ) (A β; g ), where V = Ĥ Ω Ĥ. We observe that I χ 2 k followig the secod part of Lemma 2; I 3 = (A β; g ) T M (A β; g ){+o P ()} by Propositio ; I 2 = O P ( ) by Cauchy-Schwartz iequality. Thus I 3 λ mi (M ) A β; g 2 {+o P ()} = λ max(m) A β; g 2 +o P (). These complete the proof for W. L

10 S Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag Proof of Theorem 6 Followigthe secodpartoflemma 2, we observethat (A V A T ) /2 L (A β g ) N(M /2 c,i k ), which completes the proof. Proof of Theorem 7 We first eed to show Lemma 3. Lemma 3 Suppose that (X o,y o ) follows the distributio of (X,Y) ad is idepedet of the traiig set T. If Q is a BD, the E{Q(Y o, m(x o ))} = E{Q(Y o,m(x o ))}+E{Q(m(X o ), m(x o ))}. Proof: Let q be the geeratig fuctio of Q. The Q(Y o, m(x o )) = [q(m(xo )) E{q(Y o ) T,X o }]+[E{q(Y o ) T,X o } q(y o )] q(m(x o ))+q( m(x o ))+{Y o m(x o )}q ( m(x o )). (S2.4) Sice (X o,y o ) is idepedet of T, we deduce from Chow ad Teicher (989, Corollary 3, p. 223) that E{q(Y o ) T,X o } = E{q(Y o ) X o }. (S2.5) Similarly, E{Y o q ( m(x o )) T,X o } = E(Y o X o )q ( m(x o )) = m(xo )q ( m(x o )). (S2.6) Applyig (S2.5) ad (S2.6) to (S2.4) results i E{Q(Y o, m(x o )) T,X o } = E{Q(Y o,m(x o )) X o }+Q(m(X o ), m(x o )) ad thus the coclusio. Now show Theorem 7. Settig Q i Lemma 3 to be the misclassificatio loss gives /2[E{R( φ )} R(φ,B )] E[ m(x o ).5 I{m(X o ).5, m(x o ) >.5}] +E[ m(x o ).5 I{m(Xo ) >.5, m(xo ).5}] = I +I 2. For ay ǫ >, it follows that I = E[ m(x o ).5 I{m(Xo ) <.5 ǫ, m(xo ) >.5}] +E[ m(x o ).5 I{.5 ǫ m(x o ).5, m(x o ) >.5}] P{ m(x o ) m(xo ) > ǫ}+ǫ ad similarly, I 2 ǫ+p{ m(x o ) m(xo ) ǫ}. Recall that m(x o ) m(xo ) = F ( X o T β) F ( X o T β; ) (F ) ( X o T β ) Xo β β;,

11 Robust-BD Estimatio ad Iferece S forsome β betwee β ; ad β, where Xo = (,X o T ) T. ByCoditioA4, wecoclude that (F ) ( X o T β ) = O P (). This alog with β β; = O P () ad X o = O P ( p ) implies that m(x o ) m(xo ) = O P(r p ) = o P (). Therefore I ad I 2, which completes the proof. S3 Figures 7 i Sectio 6.2 cotamiated data, classical dev. loss, w= cotamiated data, robust dev. loss, with w j cotamiated data, classical exp. loss, w= 2 j cotamiated data, robust exp. loss, with w j 2 j Figure 7: (Simulated Beroulli respose data with cotamiatio) Boxplots of β j β j;, j =,,...,p (from left to right i each pael). Left paels: the o-robust estimates; right paels: the robust estimates.

12 S2 Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag cotamiated data, robust dev. loss, with w, H cotamiated data, robust dev. loss, with w, composite H (II) cotamiated data, robust exp. loss, with w, H cotamiated data, robust exp. loss, with w, composite H (II) Figure 8: (Simulated Beroulli respose data with cotamiatio) Empirical quatiles (othey-axis)ofteststatisticsw versusquatiles(othex-axis)oftheχ 2 k distributio. Solid lie: the 45 degree referece lie. Left paels: for testig H ; right paels: for testig H (II).

13 Robust-BD Estimatio ad Iferece S3 deviace loss, H deviace loss, composite H (II) d expoetial loss, H d d expoetial loss, composite H (II) d Figure 9: Level of tests for the Beroulli respose data. The dashed lie correspods to the o-robust Wald-type test; the solid lie correspods to the robust Wald-type test; the dotted lie idicates the 5% omial level. Left paels: for testig H ; right paels: for testig H (II).

14 S4 Chumig Zhag Xiao Guo Che Cheg Zhegju Zhag o cotamiated case, deviace loss, H cotamiated case, deviace loss, H o cotamiated case, expoetial loss, H cotamiated case, expoetial loss, H Figure : Observed power fuctios of tests for the Beroulli respose data. The dashed lie correspods to the o-robust Wald-type test; the solid lie correspods to the robust Wald-type test; the dotted lie idicates the 5% omial level. Left paels: o-cotamiated case; right paels: cotamiated case.

Supplemental Material: Proofs

Supplemental Material: Proofs Proof to Theorem Supplemetal Material: Proofs Proof. Let be the miimal umber of traiig items to esure a uique solutio θ. First cosider the case. It happes if ad oly if θ ad Rak(A) d, which is a special

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS MASSACHUSTTS INSTITUT OF TCHNOLOGY 6.436J/5.085J Fall 2008 Lecture 9 /7/2008 LAWS OF LARG NUMBRS II Cotets. The strog law of large umbers 2. The Cheroff boud TH STRONG LAW OF LARG NUMBRS While the weak

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d

Linear regression. Daniel Hsu (COMS 4771) (y i x T i β)2 2πσ. 2 2σ 2. 1 n. (x T i β y i ) 2. 1 ˆβ arg min. β R n d Liear regressio Daiel Hsu (COMS 477) Maximum likelihood estimatio Oe of the simplest liear regressio models is the followig: (X, Y ),..., (X, Y ), (X, Y ) are iid radom pairs takig values i R d R, ad Y

More information

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Convergence of random variables. (telegram style notes) P.J.C. Spreij Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space

More information

Lecture 7: Properties of Random Samples

Lecture 7: Properties of Random Samples Lecture 7: Properties of Radom Samples 1 Cotiued From Last Class Theorem 1.1. Let X 1, X,...X be a radom sample from a populatio with mea µ ad variace σ

More information

Sequences and Series of Functions

Sequences and Series of Functions Chapter 6 Sequeces ad Series of Fuctios 6.1. Covergece of a Sequece of Fuctios Poitwise Covergece. Defiitio 6.1. Let, for each N, fuctio f : A R be defied. If, for each x A, the sequece (f (x)) coverges

More information

TESTING FOR THE BUFFERED AUTOREGRESSIVE PROCESSES (SUPPLEMENTARY MATERIAL)

TESTING FOR THE BUFFERED AUTOREGRESSIVE PROCESSES (SUPPLEMENTARY MATERIAL) TESTING FOR THE BUFFERED AUTOREGRESSIVE PROCESSES SUPPLEMENTARY MATERIAL) By Ke Zhu, Philip L.H. Yu ad Wai Keug Li Chiese Academy of Scieces ad Uiversity of Hog Kog APPENDIX: PROOFS I this appedix, we

More information

MAS111 Convergence and Continuity

MAS111 Convergence and Continuity MAS Covergece ad Cotiuity Key Objectives At the ed of the course, studets should kow the followig topics ad be able to apply the basic priciples ad theorems therei to solvig various problems cocerig covergece

More information

arxiv: v1 [math.pr] 4 Dec 2013

arxiv: v1 [math.pr] 4 Dec 2013 Squared-Norm Empirical Process i Baach Space arxiv:32005v [mathpr] 4 Dec 203 Vicet Q Vu Departmet of Statistics The Ohio State Uiversity Columbus, OH vqv@statosuedu Abstract Jig Lei Departmet of Statistics

More information

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i

More information

Lecture 20: Multivariate convergence and the Central Limit Theorem

Lecture 20: Multivariate convergence and the Central Limit Theorem Lecture 20: Multivariate covergece ad the Cetral Limit Theorem Covergece i distributio for radom vectors Let Z,Z 1,Z 2,... be radom vectors o R k. If the cdf of Z is cotiuous, the we ca defie covergece

More information

Technical Proofs for Homogeneity Pursuit

Technical Proofs for Homogeneity Pursuit Techical Proofs for Homogeeity Pursuit bstract This is the supplemetal material for the article Homogeeity Pursuit, submitted for publicatio i Joural of the merica Statistical ssociatio. B Proofs B. Proof

More information

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1

1 = δ2 (0, ), Y Y n nδ. , T n = Y Y n n. ( U n,k + X ) ( f U n,k + Y ) n 2n f U n,k + θ Y ) 2 E X1 2 X1 8. The cetral limit theorems 8.1. The cetral limit theorem for i.i.d. sequeces. ecall that C ( is N -separatig. Theorem 8.1. Let X 1, X,... be i.i.d. radom variables with EX 1 = ad EX 1 = σ (,. Suppose

More information

Ma 4121: Introduction to Lebesgue Integration Solutions to Homework Assignment 5

Ma 4121: Introduction to Lebesgue Integration Solutions to Homework Assignment 5 Ma 42: Itroductio to Lebesgue Itegratio Solutios to Homework Assigmet 5 Prof. Wickerhauser Due Thursday, April th, 23 Please retur your solutios to the istructor by the ed of class o the due date. You

More information

Supplementary Material for Fast Stochastic AUC Maximization with O(1/n)-Convergence Rate

Supplementary Material for Fast Stochastic AUC Maximization with O(1/n)-Convergence Rate Supplemetary Material for Fast Stochastic AUC Maximizatio with O/-Covergece Rate Migrui Liu Xiaoxua Zhag Zaiyi Che Xiaoyu Wag 3 iabao Yag echical Lemmas ized versio of Hoeffdig s iequality, ote that We

More information

Definition 4.2. (a) A sequence {x n } in a Banach space X is a basis for X if. unique scalars a n (x) such that x = n. a n (x) x n. (4.

Definition 4.2. (a) A sequence {x n } in a Banach space X is a basis for X if. unique scalars a n (x) such that x = n. a n (x) x n. (4. 4. BASES I BAACH SPACES 39 4. BASES I BAACH SPACES Sice a Baach space X is a vector space, it must possess a Hamel, or vector space, basis, i.e., a subset {x γ } γ Γ whose fiite liear spa is all of X ad

More information

Summary and Discussion on Simultaneous Analysis of Lasso and Dantzig Selector

Summary and Discussion on Simultaneous Analysis of Lasso and Dantzig Selector Summary ad Discussio o Simultaeous Aalysis of Lasso ad Datzig Selector STAT732, Sprig 28 Duzhe Wag May 4, 28 Abstract This is a discussio o the work i Bickel, Ritov ad Tsybakov (29). We begi with a short

More information

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 12

Machine Learning Theory Tübingen University, WS 2016/2017 Lecture 12 Machie Learig Theory Tübige Uiversity, WS 06/07 Lecture Tolstikhi Ilya Abstract I this lecture we derive risk bouds for kerel methods. We will start by showig that Soft Margi kerel SVM correspods to miimizig

More information

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)].

Probability 2 - Notes 10. Lemma. If X is a random variable and g(x) 0 for all x in the support of f X, then P(g(X) 1) E[g(X)]. Probability 2 - Notes 0 Some Useful Iequalities. Lemma. If X is a radom variable ad g(x 0 for all x i the support of f X, the P(g(X E[g(X]. Proof. (cotiuous case P(g(X Corollaries x:g(x f X (xdx x:g(x

More information

18.657: Mathematics of Machine Learning

18.657: Mathematics of Machine Learning 8.657: Mathematics of Machie Learig Lecturer: Philippe Rigollet Lecture 0 Scribe: Ade Forrow Oct. 3, 05 Recall the followig defiitios from last time: Defiitio: A fuctio K : X X R is called a positive symmetric

More information

Econometrica Supplementary Material

Econometrica Supplementary Material Ecoometrica Supplemetary Material SUPPLEMENT TO NONPARAMETRIC INSTRUMENTAL VARIABLES ESTIMATION OF A QUANTILE REGRESSION MODEL : MATHEMATICAL APPENDIX (Ecoometrica, Vol. 75, No. 4, July 27, 1191 128) BY

More information

Optimally Sparse SVMs

Optimally Sparse SVMs A. Proof of Lemma 3. We here prove a lower boud o the umber of support vectors to achieve geeralizatio bouds of the form which we cosider. Importatly, this result holds ot oly for liear classifiers, but

More information

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss ECE 90 Lecture : Complexity Regularizatio ad the Squared Loss R. Nowak 5/7/009 I the previous lectures we made use of the Cheroff/Hoeffdig bouds for our aalysis of classifier errors. Hoeffdig s iequality

More information

IIT JAM Mathematical Statistics (MS) 2006 SECTION A

IIT JAM Mathematical Statistics (MS) 2006 SECTION A IIT JAM Mathematical Statistics (MS) 6 SECTION A. If a > for ad lim a / L >, the which of the followig series is ot coverget? (a) (b) (c) (d) (d) = = a = a = a a + / a lim a a / + = lim a / a / + = lim

More information

Math Solutions to homework 6

Math Solutions to homework 6 Math 175 - Solutios to homework 6 Cédric De Groote November 16, 2017 Problem 1 (8.11 i the book): Let K be a compact Hermitia operator o a Hilbert space H ad let the kerel of K be {0}. Show that there

More information

5. Matrix exponentials and Von Neumann s theorem The matrix exponential. For an n n matrix X we define

5. Matrix exponentials and Von Neumann s theorem The matrix exponential. For an n n matrix X we define 5. Matrix expoetials ad Vo Neuma s theorem 5.1. The matrix expoetial. For a matrix X we defie e X = exp X = I + X + X2 2! +... = 0 X!. We assume that the etries are complex so that exp is well defied o

More information

Sieve Estimators: Consistency and Rates of Convergence

Sieve Estimators: Consistency and Rates of Convergence EECS 598: Statistical Learig Theory, Witer 2014 Topic 6 Sieve Estimators: Cosistecy ad Rates of Covergece Lecturer: Clayto Scott Scribe: Julia Katz-Samuels, Brado Oselio, Pi-Yu Che Disclaimer: These otes

More information

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

Matrix Representation of Data in Experiment

Matrix Representation of Data in Experiment Matrix Represetatio of Data i Experimet Cosider a very simple model for resposes y ij : y ij i ij, i 1,; j 1,,..., (ote that for simplicity we are assumig the two () groups are of equal sample size ) Y

More information

Ω ). Then the following inequality takes place:

Ω ). Then the following inequality takes place: Lecture 8 Lemma 5. Let f : R R be a cotiuously differetiable covex fuctio. Choose a costat δ > ad cosider the subset Ωδ = { R f δ } R. Let Ωδ ad assume that f < δ, i.e., is ot o the boudary of f = δ, i.e.,

More information

Empirical Processes: Glivenko Cantelli Theorems

Empirical Processes: Glivenko Cantelli Theorems Empirical Processes: Gliveko Catelli Theorems Mouliath Baerjee Jue 6, 200 Gliveko Catelli classes of fuctios The reader is referred to Chapter.6 of Weller s Torgo otes, Chapter??? of VDVW ad Chapter 8.3

More information

Appendix to: Hypothesis Testing for Multiple Mean and Correlation Curves with Functional Data

Appendix to: Hypothesis Testing for Multiple Mean and Correlation Curves with Functional Data Appedix to: Hypothesis Testig for Multiple Mea ad Correlatio Curves with Fuctioal Data Ao Yua 1, Hog-Bi Fag 1, Haiou Li 1, Coli O. Wu, Mig T. Ta 1, 1 Departmet of Biostatistics, Bioiformatics ad Biomathematics,

More information

Advanced Stochastic Processes.

Advanced Stochastic Processes. Advaced Stochastic Processes. David Gamarik LECTURE 2 Radom variables ad measurable fuctios. Strog Law of Large Numbers (SLLN). Scary stuff cotiued... Outlie of Lecture Radom variables ad measurable fuctios.

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit theorems Throughout this sectio we will assume a probability space (Ω, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

11 THE GMM ESTIMATION

11 THE GMM ESTIMATION Cotets THE GMM ESTIMATION 2. Cosistecy ad Asymptotic Normality..................... 3.2 Regularity Coditios ad Idetificatio..................... 4.3 The GMM Iterpretatio of the OLS Estimatio.................

More information

Lecture 7: October 18, 2017

Lecture 7: October 18, 2017 Iformatio ad Codig Theory Autum 207 Lecturer: Madhur Tulsiai Lecture 7: October 8, 207 Biary hypothesis testig I this lecture, we apply the tools developed i the past few lectures to uderstad the problem

More information

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator

Slide Set 13 Linear Model with Endogenous Regressors and the GMM estimator Slide Set 13 Liear Model with Edogeous Regressors ad the GMM estimator Pietro Coretto pcoretto@uisa.it Ecoometrics Master i Ecoomics ad Fiace (MEF) Uiversità degli Studi di Napoli Federico II Versio: Friday

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 3 9/11/2013. Large deviations Theory. Cramér s Theorem MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/5.070J Fall 203 Lecture 3 9//203 Large deviatios Theory. Cramér s Theorem Cotet.. Cramér s Theorem. 2. Rate fuctio ad properties. 3. Chage of measure techique.

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Fundamental Theorem of Algebra. Yvonne Lai March 2010

Fundamental Theorem of Algebra. Yvonne Lai March 2010 Fudametal Theorem of Algebra Yvoe Lai March 010 We prove the Fudametal Theorem of Algebra: Fudametal Theorem of Algebra. Let f be a o-costat polyomial with real coefficiets. The f has at least oe complex

More information

Notes 27 : Brownian motion: path properties

Notes 27 : Brownian motion: path properties Notes 27 : Browia motio: path properties Math 733-734: Theory of Probability Lecturer: Sebastie Roch Refereces:[Dur10, Sectio 8.1], [MP10, Sectio 1.1, 1.2, 1.3]. Recall: DEF 27.1 (Covariace) Let X = (X

More information

Solution. 1 Solutions of Homework 1. Sangchul Lee. October 27, Problem 1.1

Solution. 1 Solutions of Homework 1. Sangchul Lee. October 27, Problem 1.1 Solutio Sagchul Lee October 7, 017 1 Solutios of Homework 1 Problem 1.1 Let Ω,F,P) be a probability space. Show that if {A : N} F such that A := lim A exists, the PA) = lim PA ). Proof. Usig the cotiuity

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

REGRESSION WITH QUADRATIC LOSS

REGRESSION WITH QUADRATIC LOSS REGRESSION WITH QUADRATIC LOSS MAXIM RAGINSKY Regressio with quadratic loss is aother basic problem studied i statistical learig theory. We have a radom couple Z = X, Y ), where, as before, X is a R d

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation ECE 645: Estimatio Theory Sprig 2015 Istructor: Prof. Staley H. Cha Maximum Likelihood Estimatio (LaTeX prepared by Shaobo Fag) April 14, 2015 This lecture ote is based o ECE 645(Sprig 2015) by Prof. Staley

More information

Mathematics 170B Selected HW Solutions.

Mathematics 170B Selected HW Solutions. Mathematics 17B Selected HW Solutios. F 4. Suppose X is B(,p). (a)fidthemometgeeratigfuctiom (s)of(x p)/ p(1 p). Write q = 1 p. The MGF of X is (pe s + q), sice X ca be writte as the sum of idepedet Beroulli

More information

arxiv: v1 [math.pr] 13 Oct 2011

arxiv: v1 [math.pr] 13 Oct 2011 A tail iequality for quadratic forms of subgaussia radom vectors Daiel Hsu, Sham M. Kakade,, ad Tog Zhag 3 arxiv:0.84v math.pr] 3 Oct 0 Microsoft Research New Eglad Departmet of Statistics, Wharto School,

More information

Lecture 8: Convergence of transformations and law of large numbers

Lecture 8: Convergence of transformations and law of large numbers Lecture 8: Covergece of trasformatios ad law of large umbers Trasformatio ad covergece Trasformatio is a importat tool i statistics. If X coverges to X i some sese, we ofte eed to check whether g(x ) coverges

More information

Lecture 19. sup y 1,..., yn B d n

Lecture 19. sup y 1,..., yn B d n STAT 06A: Polyomials of adom Variables Lecture date: Nov Lecture 19 Grothedieck s Iequality Scribe: Be Hough The scribes are based o a guest lecture by ya O Doell. I this lecture we prove Grothedieck s

More information

LECTURE 8: ASYMPTOTICS I

LECTURE 8: ASYMPTOTICS I LECTURE 8: ASYMPTOTICS I We are iterested i the properties of estimators as. Cosider a sequece of radom variables {, X 1}. N. M. Kiefer, Corell Uiversity, Ecoomics 60 1 Defiitio: (Weak covergece) A sequece

More information

Chapter 5. Inequalities. 5.1 The Markov and Chebyshev inequalities

Chapter 5. Inequalities. 5.1 The Markov and Chebyshev inequalities Chapter 5 Iequalities 5.1 The Markov ad Chebyshev iequalities As you have probably see o today s frot page: every perso i the upper teth percetile ears at least 1 times more tha the average salary. I other

More information

Summary. Recap ... Last Lecture. Summary. Theorem

Summary. Recap ... Last Lecture. Summary. Theorem Last Lecture Biostatistics 602 - Statistical Iferece Lecture 23 Hyu Mi Kag April 11th, 2013 What is p-value? What is the advatage of p-value compared to hypothesis testig procedure with size α? How ca

More information

Solution to Chapter 2 Analytical Exercises

Solution to Chapter 2 Analytical Exercises Nov. 25, 23, Revised Dec. 27, 23 Hayashi Ecoometrics Solutio to Chapter 2 Aalytical Exercises. For ay ε >, So, plim z =. O the other had, which meas that lim E(z =. 2. As show i the hit, Prob( z > ε =

More information

MINIMAX RATES OF CONVERGENCE AND OPTIMALITY OF BAYES FACTOR WAVELET REGRESSION ESTIMATORS UNDER POINTWISE RISKS

MINIMAX RATES OF CONVERGENCE AND OPTIMALITY OF BAYES FACTOR WAVELET REGRESSION ESTIMATORS UNDER POINTWISE RISKS Statistica Siica 19 (2009) 1389 1406: Supplemet S1 S17 1 MINIMAX RATES OF ONVERGENE AND OPTIMALITY OF BAYES FATOR WAVELET REGRESSION ESTIMATORS UNDER POINTWISE RISKS Natalia Bochkia ad Theofais Sapatias

More information

Lecture 10: Bounded Linear Operators and Orthogonality in Hilbert Spaces

Lecture 10: Bounded Linear Operators and Orthogonality in Hilbert Spaces Lecture : Bouded Liear Operators ad Orthogoality i Hilbert Spaces 34 Bouded Liear Operator Let ( X, ), ( Y, ) i i be ored liear vector spaces ad { } X Y The, T is said to be bouded if a real uber c such

More information

Problem Set 2 Solutions

Problem Set 2 Solutions CS271 Radomess & Computatio, Sprig 2018 Problem Set 2 Solutios Poit totals are i the margi; the maximum total umber of poits was 52. 1. Probabilistic method for domiatig sets 6pts Pick a radom subset S

More information

Sequences and Limits

Sequences and Limits Chapter Sequeces ad Limits Let { a } be a sequece of real or complex umbers A ecessary ad sufficiet coditio for the sequece to coverge is that for ay ɛ > 0 there exists a iteger N > 0 such that a p a q

More information

The natural exponential function

The natural exponential function The atural expoetial fuctio Attila Máté Brookly College of the City Uiversity of New York December, 205 Cotets The atural expoetial fuctio for real x. Beroulli s iequality.....................................2

More information

b i u x i U a i j u x i u x j

b i u x i U a i j u x i u x j M ath 5 2 7 Fall 2 0 0 9 L ecture 1 9 N ov. 1 6, 2 0 0 9 ) S ecod- Order Elliptic Equatios: Weak S olutios 1. Defiitios. I this ad the followig two lectures we will study the boudary value problem Here

More information

Self-normalized deviation inequalities with application to t-statistic

Self-normalized deviation inequalities with application to t-statistic Self-ormalized deviatio iequalities with applicatio to t-statistic Xiequa Fa Ceter for Applied Mathematics, Tiaji Uiversity, 30007 Tiaji, Chia Abstract Let ξ i i 1 be a sequece of idepedet ad symmetric

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

Integrable Functions. { f n } is called a determining sequence for f. If f is integrable with respect to, then f d does exist as a finite real number

Integrable Functions. { f n } is called a determining sequence for f. If f is integrable with respect to, then f d does exist as a finite real number MATH 532 Itegrable Fuctios Dr. Neal, WKU We ow shall defie what it meas for a measurable fuctio to be itegrable, show that all itegral properties of simple fuctios still hold, ad the give some coditios

More information

Inverse Nodal Problems for Differential Equation on the Half-line

Inverse Nodal Problems for Differential Equation on the Half-line Australia Joural of Basic ad Applied Scieces, 3(4): 4498-4502, 2009 ISSN 1991-8178 Iverse Nodal Problems for Differetial Equatio o the Half-lie 1 2 3 A. Dabbaghia, A. Nematy ad Sh. Akbarpoor 1 Islamic

More information

LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK)

LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK) LECTURE 8: ORTHOGONALITY (CHAPTER 5 IN THE BOOK) Everythig marked by is ot required by the course syllabus I this lecture, all vector spaces is over the real umber R. All vectors i R is viewed as a colum

More information

A NOTE ON LEBESGUE SPACES

A NOTE ON LEBESGUE SPACES Volume 6, 1981 Pages 363 369 http://topology.aubur.edu/tp/ A NOTE ON LEBESGUE SPACES by Sam B. Nadler, Jr. ad Thelma West Topology Proceedigs Web: http://topology.aubur.edu/tp/ Mail: Topology Proceedigs

More information

CHAPTER 5. Theory and Solution Using Matrix Techniques

CHAPTER 5. Theory and Solution Using Matrix Techniques A SERIES OF CLASS NOTES FOR 2005-2006 TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS DE CLASS NOTES 3 A COLLECTION OF HANDOUTS ON SYSTEMS OF ORDINARY DIFFERENTIAL

More information

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT

Geometry of LS. LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT OCTOBER 7, 2016 LECTURE 3 GEOMETRY OF LS, PROPERTIES OF σ 2, PARTITIONED REGRESSION, GOODNESS OF FIT Geometry of LS We ca thik of y ad the colums of X as members of the -dimesioal Euclidea space R Oe ca

More information

Supplementary Material for Fast Stochastic AUC Maximization with O(1/n)-Convergence Rate

Supplementary Material for Fast Stochastic AUC Maximization with O(1/n)-Convergence Rate Supplemetary Material for Fast Stochastic AUC Maximizatio with O/-Covergece Rate Migrui Liu Xiaoxua Zhag Zaiyi Che Xiaoyu Wag 3 iabao Yag echical Lemmas ized versio of Hoeffdig s iequality, ote that We

More information

On forward improvement iteration for stopping problems

On forward improvement iteration for stopping problems O forward improvemet iteratio for stoppig problems Mathematical Istitute, Uiversity of Kiel, Ludewig-Mey-Str. 4, D-24098 Kiel, Germay irle@math.ui-iel.de Albrecht Irle Abstract. We cosider the optimal

More information

A Hadamard-type lower bound for symmetric diagonally dominant positive matrices

A Hadamard-type lower bound for symmetric diagonally dominant positive matrices A Hadamard-type lower boud for symmetric diagoally domiat positive matrices Christopher J. Hillar, Adre Wibisoo Uiversity of Califoria, Berkeley Jauary 7, 205 Abstract We prove a ew lower-boud form of

More information

CHAPTER 5 SOME MINIMAX AND SADDLE POINT THEOREMS

CHAPTER 5 SOME MINIMAX AND SADDLE POINT THEOREMS CHAPTR 5 SOM MINIMA AND SADDL POINT THORMS 5. INTRODUCTION Fied poit theorems provide importat tools i game theory which are used to prove the equilibrium ad eistece theorems. For istace, the fied poit

More information

1+x 1 + α+x. x = 2(α x2 ) 1+x

1+x 1 + α+x. x = 2(α x2 ) 1+x Math 2030 Homework 6 Solutios # [Problem 5] For coveiece we let α lim sup a ad β lim sup b. Without loss of geerality let us assume that α β. If α the by assumptio β < so i this case α + β. By Theorem

More information

Lecture 27: Optimal Estimators and Functional Delta Method

Lecture 27: Optimal Estimators and Functional Delta Method Stat210B: Theoretical Statistics Lecture Date: April 19, 2007 Lecture 27: Optimal Estimators ad Fuctioal Delta Method Lecturer: Michael I. Jorda Scribe: Guilherme V. Rocha 1 Achievig Optimal Estimators

More information

Lecture 3 : Random variables and their distributions

Lecture 3 : Random variables and their distributions Lecture 3 : Radom variables ad their distributios 3.1 Radom variables Let (Ω, F) ad (S, S) be two measurable spaces. A map X : Ω S is measurable or a radom variable (deoted r.v.) if X 1 (A) {ω : X(ω) A}

More information

Regression with quadratic loss

Regression with quadratic loss Regressio with quadratic loss Maxim Ragisky October 13, 2015 Regressio with quadratic loss is aother basic problem studied i statistical learig theory. We have a radom couple Z = X,Y, where, as before,

More information

Theorem 3. A subset S of a topological space X is compact if and only if every open cover of S by open sets in X has a finite subcover.

Theorem 3. A subset S of a topological space X is compact if and only if every open cover of S by open sets in X has a finite subcover. Compactess Defiitio 1. A cover or a coverig of a topological space X is a family C of subsets of X whose uio is X. A subcover of a cover C is a subfamily of C which is a cover of X. A ope cover of X is

More information

Chapter 6 Infinite Series

Chapter 6 Infinite Series Chapter 6 Ifiite Series I the previous chapter we cosidered itegrals which were improper i the sese that the iterval of itegratio was ubouded. I this chapter we are goig to discuss a topic which is somewhat

More information

1. By using truth tables prove that, for all statements P and Q, the statement

1. By using truth tables prove that, for all statements P and Q, the statement Author: Satiago Salazar Problems I: Mathematical Statemets ad Proofs. By usig truth tables prove that, for all statemets P ad Q, the statemet P Q ad its cotrapositive ot Q (ot P) are equivalet. I example.2.3

More information

Chapter 7 Isoperimetric problem

Chapter 7 Isoperimetric problem Chapter 7 Isoperimetric problem Recall that the isoperimetric problem (see the itroductio its coectio with ido s proble) is oe of the most classical problem of a shape optimizatio. It ca be formulated

More information

Exponential Families and Bayesian Inference

Exponential Families and Bayesian Inference Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where

More information

Mi-Hwa Ko and Tae-Sung Kim

Mi-Hwa Ko and Tae-Sung Kim J. Korea Math. Soc. 42 2005), No. 5, pp. 949 957 ALMOST SURE CONVERGENCE FOR WEIGHTED SUMS OF NEGATIVELY ORTHANT DEPENDENT RANDOM VARIABLES Mi-Hwa Ko ad Tae-Sug Kim Abstract. For weighted sum of a sequece

More information

Detailed proofs of Propositions 3.1 and 3.2

Detailed proofs of Propositions 3.1 and 3.2 Detailed proofs of Propositios 3. ad 3. Proof of Propositio 3. NB: itegratio sets are geerally omitted for itegrals defied over a uit hypercube [0, s with ay s d. We first give four lemmas. The proof of

More information

Comparison of Minimum Initial Capital with Investment and Non-investment Discrete Time Surplus Processes

Comparison of Minimum Initial Capital with Investment and Non-investment Discrete Time Surplus Processes The 22 d Aual Meetig i Mathematics (AMM 207) Departmet of Mathematics, Faculty of Sciece Chiag Mai Uiversity, Chiag Mai, Thailad Compariso of Miimum Iitial Capital with Ivestmet ad -ivestmet Discrete Time

More information

ON POINTWISE BINOMIAL APPROXIMATION

ON POINTWISE BINOMIAL APPROXIMATION Iteratioal Joural of Pure ad Applied Mathematics Volume 71 No. 1 2011, 57-66 ON POINTWISE BINOMIAL APPROXIMATION BY w-functions K. Teerapabolar 1, P. Wogkasem 2 Departmet of Mathematics Faculty of Sciece

More information

PROBLEM SET 5 SOLUTIONS 126 = , 37 = , 15 = , 7 = 7 1.

PROBLEM SET 5 SOLUTIONS 126 = , 37 = , 15 = , 7 = 7 1. Math 7 Sprig 06 PROBLEM SET 5 SOLUTIONS Notatios. Give a real umber x, we will defie sequeces (a k ), (x k ), (p k ), (q k ) as i lecture.. (a) (5 pts) Fid the simple cotiued fractio represetatios of 6

More information

An Introduction to Randomized Algorithms

An Introduction to Randomized Algorithms A Itroductio to Radomized Algorithms The focus of this lecture is to study a radomized algorithm for quick sort, aalyze it usig probabilistic recurrece relatios, ad also provide more geeral tools for aalysis

More information

Chapter 8. Euler s Gamma function

Chapter 8. Euler s Gamma function Chapter 8 Euler s Gamma fuctio The Gamma fuctio plays a importat role i the fuctioal equatio for ζ(s that we will derive i the ext chapter. I the preset chapter we have collected some properties of the

More information

Exponential Functions and Taylor Series

Exponential Functions and Taylor Series MATH 4530: Aalysis Oe Expoetial Fuctios ad Taylor Series James K. Peterso Departmet of Biological Scieces ad Departmet of Mathematical Scieces Clemso Uiversity March 29, 2017 MATH 4530: Aalysis Oe Outlie

More information

Iterative Techniques for Solving Ax b -(3.8). Assume that the system has a unique solution. Let x be the solution. Then x A 1 b.

Iterative Techniques for Solving Ax b -(3.8). Assume that the system has a unique solution. Let x be the solution. Then x A 1 b. Iterative Techiques for Solvig Ax b -(8) Cosider solvig liear systems of them form: Ax b where A a ij, x x i, b b i Assume that the system has a uique solutio Let x be the solutio The x A b Jacobi ad Gauss-Seidel

More information

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15 17. Joit distributios of extreme order statistics Lehma 5.1; Ferguso 15 I Example 10., we derived the asymptotic distributio of the maximum from a radom sample from a uiform distributio. We did this usig

More information

Precise Rates in Complete Moment Convergence for Negatively Associated Sequences

Precise Rates in Complete Moment Convergence for Negatively Associated Sequences Commuicatios of the Korea Statistical Society 29, Vol. 16, No. 5, 841 849 Precise Rates i Complete Momet Covergece for Negatively Associated Sequeces Dae-Hee Ryu 1,a a Departmet of Computer Sciece, ChugWoo

More information

1 Convergence in Probability and the Weak Law of Large Numbers

1 Convergence in Probability and the Weak Law of Large Numbers 36-752 Advaced Probability Overview Sprig 2018 8. Covergece Cocepts: i Probability, i L p ad Almost Surely Istructor: Alessadro Rialdo Associated readig: Sec 2.4, 2.5, ad 4.11 of Ash ad Doléas-Dade; Sec

More information

Web-based Supplementary Materials for A Modified Partial Likelihood Score Method for Cox Regression with Covariate Error Under the Internal

Web-based Supplementary Materials for A Modified Partial Likelihood Score Method for Cox Regression with Covariate Error Under the Internal Web-based Supplemetary Materials for A Modified Partial Likelihood Score Method for Cox Regressio with Covariate Error Uder the Iteral Validatio Desig by David M. Zucker, Xi Zhou, Xiaomei Liao, Yi Li,

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013. Large Deviations for i.i.d. Random Variables MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 2 9/9/2013 Large Deviatios for i.i.d. Radom Variables Cotet. Cheroff boud usig expoetial momet geeratig fuctios. Properties of a momet

More information

On Nonsingularity of Saddle Point Matrices. with Vectors of Ones

On Nonsingularity of Saddle Point Matrices. with Vectors of Ones Iteratioal Joural of Algebra, Vol. 2, 2008, o. 4, 197-204 O Nosigularity of Saddle Poit Matrices with Vectors of Oes Tadeusz Ostrowski Istitute of Maagemet The State Vocatioal Uiversity -400 Gorzów, Polad

More information

PAPER : IIT-JAM 2010

PAPER : IIT-JAM 2010 MATHEMATICS-MA (CODE A) Q.-Q.5: Oly oe optio is correct for each questio. Each questio carries (+6) marks for correct aswer ad ( ) marks for icorrect aswer.. Which of the followig coditios does NOT esure

More information

MATH 112: HOMEWORK 6 SOLUTIONS. Problem 1: Rudin, Chapter 3, Problem s k < s k < 2 + s k+1

MATH 112: HOMEWORK 6 SOLUTIONS. Problem 1: Rudin, Chapter 3, Problem s k < s k < 2 + s k+1 MATH 2: HOMEWORK 6 SOLUTIONS CA PRO JIRADILOK Problem. If s = 2, ad Problem : Rudi, Chapter 3, Problem 3. s + = 2 + s ( =, 2, 3,... ), prove that {s } coverges, ad that s < 2 for =, 2, 3,.... Proof. The

More information

36-755, Fall 2017 Homework 5 Solution Due Wed Nov 15 by 5:00pm in Jisu s mailbox

36-755, Fall 2017 Homework 5 Solution Due Wed Nov 15 by 5:00pm in Jisu s mailbox Poits: 00+ pts total for the assigmet 36-755, Fall 07 Homework 5 Solutio Due Wed Nov 5 by 5:00pm i Jisu s mailbox We first review some basic relatios with orms ad the sigular value decompositio o matrices

More information

Complex Analysis Spring 2001 Homework I Solution

Complex Analysis Spring 2001 Homework I Solution Complex Aalysis Sprig 2001 Homework I Solutio 1. Coway, Chapter 1, sectio 3, problem 3. Describe the set of poits satisfyig the equatio z a z + a = 2c, where c > 0 ad a R. To begi, we see from the triagle

More information