An introduction to chaining, and applications to sublinear algorithms

Size: px
Start display at page:

Download "An introduction to chaining, and applications to sublinear algorithms"

Transcription

1 An ntroducton to channg, and applcatons to sublnear algorthms Jelan Nelson Harvard August 28, 2015

2 What s ths talk about?

3 What s ths talk about? Gven a collecton of random varables X 1, X 2,...,, we would lke to say that max X s small wth hgh probablty. (Happens all over computer scence, e.g. Chernon (Chernoff+Unon) bound)

4 What s ths talk about? Gven a collecton of random varables X 1, X 2,...,, we would lke to say that max X s small wth hgh probablty. (Happens all over computer scence, e.g. Chernon (Chernoff+Unon) bound) Today s topc: Beatng the Unon Bound

5 What s ths talk about? Gven a collecton of random varables X 1, X 2,...,, we would lke to say that max X s small wth hgh probablty. (Happens all over computer scence, e.g. Chernon (Chernoff+Unon) bound) Today s topc: Beatng the Unon Bound Dsclamer: Ths s an educatonal talk, about deas whch aren t mne.

6 T B l n 2 A frst example

7 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres

8 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres Defne gaussan mean wdth g(t ) = E g sup Z x

9 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres Defne gaussan mean wdth g(t ) = E g sup Z x How can we bound g(t )?

10 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres Defne gaussan mean wdth g(t ) = E g sup Z x How can we bound g(t )? Ths talk: four progressvely tghter ways to bound g(t ), then applcatons of technques to some TCS problems

11 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x

12 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one

13 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one E sup Z x = = 0 u 0 P(sup Z x > u)du P(sup Z x > u) } {{ } 1 u + T e u2 /2 du + u log T (set u = 2 log T ) P(sup Z x > u) }{{} T e u2 /2 (unon bound) du

14 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one E sup Z x = = 0 u 0 P(sup Z x > u)du P(sup Z x > u) } {{ } 1 u + T e u2 /2 du + u log T (set u = 2 log T ) P(sup Z x > u) }{{} T e u2 /2 (unon bound) du

15 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one E sup Z x = = 0 u 0 P(sup Z x > u)du P(sup Z x > u) } {{ } 1 u + T e u2 /2 du + u log T (set u = 2 log T ) P(sup Z x > u) }{{} T e u2 /2 (unon bound) du

16 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 )

17 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 ) g, x = g, x + g, x x (x = argmn y T x y 2 ) g(t ) g(s ε ) + E g sup g, x x }{{} ε g 2

18 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 ) g, x = g, x + g, x x (x = argmn y T x y 2 ) g(t ) g(s ε ) + E g sup g, x x }{{} ε g 2 log S ε + ε(e g g 2 2 )1/2 log 1/2 N (T, l 2, ε) +ε n }{{} smallest ε net sze

19 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 ) g, x = g, x + g, x x (x = argmn y T x y 2 ) g(t ) g(s ε ) + E g sup g, x x }{{} ε g 2 log S ε + ε(e g g 2 2 )1/2 log 1/2 N (T, l 2, ε) }{{} smallest ε net sze +ε n Choose ε to optmze bound; can never be worse than last slde (whch amounts to choosng ε = 0)

20 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x

21 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx

22 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx g(t ) E sup g, π 0 x g } {{ } 0 + k=1 E g sup g, k x

23 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx g(t ) E sup g, π 0 x g + k=1 E g sup g, k x } {{ } 0 { k x : x T } N (T, l 2, 1 /2 k ) N (T, l 2, 1 /2 k 1 ) (N (T, l 2, 1 /2 k )) 2

24 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx g(t ) E sup g, π 0 x g + k=1 E g sup g, k x } {{ } 0 { k x : x T } N (T, l 2, 1 /2 k ) N (T, l 2, 1 /2 k 1 ) (N (T, l 2, 1 /2 k )) 2 g(t ) k=1 (1 /2 k ) log 1/2 N (T, l 2, 1 /2 k ) 0 log 1/2 N (T, l 2, u)du (Dudley s theorem)

25 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble )

26 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble ) Exercse: show Dudley s theorem s equvalent to g(t ) nf {Tk } admssble k=1 2k/2 sup d l2 (x, T k ) (should pck T k to be the best ε = ε(k) net of sze 2 2k )

27 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble ) Exercse: show Dudley s theorem s equvalent to g(t ) nf {Tk } admssble k=1 2k/2 sup d l2 (x, T k ) (should pck T k to be the best ε = ε(k) net of sze 2 2k ) Fernque 76 : can pull the sup x outsde the sum g(t ) nf {Tk } sup k=1 2k/2 d l2 (x, T k ) def = γ 2 (T, l 2 )

28 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble ) Exercse: show Dudley s theorem s equvalent to g(t ) nf {Tk } admssble k=1 2k/2 sup d l2 (x, T k ) (should pck T k to be the best ε = ε(k) net of sze 2 2k ) Fernque 76 : can pull the sup x outsde the sum g(t ) nf {Tk } sup k=1 2k/2 d l2 (x, T k ) def = γ 2 (T, l 2 ) equvalent upper bound proven by Fernque (who mnmzed some ntegral over all measures over T ), but reformulated n terms of admssble sequences by Talgarand

29 Gaussan mean wdth bound 4: generc channg Proof of Fernque s bound g(t ) E sup g, π 0 x g } {{ } 0 + E g sup g, k x }{{} Y k k=1 (from before) t, P(Y k > t2 k/2 k x 2 ) e t2 2 k /2 (gaussan decay)

30 Gaussan mean wdth bound 4: generc channg Proof of Fernque s bound g(t ) E sup g, π 0 x g } {{ } 0 + E g sup g, k x }{{} Y k k=1 (from before) t, P(Y k > t2 k/2 k x 2 ) e t2 2 k /2 (gaussan decay) P( x, k Y k > t2 k/2 k x 2 ) k (22k ) 2 e t2 2 k /2

31 Gaussan mean wdth bound 4: generc channg Proof of Fernque s bound g(t ) E sup g, π 0 x g } {{ } 0 + E g sup g, k x }{{} Y k k=1 (from before) t, P(Y k > t2 k/2 k x 2 ) e t2 2 k /2 (gaussan decay) P( x, k Y k > t2 k/2 k x 2 ) k (22k ) 2 e t2 2 k /2 E g sup Y k = k 0 P(sup Y k > u)du k

32 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) k = γ 2 (T, l 2 ) [t + (2 2k ) 2 e t2 2 k /2 dt] γ 2 (T, l 2 ) t k=1 k Concluson: g(t ) γ 2 (T, l 2 ) Talagrand: g(t ) γ 2 (T, l 2 ) (won t show today) ( Majorzng measures theorem )

33 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) k = γ 2 (T, l 2 ) [t + (2 2k ) 2 e t2 2 k /2 dt] γ 2 (T, l 2 ) t k=1 k

34 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) γ 2 (T, l 2 ) [2 + γ 2 (T, l 2 ) 2 k ( ) (2 2k ) 2 e t2 2 k /2 dt] k=1 k

35 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) γ 2 (T, l 2 ) [2 + γ 2 (T, l 2 ) 2 k ( ) (2 2k ) 2 e t2 2 k /2 dt] k=1 k Concluson: g(t ) γ 2 (T, l 2 )

36 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) γ 2 (T, l 2 ) [2 + γ 2 (T, l 2 ) 2 k ( ) (2 2k ) 2 e t2 2 k /2 dt] k=1 k Concluson: g(t ) γ 2 (T, l 2 ) Talagrand: g(t ) γ 2 (T, l 2 ) (won t show today) ( Majorzng measures theorem )

37 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du

38 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1

39 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1 sup x Bl n 1 g, x = g, so g(t ) log n Exercse: Come up wth admssble {T k } yeldng γ 2 log n (must exst by majorzng measures)

40 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1 sup x Bl n 1 g, x = g, so g(t ) log n Exercse: Come up wth admssble {T k } yeldng γ 2 log n (must exst by majorzng measures) Dudley: log N (B l n 1, l 2, u) ( 1 /u 2 ) log n for u not too small (consder just coverng ( 1 /u 2 )-sparse vectors wth u 2 n each coordnate). Dudley can only gve g(b l n 1 ) log 3/2 n.

41 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1 sup x Bl n 1 g, x = g, so g(t ) log n Exercse: Come up wth admssble {T k } yeldng γ 2 log n (must exst by majorzng measures) Dudley: log N (B l n 1, l 2, u) ( 1 /u 2 ) log n for u not too small (consder just coverng ( 1 /u 2 )-sparse vectors wth u 2 n each coordnate). Dudley can only gve g(b l n 1 ) log 3/2 n. Smple vanlla ε-net argument gves g(b l n 1 ) poly(n).

42 Hgh probablty So far just talked about g(t ) = E g sup Z x But what f we want to know sup Z x s small whp, not just n expectaton?

43 Hgh probablty So far just talked about g(t ) = E g sup Z x But what f we want to know sup Z x s small whp, not just n expectaton? Usual approach: bound E g sup Zx p for large p and do Markov ( moment method ) Can bound moments usng channg too; see (Drksen 13)

44 Applcatons n computer scence Fast RIP matrces (Candès, Tao 06), (Rudelson, Vershynn 06), (Cheragch, Guruswam, Velngker 13), (N., Prce, Wootters 14), (Bourgan 14), (Havv, Regev 15) Fast JL (Alon, Lberty 11), (Krahmer, Ward 11), (Bourgan, Drksen, N. 15), (Oymak, Recht, Soltanolkotab 15) Instance-wse JL bounds (Gordon 88), (Klartag, Mendelson 05), (Mendelson, Pajor, Tomczak-Jaegermann 07), (Drksen 14) Approxmate nearest neghbor (Indyk, Naor 07) Determnstc algorthm to estmate graph cover tme (Dng, Lee, Peres 11) Lst-decodablty of random codes (Wootters 13), (Rudra, Wootters 14)...

45 A channg result for quadratc forms Theorem [Krahmer, Mendelson, Rauhut 14] Let A R n n be a famly of matrces, and let σ 1,..., σ n be ndependent subgaussans. Then E sup Aσ 2 2 E Aσ 2 A A σ 2 γ 2 2(A, l2 l 2 ) + γ 2 (A, l2 l 2 ) F (A) + l2 l 2 (A) F (A) ( X s dameter under X -norm)

46 A channg result for quadratc forms Theorem [Krahmer, Mendelson, Rauhut 14] Let A R n n be a famly of matrces, and let σ 1,..., σ n be ndependent subgaussans. Then E sup Aσ 2 2 E Aσ 2 A A σ 2 γ 2 2(A, l2 l 2 ) + γ 2 (A, l2 l 2 ) F (A) + l2 l 2 (A) F (A) ( X s dameter under X -norm) Won t show proof today, but t s smlar to boundng g(t ) (wth some extra trcks). See edu/ mnlek/madalgo2015/, Lecture 3.

47 Instance-wse bounds for JL Corollary (Gordon 88, Klartag-Mendelson 05, Mendelson, Pajor, Tomczak-Jaegermann 07, Drksen 14) For T S n 1 and 0 < ε < 1/2, let Π R m n have ndependent subgaussan ndependent entres wth mean zero and varance 1/m for m (g2 (T )+1)/ε 2. Then E sup Πx < ε Π

48 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n

49 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m).

50 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t )

51 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t ) F (A T ) = 1, l2 l 2 (A T ) = 1/ m

52 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t ) F (A T ) = 1, l2 l 2 (A T ) = 1/ m Thus E Π sup Πx g2 (T )/m + g(t ) / m + 1 / m

53 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t ) F (A T ) = 1, l2 l 2 (A T ) = 1/ m Thus E Π sup Πx g2 (T )/m + g(t ) / m + 1 / m Set m (g2 (T )+1)/ε 2

54 Consequences of Gordon s theorem m (g2 (T )+1)/ε 2 T < : g 2 (T ) log T (JL) T a d-dm subspace: g 2 (T ) d (subspace embeddngs) T all k-sparse vectors: g 2 (T ) k log(n/k) (RIP)

55 Consequences of Gordon s theorem m (g2 (T )+1)/ε 2 T < : g 2 (T ) log T (JL) T a d-dm subspace: g 2 (T ) d (subspace embeddngs) T all k-sparse vectors: g 2 (T ) k log(n/k) (RIP) more applcatons to constraned least squares, manfold learnng, model-based compressed sensng,... (see (Drksen 14) and (Bourgan, Drksen, N. 15))

56 Channg sn t just for gaussans

57 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) Restrcted sometry property useful n compressed sensng. T = {x : x 0 k, x 2 = 1}. Theorem (Candès-Tao 06, Donoho 06, Candés 08) If Π satsfes (ε, k)-rip for ε < 2 1 then there s a lnear program whch, gven Πx and Π as nput, recovers x n polynomal tme such that x x 2 O( 1 / k) mn y 0 k x y 1.

58 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) Restrcted sometry property useful n compressed sensng. T = {x : x 0 k, x 2 = 1}. Theorem (Candès-Tao 06, Donoho 06, Candés 08) If Π satsfes (ε, k)-rip for ε < 2 1 then there s a lnear program whch, gven Πx and Π as nput, recovers x n polynomal tme such that x x 2 O( 1 / k) mn y 0 k x y 1. Of nterest to show samplng rows of dscrete Fourer matrx s RIP

59 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) (Unnormalzed) Fourer matrx F, rows: z 1,..., z n δ 1,..., δ n ndependent Bernoull wth expectaton m/n

60 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) (Unnormalzed) Fourer matrx F, rows: z 1,..., z n δ 1,..., δ n ndependent Bernoull wth expectaton m/n Want E δ sup T [n] T k I T 1 m n δ z (T ) =1 (T ) z < ε

61 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ 1 m E = sup T [n] T k sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1

62 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1

63 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g sup δ,g T sup x B n,k 2 T (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1

64 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1

65 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1

66 The End

67 June 22 nd +23 rd : workshop on concentraton of measure / channg at Harvard, after STOC 16. Detals+webste forthcomng.

Dimensionality Reduction Notes 1

Dimensionality Reduction Notes 1 Dmensonalty Reducton Notes 1 Jelan Nelson mnlek@seas.harvard.edu August 10, 2015 1 Prelmnares Here we collect some notaton and basc lemmas used throughout ths note. Throughout, for a random varable X,

More information

HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION

HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION MARK RUDELSON AND ROMAN VERSHYNIN Abstract. In ths expostory note, we gve a modern proof of Hanson-Wrght nequalty for quadratc forms n sub-gaussan

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017 U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that

More information

Which Separator? Spring 1

Which Separator? Spring 1 Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal

More information

Lecture 9 Sept 29, 2017

Lecture 9 Sept 29, 2017 Sketchng Algorthms for Bg Data Fall 2017 Prof. Jelan Nelson Lecture 9 Sept 29, 2017 Scrbe: Mtal Bafna 1 Fast JL transform Typcally we have some hgh-mensonal computatonal geometry problem, an we use JL

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

CIS 700: algorithms for Big Data

CIS 700: algorithms for Big Data CIS 700: algorthms for Bg Data Lecture 5: Dmenson Reducton Sldes at htt://grgory.us/bg-data-class.html Grgory Yaroslavtsev htt://grgory.us Today Dmensonalty reducton AMS as dmensonalty reducton Johnson-Lndenstrauss

More information

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD Matrx Approxmaton va Samplng, Subspace Embeddng Lecturer: Anup Rao Scrbe: Rashth Sharma, Peng Zhang 0/01/016 1 Solvng Lnear Systems Usng SVD Two applcatons of SVD have been covered so far. Today we loo

More information

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable

More information

P exp(tx) = 1 + t 2k M 2k. k N

P exp(tx) = 1 + t 2k M 2k. k N 1. Subgaussan tals Defnton. Say that a random varable X has a subgaussan dstrbuton wth scale factor σ< f P exp(tx) exp(σ 2 t 2 /2) for all real t. For example, f X s dstrbuted N(,σ 2 ) then t s subgaussan.

More information

18.1 Introduction and Recap

18.1 Introduction and Recap CS787: Advanced Algorthms Scrbe: Pryananda Shenoy and Shjn Kong Lecturer: Shuch Chawla Topc: Streamng Algorthmscontnued) Date: 0/26/2007 We contnue talng about streamng algorthms n ths lecture, ncludng

More information

Dimensionality Reduction Notes 2

Dimensionality Reduction Notes 2 Dmensonalty Reducton Notes 2 Jelan Nelson mnlek@seas.harvard.edu August 11, 2015 1 Optmalty theorems for JL Yesterday we saw for MJL that we could acheve target dmenson m = O(ε 2 log N), and for DJL we

More information

Mining Data Streams-Estimating Frequency Moment

Mining Data Streams-Estimating Frequency Moment Mnng Data Streams-Estmatng Frequency Moment Barna Saha October 26, 2017 Frequency Moment Computng moments nvolves dstrbuton of frequences of dfferent elements n the stream. Frequency Moment Computng moments

More information

HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION

HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION MARK RUDELSON AND ROMAN VERSHYNIN Abstract. In ths expostory note, we gve a modern proof of Hanson-Wrght nequalty for quadratc forms n sub-gaussan

More information

Finding Dense Subgraphs in G(n, 1/2)

Finding Dense Subgraphs in G(n, 1/2) Fndng Dense Subgraphs n Gn, 1/ Atsh Das Sarma 1, Amt Deshpande, and Rav Kannan 1 Georga Insttute of Technology,atsh@cc.gatech.edu Mcrosoft Research-Bangalore,amtdesh,annan@mcrosoft.com Abstract. Fndng

More information

Lecture 4: September 12

Lecture 4: September 12 36-755: Advanced Statstcal Theory Fall 016 Lecture 4: September 1 Lecturer: Alessandro Rnaldo Scrbe: Xao Hu Ta Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer: These notes have not been

More information

Lecture 3. Ax x i a i. i i

Lecture 3. Ax x i a i. i i 18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest

More information

Lecture 10: May 6, 2013

Lecture 10: May 6, 2013 TTIC/CMSC 31150 Mathematcal Toolkt Sprng 013 Madhur Tulsan Lecture 10: May 6, 013 Scrbe: Wenje Luo In today s lecture, we manly talked about random walk on graphs and ntroduce the concept of graph expander,

More information

Lecture 4: Universal Hash Functions/Streaming Cont d

Lecture 4: Universal Hash Functions/Streaming Cont d CSE 5: Desgn and Analyss of Algorthms I Sprng 06 Lecture 4: Unversal Hash Functons/Streamng Cont d Lecturer: Shayan Oves Gharan Aprl 6th Scrbe: Jacob Schreber Dsclamer: These notes have not been subjected

More information

Dimensionality Reduction Notes 3

Dimensionality Reduction Notes 3 Dimensionality Reduction Notes 3 Jelani Nelson minilek@seas.harvard.edu August 13, 2015 1 Gordon s theorem Let T be a finite subset of some normed vector space with norm X. We say that a sequence T 0 T

More information

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution. Solutons HW #2 Dual of general LP. Fnd the dual functon of the LP mnmze subject to c T x Gx h Ax = b. Gve the dual problem, and make the mplct equalty constrants explct. Soluton. 1. The Lagrangan s L(x,

More information

Lecture 3 January 31, 2017

Lecture 3 January 31, 2017 CS 224: Advanced Algorthms Sprng 207 Prof. Jelan Nelson Lecture 3 January 3, 207 Scrbe: Saketh Rama Overvew In the last lecture we covered Y-fast tres and Fuson Trees. In ths lecture we start our dscusson

More information

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0 MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector

More information

Lecture 4: Constant Time SVD Approximation

Lecture 4: Constant Time SVD Approximation Spectral Algorthms and Representatons eb. 17, Mar. 3 and 8, 005 Lecture 4: Constant Tme SVD Approxmaton Lecturer: Santosh Vempala Scrbe: Jangzhuo Chen Ths topc conssts of three lectures 0/17, 03/03, 03/08),

More information

Hanson-Wright inequality and sub-gaussian concentration

Hanson-Wright inequality and sub-gaussian concentration Electron. Commun. Probab. 18 (013, no. 8, 1 9. DOI: 10.114/ECP.v18-865 ISSN: 1083-589X ELECTRONIC COMMUNICATIONS n PROBABILITY Hanson-Wrght nequalty and sub-gaussan concentraton Mark Rudelson Roman Vershynn

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 6 Luca Trevisan September 12, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 6 Luca Trevisan September 12, 2017 U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 6 Luca Trevsan September, 07 Scrbed by Theo McKenze Lecture 6 In whch we study the spectrum of random graphs. Overvew When attemptng to fnd n polynomal

More information

Support Vector Machines

Support Vector Machines Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Differentiating Gaussian Processes

Differentiating Gaussian Processes Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the

More information

COS 521: Advanced Algorithms Game Theory and Linear Programming

COS 521: Advanced Algorithms Game Theory and Linear Programming COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton

More information

Feature Selection: Part 1

Feature Selection: Part 1 CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Eigenvalues of Random Graphs

Eigenvalues of Random Graphs Spectral Graph Theory Lecture 2 Egenvalues of Random Graphs Danel A. Spelman November 4, 202 2. Introducton In ths lecture, we consder a random graph on n vertces n whch each edge s chosen to be n the

More information

Lecture 5 September 17, 2015

Lecture 5 September 17, 2015 CS 229r: Algorthms for Bg Data Fall 205 Prof. Jelan Nelson Lecture 5 September 7, 205 Scrbe: Yakr Reshef Recap and overvew Last tme we dscussed the problem of norm estmaton for p-norms wth p > 2. We had

More information

10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)

10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014) 0-80: Advanced Optmzaton and Randomzed Methods Lecture : Convex functons (Jan 5, 04) Lecturer: Suvrt Sra Addr: Carnege Mellon Unversty, Sprng 04 Scrbes: Avnava Dubey, Ahmed Hefny Dsclamer: These notes

More information

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011 Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Properties of Least Squares

Properties of Least Squares Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures

More information

Support Vector Machines

Support Vector Machines /14/018 Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x

More information

Notes on Frequency Estimation in Data Streams

Notes on Frequency Estimation in Data Streams Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to

More information

Introduction to Algorithms

Introduction to Algorithms Introducton to Algorthms 6.046J/8.40J Lecture 7 Prof. Potr Indyk Data Structures Role of data structures: Encapsulate data Support certan operatons (e.g., INSERT, DELETE, SEARCH) Our focus: effcency of

More information

1 The Mistake Bound Model

1 The Mistake Bound Model 5-850: Advanced Algorthms CMU, Sprng 07 Lecture #: Onlne Learnng and Multplcatve Weghts February 7, 07 Lecturer: Anupam Gupta Scrbe: Bryan Lee,Albert Gu, Eugene Cho he Mstake Bound Model Suppose there

More information

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES

TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES SVANTE JANSON Abstract. We gve explct bounds for the tal probabltes for sums of ndependent geometrc or exponental varables, possbly wth dfferent

More information

Lecture 21: Numerical methods for pricing American type derivatives

Lecture 21: Numerical methods for pricing American type derivatives Lecture 21: Numercal methods for prcng Amercan type dervatves Xaoguang Wang STAT 598W Aprl 10th, 2014 (STAT 598W) Lecture 21 1 / 26 Outlne 1 Fnte Dfference Method Explct Method Penalty Method (STAT 598W)

More information

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009 College of Computer & Informaton Scence Fall 2009 Northeastern Unversty 20 October 2009 CS7880: Algorthmc Power Tools Scrbe: Jan Wen and Laura Poplawsk Lecture Outlne: Prmal-dual schema Network Desgn:

More information

Maximum Likelihood Estimation (MLE)

Maximum Likelihood Estimation (MLE) Maxmum Lkelhood Estmaton (MLE) Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 01 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y

More information

Lecture 3: Dual problems and Kernels

Lecture 3: Dual problems and Kernels Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture # 15 Scribe: Jieming Mao April 1, 2013

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture # 15 Scribe: Jieming Mao April 1, 2013 COS 511: heoretcal Machne Learnng Lecturer: Rob Schapre Lecture # 15 Scrbe: Jemng Mao Aprl 1, 013 1 Bref revew 1.1 Learnng wth expert advce Last tme, we started to talk about learnng wth expert advce.

More information

Lecture 20: November 7

Lecture 20: November 7 0-725/36-725: Convex Optmzaton Fall 205 Lecturer: Ryan Tbshran Lecture 20: November 7 Scrbes: Varsha Chnnaobreddy, Joon Sk Km, Lngyao Zhang Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer:

More information

Linear Classification, SVMs and Nearest Neighbors

Linear Classification, SVMs and Nearest Neighbors 1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

Vapnik-Chervonenkis theory

Vapnik-Chervonenkis theory Vapnk-Chervonenks theory Rs Kondor June 13, 2008 For the purposes of ths lecture, we restrct ourselves to the bnary supervsed batch learnng settng. We assume that we have an nput space X, and an unknown

More information

Probability Theory (revisited)

Probability Theory (revisited) Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

Lecture 10 Support Vector Machines. Oct

Lecture 10 Support Vector Machines. Oct Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron

More information

Randomness and Computation

Randomness and Computation Randomness and Computaton or, Randomzed Algorthms Mary Cryan School of Informatcs Unversty of Ednburgh RC 208/9) Lecture 0 slde Balls n Bns m balls, n bns, and balls thrown unformly at random nto bns usually

More information

The Second Anti-Mathima on Game Theory

The Second Anti-Mathima on Game Theory The Second Ant-Mathma on Game Theory Ath. Kehagas December 1 2006 1 Introducton In ths note we wll examne the noton of game equlbrum for three types of games 1. 2-player 2-acton zero-sum games 2. 2-player

More information

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and

More information

NP-Completeness : Proofs

NP-Completeness : Proofs NP-Completeness : Proofs Proof Methods A method to show a decson problem Π NP-complete s as follows. (1) Show Π NP. (2) Choose an NP-complete problem Π. (3) Show Π Π. A method to show an optmzaton problem

More information

Lagrange Multipliers Kernel Trick

Lagrange Multipliers Kernel Trick Lagrange Multplers Kernel Trck Ncholas Ruozz Unversty of Texas at Dallas Based roughly on the sldes of Davd Sontag General Optmzaton A mathematcal detour, we ll come back to SVMs soon! subject to: f x

More information

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to THE INVERSE POWER METHOD (or INVERSE ITERATION) -- applcaton of the Power method to A some fxed constant ρ (whch s called a shft), x λ ρ If the egenpars of A are { ( λ, x ) } ( ), or (more usually) to,

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

Linear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.

Linear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space. Lnear, affne, and convex sets and hulls In the sequel, unless otherwse specfed, X wll denote a real vector space. Lnes and segments. Gven two ponts x, y X, we defne xy = {x + t(y x) : t R} = {(1 t)x +

More information

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS SECTION 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS 493 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS All the vector spaces you have studed thus far n the text are real vector spaces because the scalars

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

Meshless Surfaces. presented by Niloy J. Mitra. An Nguyen

Meshless Surfaces. presented by Niloy J. Mitra. An Nguyen Meshless Surfaces presented by Nloy J. Mtra An Nguyen Outlne Mesh-Independent Surface Interpolaton D. Levn Outlne Mesh-Independent Surface Interpolaton D. Levn Pont Set Surfaces M. Alexa, J. Behr, D. Cohen-Or,

More information

Google PageRank with Stochastic Matrix

Google PageRank with Stochastic Matrix Google PageRank wth Stochastc Matrx Md. Sharq, Puranjt Sanyal, Samk Mtra (M.Sc. Applcatons of Mathematcs) Dscrete Tme Markov Chan Let S be a countable set (usually S s a subset of Z or Z d or R or R d

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

Communication Complexity 16:198: February Lecture 4. x ij y ij

Communication Complexity 16:198: February Lecture 4. x ij y ij Communcaton Complexty 16:198:671 09 February 2010 Lecture 4 Lecturer: Troy Lee Scrbe: Rajat Mttal 1 Homework problem : Trbes We wll solve the thrd queston n the homework. The goal s to show that the nondetermnstc

More information

6.842 Randomness and Computation February 18, Lecture 4

6.842 Randomness and Computation February 18, Lecture 4 6.842 Randomness and Computaton February 18, 2014 Lecture 4 Lecturer: Rontt Rubnfeld Scrbe: Amartya Shankha Bswas Topcs 2-Pont Samplng Interactve Proofs Publc cons vs Prvate cons 1 Two Pont Samplng 1.1

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

Lecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.

Lecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem. prnceton u. sp 02 cos 598B: algorthms and complexty Lecture 20: Lft and Project, SDP Dualty Lecturer: Sanjeev Arora Scrbe:Yury Makarychev Today we wll study the Lft and Project method. Then we wll prove

More information

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there

More information

p 1 c 2 + p 2 c 2 + p 3 c p m c 2

p 1 c 2 + p 2 c 2 + p 3 c p m c 2 Where to put a faclty? Gven locatons p 1,..., p m n R n of m houses, want to choose a locaton c n R n for the fre staton. Want c to be as close as possble to all the house. We know how to measure dstance

More information

Stanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7

Stanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7 Stanford Unversty CS54: Computatonal Complexty Notes 7 Luca Trevsan January 9, 014 Notes for Lecture 7 1 Approxmate Countng wt an N oracle We complete te proof of te followng result: Teorem 1 For every

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13 CME 30: NUMERICAL LINEAR ALGEBRA FALL 005/06 LECTURE 13 GENE H GOLUB 1 Iteratve Methods Very large problems (naturally sparse, from applcatons): teratve methods Structured matrces (even sometmes dense,

More information

Generative classification models

Generative classification models CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn

More information

Error Bars in both X and Y

Error Bars in both X and Y Error Bars n both X and Y Wrong ways to ft a lne : 1. y(x) a x +b (σ x 0). x(y) c y + d (σ y 0) 3. splt dfference between 1 and. Example: Prmordal He abundance: Extrapolate ft lne to [ O / H ] 0. [ He

More information

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

THE WEIGHTED WEAK TYPE INEQUALITY FOR THE STRONG MAXIMAL FUNCTION

THE WEIGHTED WEAK TYPE INEQUALITY FOR THE STRONG MAXIMAL FUNCTION THE WEIGHTED WEAK TYPE INEQUALITY FO THE STONG MAXIMAL FUNCTION THEMIS MITSIS Abstract. We prove the natural Fefferman-Sten weak type nequalty for the strong maxmal functon n the plane, under the assumpton

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

Quantifying Uncertainty

Quantifying Uncertainty Partcle Flters Quantfyng Uncertanty Sa Ravela M. I. T Last Updated: Sprng 2013 1 Quantfyng Uncertanty Partcle Flters Partcle Flters Appled to Sequental flterng problems Can also be appled to smoothng problems

More information

Week 2. This week, we covered operations on sets and cardinality.

Week 2. This week, we covered operations on sets and cardinality. Week 2 Ths week, we covered operatons on sets and cardnalty. Defnton 0.1 (Correspondence). A correspondence between two sets A and B s a set S contaned n A B = {(a, b) a A, b B}. A correspondence from

More information

2.3 Nilpotent endomorphisms

2.3 Nilpotent endomorphisms s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms

More information

APPENDIX A Some Linear Algebra

APPENDIX A Some Linear Algebra APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,

More information

Calculation of time complexity (3%)

Calculation of time complexity (3%) Problem 1. (30%) Calculaton of tme complexty (3%) Gven n ctes, usng exhaust search to see every result takes O(n!). Calculaton of tme needed to solve the problem (2%) 40 ctes:40! dfferent tours 40 add

More information

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING 1 ADVANCED ACHINE LEARNING ADVANCED ACHINE LEARNING Non-lnear regresson technques 2 ADVANCED ACHINE LEARNING Regresson: Prncple N ap N-dm. nput x to a contnuous output y. Learn a functon of the type: N

More information

20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The first idea is connectedness.

20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The first idea is connectedness. 20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The frst dea s connectedness. Essentally, we want to say that a space cannot be decomposed

More information

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016 U.C. Berkeley CS94: Spectral Methods and Expanders Handout 8 Luca Trevsan February 7, 06 Lecture 8: Spectral Algorthms Wrap-up In whch we talk about even more generalzatons of Cheeger s nequaltes, and

More information

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14 APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

15-381: Artificial Intelligence. Regression and cross validation

15-381: Artificial Intelligence. Regression and cross validation 15-381: Artfcal Intellgence Regresson and cross valdaton Where e are Inputs Densty Estmator Probablty Inputs Classfer Predct category Inputs Regressor Predct real no. Today Lnear regresson Gven an nput

More information

The conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above

The conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ

More information