An introduction to chaining, and applications to sublinear algorithms
|
|
- Flora Garrett
- 5 years ago
- Views:
Transcription
1 An ntroducton to channg, and applcatons to sublnear algorthms Jelan Nelson Harvard August 28, 2015
2 What s ths talk about?
3 What s ths talk about? Gven a collecton of random varables X 1, X 2,...,, we would lke to say that max X s small wth hgh probablty. (Happens all over computer scence, e.g. Chernon (Chernoff+Unon) bound)
4 What s ths talk about? Gven a collecton of random varables X 1, X 2,...,, we would lke to say that max X s small wth hgh probablty. (Happens all over computer scence, e.g. Chernon (Chernoff+Unon) bound) Today s topc: Beatng the Unon Bound
5 What s ths talk about? Gven a collecton of random varables X 1, X 2,...,, we would lke to say that max X s small wth hgh probablty. (Happens all over computer scence, e.g. Chernon (Chernoff+Unon) bound) Today s topc: Beatng the Unon Bound Dsclamer: Ths s an educatonal talk, about deas whch aren t mne.
6 T B l n 2 A frst example
7 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres
8 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres Defne gaussan mean wdth g(t ) = E g sup Z x
9 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres Defne gaussan mean wdth g(t ) = E g sup Z x How can we bound g(t )?
10 A frst example T B l n 2 Random varables (Z x ) Z x = g, x for a vector g wth..d. N (0, 1) entres Defne gaussan mean wdth g(t ) = E g sup Z x How can we bound g(t )? Ths talk: four progressvely tghter ways to bound g(t ), then applcatons of technques to some TCS problems
11 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x
12 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one
13 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one E sup Z x = = 0 u 0 P(sup Z x > u)du P(sup Z x > u) } {{ } 1 u + T e u2 /2 du + u log T (set u = 2 log T ) P(sup Z x > u) }{{} T e u2 /2 (unon bound) du
14 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one E sup Z x = = 0 u 0 P(sup Z x > u)du P(sup Z x > u) } {{ } 1 u + T e u2 /2 du + u log T (set u = 2 log T ) P(sup Z x > u) }{{} T e u2 /2 (unon bound) du
15 Gaussan mean wdth bound 1: unon bound g(t ) = E sup Z x = E sup g, x Z x s a gaussan wth varance one E sup Z x = = 0 u 0 P(sup Z x > u)du P(sup Z x > u) } {{ } 1 u + T e u2 /2 du + u log T (set u = 2 log T ) P(sup Z x > u) }{{} T e u2 /2 (unon bound) du
16 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 )
17 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 ) g, x = g, x + g, x x (x = argmn y T x y 2 ) g(t ) g(s ε ) + E g sup g, x x }{{} ε g 2
18 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 ) g, x = g, x + g, x x (x = argmn y T x y 2 ) g(t ) g(s ε ) + E g sup g, x x }{{} ε g 2 log S ε + ε(e g g 2 2 )1/2 log 1/2 N (T, l 2, ε) +ε n }{{} smallest ε net sze
19 Gaussan mean wdth bound 2: ε-net g(t ) = E sup g, x Let S ε be ε-net of (T, l 2 ) g, x = g, x + g, x x (x = argmn y T x y 2 ) g(t ) g(s ε ) + E g sup g, x x }{{} ε g 2 log S ε + ε(e g g 2 2 )1/2 log 1/2 N (T, l 2, ε) }{{} smallest ε net sze +ε n Choose ε to optmze bound; can never be worse than last slde (whch amounts to choosng ε = 0)
20 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x
21 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx
22 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx g(t ) E sup g, π 0 x g } {{ } 0 + k=1 E g sup g, k x
23 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx g(t ) E sup g, π 0 x g + k=1 E g sup g, k x } {{ } 0 { k x : x T } N (T, l 2, 1 /2 k ) N (T, l 2, 1 /2 k 1 ) (N (T, l 2, 1 /2 k )) 2
24 Gaussan mean wdth bound 3: ε-net sequence S k s a ( 1 /2 k )-net of T, k 0 π k x s closest pont n S k to x T, k x = π k x π k 1 x wlog T < (else apply ths slde to ε-net of T for ε small) g, x = g, π 0 x + k=1 g, kx g(t ) E sup g, π 0 x g + k=1 E g sup g, k x } {{ } 0 { k x : x T } N (T, l 2, 1 /2 k ) N (T, l 2, 1 /2 k 1 ) (N (T, l 2, 1 /2 k )) 2 g(t ) k=1 (1 /2 k ) log 1/2 N (T, l 2, 1 /2 k ) 0 log 1/2 N (T, l 2, u)du (Dudley s theorem)
25 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble )
26 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble ) Exercse: show Dudley s theorem s equvalent to g(t ) nf {Tk } admssble k=1 2k/2 sup d l2 (x, T k ) (should pck T k to be the best ε = ε(k) net of sze 2 2k )
27 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble ) Exercse: show Dudley s theorem s equvalent to g(t ) nf {Tk } admssble k=1 2k/2 sup d l2 (x, T k ) (should pck T k to be the best ε = ε(k) net of sze 2 2k ) Fernque 76 : can pull the sup x outsde the sum g(t ) nf {Tk } sup k=1 2k/2 d l2 (x, T k ) def = γ 2 (T, l 2 )
28 Gaussan mean wdth bound 4: generc channg Agan, wlog T <. Defne T 0 T 1 T k = T T 0 = 1, T k 2 2k (call such a sequence admssble ) Exercse: show Dudley s theorem s equvalent to g(t ) nf {Tk } admssble k=1 2k/2 sup d l2 (x, T k ) (should pck T k to be the best ε = ε(k) net of sze 2 2k ) Fernque 76 : can pull the sup x outsde the sum g(t ) nf {Tk } sup k=1 2k/2 d l2 (x, T k ) def = γ 2 (T, l 2 ) equvalent upper bound proven by Fernque (who mnmzed some ntegral over all measures over T ), but reformulated n terms of admssble sequences by Talgarand
29 Gaussan mean wdth bound 4: generc channg Proof of Fernque s bound g(t ) E sup g, π 0 x g } {{ } 0 + E g sup g, k x }{{} Y k k=1 (from before) t, P(Y k > t2 k/2 k x 2 ) e t2 2 k /2 (gaussan decay)
30 Gaussan mean wdth bound 4: generc channg Proof of Fernque s bound g(t ) E sup g, π 0 x g } {{ } 0 + E g sup g, k x }{{} Y k k=1 (from before) t, P(Y k > t2 k/2 k x 2 ) e t2 2 k /2 (gaussan decay) P( x, k Y k > t2 k/2 k x 2 ) k (22k ) 2 e t2 2 k /2
31 Gaussan mean wdth bound 4: generc channg Proof of Fernque s bound g(t ) E sup g, π 0 x g } {{ } 0 + E g sup g, k x }{{} Y k k=1 (from before) t, P(Y k > t2 k/2 k x 2 ) e t2 2 k /2 (gaussan decay) P( x, k Y k > t2 k/2 k x 2 ) k (22k ) 2 e t2 2 k /2 E g sup Y k = k 0 P(sup Y k > u)du k
32 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) k = γ 2 (T, l 2 ) [t + (2 2k ) 2 e t2 2 k /2 dt] γ 2 (T, l 2 ) t k=1 k Concluson: g(t ) γ 2 (T, l 2 ) Talagrand: g(t ) γ 2 (T, l 2 ) (won t show today) ( Majorzng measures theorem )
33 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) k = γ 2 (T, l 2 ) [t + (2 2k ) 2 e t2 2 k /2 dt] γ 2 (T, l 2 ) t k=1 k
34 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) γ 2 (T, l 2 ) [2 + γ 2 (T, l 2 ) 2 k ( ) (2 2k ) 2 e t2 2 k /2 dt] k=1 k
35 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) γ 2 (T, l 2 ) [2 + γ 2 (T, l 2 ) 2 k ( ) (2 2k ) 2 e t2 2 k /2 dt] k=1 k Concluson: g(t ) γ 2 (T, l 2 )
36 Gaussan mean wdth bound 4: generc channg E g sup Y k = k = γ 2 (T, l 2 ) 0 0 P(sup P(sup Y k > u)du k Y k > t sup 2 k/2 k x 2 )dt k (change of varables: u = t sup 2 k/2 k x 2 tγ 2 (T, l 2 )) γ 2 (T, l 2 ) [2 + γ 2 (T, l 2 ) 2 k ( ) (2 2k ) 2 e t2 2 k /2 dt] k=1 k Concluson: g(t ) γ 2 (T, l 2 ) Talagrand: g(t ) γ 2 (T, l 2 ) (won t show today) ( Majorzng measures theorem )
37 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du
38 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1
39 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1 sup x Bl n 1 g, x = g, so g(t ) log n Exercse: Come up wth admssble {T k } yeldng γ 2 log n (must exst by majorzng measures)
40 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1 sup x Bl n 1 g, x = g, so g(t ) log n Exercse: Come up wth admssble {T k } yeldng γ 2 log n (must exst by majorzng measures) Dudley: log N (B l n 1, l 2, u) ( 1 /u 2 ) log n for u not too small (consder just coverng ( 1 /u 2 )-sparse vectors wth u 2 n each coordnate). Dudley can only gve g(b l n 1 ) log 3/2 n.
41 Are these bounds really dfferent? γ 2 (T, l 2 ): nf {Tk } sup k=1 2k/2 d l2 (x, T k ) Dudley: nf {Tk } k=1 2k/2 sup d l2 (x, T k ) 0 log 1/2 N (T, l 2, u)du Dudley not optmal: T = B l n 1 sup x Bl n 1 g, x = g, so g(t ) log n Exercse: Come up wth admssble {T k } yeldng γ 2 log n (must exst by majorzng measures) Dudley: log N (B l n 1, l 2, u) ( 1 /u 2 ) log n for u not too small (consder just coverng ( 1 /u 2 )-sparse vectors wth u 2 n each coordnate). Dudley can only gve g(b l n 1 ) log 3/2 n. Smple vanlla ε-net argument gves g(b l n 1 ) poly(n).
42 Hgh probablty So far just talked about g(t ) = E g sup Z x But what f we want to know sup Z x s small whp, not just n expectaton?
43 Hgh probablty So far just talked about g(t ) = E g sup Z x But what f we want to know sup Z x s small whp, not just n expectaton? Usual approach: bound E g sup Zx p for large p and do Markov ( moment method ) Can bound moments usng channg too; see (Drksen 13)
44 Applcatons n computer scence Fast RIP matrces (Candès, Tao 06), (Rudelson, Vershynn 06), (Cheragch, Guruswam, Velngker 13), (N., Prce, Wootters 14), (Bourgan 14), (Havv, Regev 15) Fast JL (Alon, Lberty 11), (Krahmer, Ward 11), (Bourgan, Drksen, N. 15), (Oymak, Recht, Soltanolkotab 15) Instance-wse JL bounds (Gordon 88), (Klartag, Mendelson 05), (Mendelson, Pajor, Tomczak-Jaegermann 07), (Drksen 14) Approxmate nearest neghbor (Indyk, Naor 07) Determnstc algorthm to estmate graph cover tme (Dng, Lee, Peres 11) Lst-decodablty of random codes (Wootters 13), (Rudra, Wootters 14)...
45 A channg result for quadratc forms Theorem [Krahmer, Mendelson, Rauhut 14] Let A R n n be a famly of matrces, and let σ 1,..., σ n be ndependent subgaussans. Then E sup Aσ 2 2 E Aσ 2 A A σ 2 γ 2 2(A, l2 l 2 ) + γ 2 (A, l2 l 2 ) F (A) + l2 l 2 (A) F (A) ( X s dameter under X -norm)
46 A channg result for quadratc forms Theorem [Krahmer, Mendelson, Rauhut 14] Let A R n n be a famly of matrces, and let σ 1,..., σ n be ndependent subgaussans. Then E sup Aσ 2 2 E Aσ 2 A A σ 2 γ 2 2(A, l2 l 2 ) + γ 2 (A, l2 l 2 ) F (A) + l2 l 2 (A) F (A) ( X s dameter under X -norm) Won t show proof today, but t s smlar to boundng g(t ) (wth some extra trcks). See edu/ mnlek/madalgo2015/, Lecture 3.
47 Instance-wse bounds for JL Corollary (Gordon 88, Klartag-Mendelson 05, Mendelson, Pajor, Tomczak-Jaegermann 07, Drksen 14) For T S n 1 and 0 < ε < 1/2, let Π R m n have ndependent subgaussan ndependent entres wth mean zero and varance 1/m for m (g2 (T )+1)/ε 2. Then E sup Πx < ε Π
48 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n
49 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m).
50 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t )
51 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t ) F (A T ) = 1, l2 l 2 (A T ) = 1/ m
52 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t ) F (A T ) = 1, l2 l 2 (A T ) = 1/ m Thus E Π sup Πx g2 (T )/m + g(t ) / m + 1 / m
53 Proof of Gordon s theorem Instance-wse bounds for JL For x T let A x denote the m mn matrx: x 1 x n 0 0 A x = x 1 x n 0 0 m x 1 x n Then Πx 2 2 = A xσ 2 2, where σ s formed by concatenatng rows of Π (multpled by m). A x A y = A x y = ( 1 / m) x y 2 γ 2 (A T, l2 l 2 ) = γ 2 (T, l 2 ) g(t ) F (A T ) = 1, l2 l 2 (A T ) = 1/ m Thus E Π sup Πx g2 (T )/m + g(t ) / m + 1 / m Set m (g2 (T )+1)/ε 2
54 Consequences of Gordon s theorem m (g2 (T )+1)/ε 2 T < : g 2 (T ) log T (JL) T a d-dm subspace: g 2 (T ) d (subspace embeddngs) T all k-sparse vectors: g 2 (T ) k log(n/k) (RIP)
55 Consequences of Gordon s theorem m (g2 (T )+1)/ε 2 T < : g 2 (T ) log T (JL) T a d-dm subspace: g 2 (T ) d (subspace embeddngs) T all k-sparse vectors: g 2 (T ) k log(n/k) (RIP) more applcatons to constraned least squares, manfold learnng, model-based compressed sensng,... (see (Drksen 14) and (Bourgan, Drksen, N. 15))
56 Channg sn t just for gaussans
57 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) Restrcted sometry property useful n compressed sensng. T = {x : x 0 k, x 2 = 1}. Theorem (Candès-Tao 06, Donoho 06, Candés 08) If Π satsfes (ε, k)-rip for ε < 2 1 then there s a lnear program whch, gven Πx and Π as nput, recovers x n polynomal tme such that x x 2 O( 1 / k) mn y 0 k x y 1.
58 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) Restrcted sometry property useful n compressed sensng. T = {x : x 0 k, x 2 = 1}. Theorem (Candès-Tao 06, Donoho 06, Candés 08) If Π satsfes (ε, k)-rip for ε < 2 1 then there s a lnear program whch, gven Πx and Π as nput, recovers x n polynomal tme such that x x 2 O( 1 / k) mn y 0 k x y 1. Of nterest to show samplng rows of dscrete Fourer matrx s RIP
59 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) (Unnormalzed) Fourer matrx F, rows: z 1,..., z n δ 1,..., δ n ndependent Bernoull wth expectaton m/n
60 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) (Unnormalzed) Fourer matrx F, rows: z 1,..., z n δ 1,..., δ n ndependent Bernoull wth expectaton m/n Want E δ sup T [n] T k I T 1 m n δ z (T ) =1 (T ) z < ε
61 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ 1 m E = sup T [n] T k sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1
62 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1
63 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g sup δ,g T sup x B n,k 2 T (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1
64 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1
65 Channg wthout gaussans: RIP (Rudelson, Vershynn 06) LHS = E δ sup T [n] T k 1 m E = sup δ,δ T π 2 1 m I T {}}{ 1 n E δ δ m z (T ) (T ) z 1 m 2π 1 m E 1 m E δ E g =1 n (δ δ )z (T ) =1 E sup E δ,δ,σ g T sup δ,g T sup x B n,k 2 (T ) z n δ z (T ) =1 (Jensen) n g σ (δ δ )z (T ) =1 n g δ z (T ) =1 (T ) z (T ) z (T ) z (Jensen+trangle neq) n g δ z, x 2 (gaussan mean wdth!) =1
66 The End
67 June 22 nd +23 rd : workshop on concentraton of measure / channg at Harvard, after STOC 16. Detals+webste forthcomng.
Dimensionality Reduction Notes 1
Dmensonalty Reducton Notes 1 Jelan Nelson mnlek@seas.harvard.edu August 10, 2015 1 Prelmnares Here we collect some notaton and basc lemmas used throughout ths note. Throughout, for a random varable X,
More informationHANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION
HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION MARK RUDELSON AND ROMAN VERSHYNIN Abstract. In ths expostory note, we gve a modern proof of Hanson-Wrght nequalty for quadratc forms n sub-gaussan
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationLecture 9 Sept 29, 2017
Sketchng Algorthms for Bg Data Fall 2017 Prof. Jelan Nelson Lecture 9 Sept 29, 2017 Scrbe: Mtal Bafna 1 Fast JL transform Typcally we have some hgh-mensonal computatonal geometry problem, an we use JL
More informationErrors for Linear Systems
Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationCIS 700: algorithms for Big Data
CIS 700: algorthms for Bg Data Lecture 5: Dmenson Reducton Sldes at htt://grgory.us/bg-data-class.html Grgory Yaroslavtsev htt://grgory.us Today Dmensonalty reducton AMS as dmensonalty reducton Johnson-Lndenstrauss
More informationMatrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD
Matrx Approxmaton va Samplng, Subspace Embeddng Lecturer: Anup Rao Scrbe: Rashth Sharma, Peng Zhang 0/01/016 1 Solvng Lnear Systems Usng SVD Two applcatons of SVD have been covered so far. Today we loo
More informationprinceton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora
prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable
More informationP exp(tx) = 1 + t 2k M 2k. k N
1. Subgaussan tals Defnton. Say that a random varable X has a subgaussan dstrbuton wth scale factor σ< f P exp(tx) exp(σ 2 t 2 /2) for all real t. For example, f X s dstrbuted N(,σ 2 ) then t s subgaussan.
More information18.1 Introduction and Recap
CS787: Advanced Algorthms Scrbe: Pryananda Shenoy and Shjn Kong Lecturer: Shuch Chawla Topc: Streamng Algorthmscontnued) Date: 0/26/2007 We contnue talng about streamng algorthms n ths lecture, ncludng
More informationDimensionality Reduction Notes 2
Dmensonalty Reducton Notes 2 Jelan Nelson mnlek@seas.harvard.edu August 11, 2015 1 Optmalty theorems for JL Yesterday we saw for MJL that we could acheve target dmenson m = O(ε 2 log N), and for DJL we
More informationMining Data Streams-Estimating Frequency Moment
Mnng Data Streams-Estmatng Frequency Moment Barna Saha October 26, 2017 Frequency Moment Computng moments nvolves dstrbuton of frequences of dfferent elements n the stream. Frequency Moment Computng moments
More informationHANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION
HANSON-WRIGHT INEQUALITY AND SUB-GAUSSIAN CONCENTRATION MARK RUDELSON AND ROMAN VERSHYNIN Abstract. In ths expostory note, we gve a modern proof of Hanson-Wrght nequalty for quadratc forms n sub-gaussan
More informationFinding Dense Subgraphs in G(n, 1/2)
Fndng Dense Subgraphs n Gn, 1/ Atsh Das Sarma 1, Amt Deshpande, and Rav Kannan 1 Georga Insttute of Technology,atsh@cc.gatech.edu Mcrosoft Research-Bangalore,amtdesh,annan@mcrosoft.com Abstract. Fndng
More informationLecture 4: September 12
36-755: Advanced Statstcal Theory Fall 016 Lecture 4: September 1 Lecturer: Alessandro Rnaldo Scrbe: Xao Hu Ta Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer: These notes have not been
More informationLecture 3. Ax x i a i. i i
18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest
More informationLecture 10: May 6, 2013
TTIC/CMSC 31150 Mathematcal Toolkt Sprng 013 Madhur Tulsan Lecture 10: May 6, 013 Scrbe: Wenje Luo In today s lecture, we manly talked about random walk on graphs and ntroduce the concept of graph expander,
More informationLecture 4: Universal Hash Functions/Streaming Cont d
CSE 5: Desgn and Analyss of Algorthms I Sprng 06 Lecture 4: Unversal Hash Functons/Streamng Cont d Lecturer: Shayan Oves Gharan Aprl 6th Scrbe: Jacob Schreber Dsclamer: These notes have not been subjected
More informationDimensionality Reduction Notes 3
Dimensionality Reduction Notes 3 Jelani Nelson minilek@seas.harvard.edu August 13, 2015 1 Gordon s theorem Let T be a finite subset of some normed vector space with norm X. We say that a sequence T 0 T
More informationSolutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.
Solutons HW #2 Dual of general LP. Fnd the dual functon of the LP mnmze subject to c T x Gx h Ax = b. Gve the dual problem, and make the mplct equalty constrants explct. Soluton. 1. The Lagrangan s L(x,
More informationLecture 3 January 31, 2017
CS 224: Advanced Algorthms Sprng 207 Prof. Jelan Nelson Lecture 3 January 3, 207 Scrbe: Saketh Rama Overvew In the last lecture we covered Y-fast tres and Fuson Trees. In ths lecture we start our dscusson
More informationn α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0
MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector
More informationLecture 4: Constant Time SVD Approximation
Spectral Algorthms and Representatons eb. 17, Mar. 3 and 8, 005 Lecture 4: Constant Tme SVD Approxmaton Lecturer: Santosh Vempala Scrbe: Jangzhuo Chen Ths topc conssts of three lectures 0/17, 03/03, 03/08),
More informationHanson-Wright inequality and sub-gaussian concentration
Electron. Commun. Probab. 18 (013, no. 8, 1 9. DOI: 10.114/ECP.v18-865 ISSN: 1083-589X ELECTRONIC COMMUNICATIONS n PROBABILITY Hanson-Wrght nequalty and sub-gaussan concentraton Mark Rudelson Roman Vershynn
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 6 Luca Trevisan September 12, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 6 Luca Trevsan September, 07 Scrbed by Theo McKenze Lecture 6 In whch we study the spectrum of random graphs. Overvew When attemptng to fnd n polynomal
More informationSupport Vector Machines
Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationDifferentiating Gaussian Processes
Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the
More informationCOS 521: Advanced Algorithms Game Theory and Linear Programming
COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton
More informationFeature Selection: Part 1
CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationEigenvalues of Random Graphs
Spectral Graph Theory Lecture 2 Egenvalues of Random Graphs Danel A. Spelman November 4, 202 2. Introducton In ths lecture, we consder a random graph on n vertces n whch each edge s chosen to be n the
More informationLecture 5 September 17, 2015
CS 229r: Algorthms for Bg Data Fall 205 Prof. Jelan Nelson Lecture 5 September 7, 205 Scrbe: Yakr Reshef Recap and overvew Last tme we dscussed the problem of norm estmaton for p-norms wth p > 2. We had
More information10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)
0-80: Advanced Optmzaton and Randomzed Methods Lecture : Convex functons (Jan 5, 04) Lecturer: Suvrt Sra Addr: Carnege Mellon Unversty, Sprng 04 Scrbes: Avnava Dubey, Ahmed Hefny Dsclamer: These notes
More informationStanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011
Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationProperties of Least Squares
Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures
More informationSupport Vector Machines
/14/018 Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x
More informationNotes on Frequency Estimation in Data Streams
Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to
More informationIntroduction to Algorithms
Introducton to Algorthms 6.046J/8.40J Lecture 7 Prof. Potr Indyk Data Structures Role of data structures: Encapsulate data Support certan operatons (e.g., INSERT, DELETE, SEARCH) Our focus: effcency of
More information1 The Mistake Bound Model
5-850: Advanced Algorthms CMU, Sprng 07 Lecture #: Onlne Learnng and Multplcatve Weghts February 7, 07 Lecturer: Anupam Gupta Scrbe: Bryan Lee,Albert Gu, Eugene Cho he Mstake Bound Model Suppose there
More informationTAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES
TAIL BOUNDS FOR SUMS OF GEOMETRIC AND EXPONENTIAL VARIABLES SVANTE JANSON Abstract. We gve explct bounds for the tal probabltes for sums of ndependent geometrc or exponental varables, possbly wth dfferent
More informationLecture 21: Numerical methods for pricing American type derivatives
Lecture 21: Numercal methods for prcng Amercan type dervatves Xaoguang Wang STAT 598W Aprl 10th, 2014 (STAT 598W) Lecture 21 1 / 26 Outlne 1 Fnte Dfference Method Explct Method Penalty Method (STAT 598W)
More informationCollege of Computer & Information Science Fall 2009 Northeastern University 20 October 2009
College of Computer & Informaton Scence Fall 2009 Northeastern Unversty 20 October 2009 CS7880: Algorthmc Power Tools Scrbe: Jan Wen and Laura Poplawsk Lecture Outlne: Prmal-dual schema Network Desgn:
More informationMaximum Likelihood Estimation (MLE)
Maxmum Lkelhood Estmaton (MLE) Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 01 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y
More informationLecture 3: Dual problems and Kernels
Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationCOS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture # 15 Scribe: Jieming Mao April 1, 2013
COS 511: heoretcal Machne Learnng Lecturer: Rob Schapre Lecture # 15 Scrbe: Jemng Mao Aprl 1, 013 1 Bref revew 1.1 Learnng wth expert advce Last tme, we started to talk about learnng wth expert advce.
More informationLecture 20: November 7
0-725/36-725: Convex Optmzaton Fall 205 Lecturer: Ryan Tbshran Lecture 20: November 7 Scrbes: Varsha Chnnaobreddy, Joon Sk Km, Lngyao Zhang Note: LaTeX template courtesy of UC Berkeley EECS dept. Dsclamer:
More informationLinear Classification, SVMs and Nearest Neighbors
1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush
More informationDurban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications
Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department
More informationVapnik-Chervonenkis theory
Vapnk-Chervonenks theory Rs Kondor June 13, 2008 For the purposes of ths lecture, we restrct ourselves to the bnary supervsed batch learnng settng. We assume that we have an nput space X, and an unknown
More informationProbability Theory (revisited)
Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationThe Expectation-Maximization Algorithm
The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.
More informationLecture 10 Support Vector Machines. Oct
Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron
More informationRandomness and Computation
Randomness and Computaton or, Randomzed Algorthms Mary Cryan School of Informatcs Unversty of Ednburgh RC 208/9) Lecture 0 slde Balls n Bns m balls, n bns, and balls thrown unformly at random nto bns usually
More informationThe Second Anti-Mathima on Game Theory
The Second Ant-Mathma on Game Theory Ath. Kehagas December 1 2006 1 Introducton In ths note we wll examne the noton of game equlbrum for three types of games 1. 2-player 2-acton zero-sum games 2. 2-player
More informationxp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ
CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and
More informationNP-Completeness : Proofs
NP-Completeness : Proofs Proof Methods A method to show a decson problem Π NP-complete s as follows. (1) Show Π NP. (2) Choose an NP-complete problem Π. (3) Show Π Π. A method to show an optmzaton problem
More informationLagrange Multipliers Kernel Trick
Lagrange Multplers Kernel Trck Ncholas Ruozz Unversty of Texas at Dallas Based roughly on the sldes of Davd Sontag General Optmzaton A mathematcal detour, we ll come back to SVMs soon! subject to: f x
More informationρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to
THE INVERSE POWER METHOD (or INVERSE ITERATION) -- applcaton of the Power method to A some fxed constant ρ (whch s called a shft), x λ ρ If the egenpars of A are { ( λ, x ) } ( ), or (more usually) to,
More informationProblem Set 9 Solutions
Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem
More informationLinear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.
Lnear, affne, and convex sets and hulls In the sequel, unless otherwse specfed, X wll denote a real vector space. Lnes and segments. Gven two ponts x, y X, we defne xy = {x + t(y x) : t R} = {(1 t)x +
More information8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS
SECTION 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS 493 8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS All the vector spaces you have studed thus far n the text are real vector spaces because the scalars
More informationThe Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD
he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s
More informationMeshless Surfaces. presented by Niloy J. Mitra. An Nguyen
Meshless Surfaces presented by Nloy J. Mtra An Nguyen Outlne Mesh-Independent Surface Interpolaton D. Levn Outlne Mesh-Independent Surface Interpolaton D. Levn Pont Set Surfaces M. Alexa, J. Behr, D. Cohen-Or,
More informationGoogle PageRank with Stochastic Matrix
Google PageRank wth Stochastc Matrx Md. Sharq, Puranjt Sanyal, Samk Mtra (M.Sc. Applcatons of Mathematcs) Dscrete Tme Markov Chan Let S be a countable set (usually S s a subset of Z or Z d or R or R d
More informationMLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012
MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:
More informationThe Geometry of Logit and Probit
The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.
More informationCommunication Complexity 16:198: February Lecture 4. x ij y ij
Communcaton Complexty 16:198:671 09 February 2010 Lecture 4 Lecturer: Troy Lee Scrbe: Rajat Mttal 1 Homework problem : Trbes We wll solve the thrd queston n the homework. The goal s to show that the nondetermnstc
More information6.842 Randomness and Computation February 18, Lecture 4
6.842 Randomness and Computaton February 18, 2014 Lecture 4 Lecturer: Rontt Rubnfeld Scrbe: Amartya Shankha Bswas Topcs 2-Pont Samplng Interactve Proofs Publc cons vs Prvate cons 1 Two Pont Samplng 1.1
More informationCS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements
CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before
More informationLecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.
prnceton u. sp 02 cos 598B: algorthms and complexty Lecture 20: Lft and Project, SDP Dualty Lecturer: Sanjeev Arora Scrbe:Yury Makarychev Today we wll study the Lft and Project method. Then we wll prove
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More informationp 1 c 2 + p 2 c 2 + p 3 c p m c 2
Where to put a faclty? Gven locatons p 1,..., p m n R n of m houses, want to choose a locaton c n R n for the fre staton. Want c to be as close as possble to all the house. We know how to measure dstance
More informationStanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7
Stanford Unversty CS54: Computatonal Complexty Notes 7 Luca Trevsan January 9, 014 Notes for Lecture 7 1 Approxmate Countng wt an N oracle We complete te proof of te followng result: Teorem 1 For every
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13
CME 30: NUMERICAL LINEAR ALGEBRA FALL 005/06 LECTURE 13 GENE H GOLUB 1 Iteratve Methods Very large problems (naturally sparse, from applcatons): teratve methods Structured matrces (even sometmes dense,
More informationGenerative classification models
CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn
More informationError Bars in both X and Y
Error Bars n both X and Y Wrong ways to ft a lne : 1. y(x) a x +b (σ x 0). x(y) c y + d (σ y 0) 3. splt dfference between 1 and. Example: Prmordal He abundance: Extrapolate ft lne to [ O / H ] 0. [ He
More informationDepartment of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING
MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More informationTHE WEIGHTED WEAK TYPE INEQUALITY FOR THE STRONG MAXIMAL FUNCTION
THE WEIGHTED WEAK TYPE INEQUALITY FO THE STONG MAXIMAL FUNCTION THEMIS MITSIS Abstract. We prove the natural Fefferman-Sten weak type nequalty for the strong maxmal functon n the plane, under the assumpton
More informationLecture 6: Introduction to Linear Regression
Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6
More informationQuantifying Uncertainty
Partcle Flters Quantfyng Uncertanty Sa Ravela M. I. T Last Updated: Sprng 2013 1 Quantfyng Uncertanty Partcle Flters Partcle Flters Appled to Sequental flterng problems Can also be appled to smoothng problems
More informationWeek 2. This week, we covered operations on sets and cardinality.
Week 2 Ths week, we covered operatons on sets and cardnalty. Defnton 0.1 (Correspondence). A correspondence between two sets A and B s a set S contaned n A B = {(a, b) a A, b B}. A correspondence from
More information2.3 Nilpotent endomorphisms
s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms
More informationAPPENDIX A Some Linear Algebra
APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,
More informationCalculation of time complexity (3%)
Problem 1. (30%) Calculaton of tme complexty (3%) Gven n ctes, usng exhaust search to see every result takes O(n!). Calculaton of tme needed to solve the problem (2%) 40 ctes:40! dfferent tours 40 add
More informationADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING
1 ADVANCED ACHINE LEARNING ADVANCED ACHINE LEARNING Non-lnear regresson technques 2 ADVANCED ACHINE LEARNING Regresson: Prncple N ap N-dm. nput x to a contnuous output y. Learn a functon of the type: N
More information20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The first idea is connectedness.
20. Mon, Oct. 13 What we have done so far corresponds roughly to Chapters 2 & 3 of Lee. Now we turn to Chapter 4. The frst dea s connectedness. Essentally, we want to say that a space cannot be decomposed
More informationU.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016
U.C. Berkeley CS94: Spectral Methods and Expanders Handout 8 Luca Trevsan February 7, 06 Lecture 8: Spectral Algorthms Wrap-up In whch we talk about even more generalzatons of Cheeger s nequaltes, and
More informationAPPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14
APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce
More informationLecture 3: Probability Distributions
Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the
More information15-381: Artificial Intelligence. Regression and cross validation
15-381: Artfcal Intellgence Regresson and cross valdaton Where e are Inputs Densty Estmator Probablty Inputs Classfer Predct category Inputs Regressor Predct real no. Today Lnear regresson Gven an nput
More informationThe conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above
The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ
More information