A Matrix Expander Chernoff Bound
|
|
- Lydia Morrison
- 5 years ago
- Views:
Transcription
1 A Matrix Expander Chernoff Bound Ankit Garg, Yin Tat Lee, Zhao Song, Nikhil Srivastava UC Berkeley
2 Vanilla Chernoff Bound Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4)
3 Vanilla Chernoff Bound Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) Two Extensions: 1.Dependent Random Variables 2.Sums of random matrices
4 Expander Chernoff Bound [AKS 87, G 94] Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0.
5 Expander Chernoff Bound [AKS 87, G 94] Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 ) Implies walk of length k 1 λ 1 concentrates around mean.
6 Intuition for Dependence on Spectral Gap 1 λ = 1 n 2 f = +1 f = 1
7 Intuition for Dependence on Spectral Gap 1 λ = 1 n 2 Typical random walk takes Ω n 2 steps to see both ±1.
8 Derandomization Motivation Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 ) To generate k indep samples from [n] need k log n bits. If G is 3 regular with constant λ, walk needs: log n + k log(3) bits.
9 Derandomization Motivation Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 ) To generate k indep samples from [n] need k log n bits. If G is 3 regular with constant λ, walk needs: log n + k log(3) bits. When k = O(log n) reduces randomness quadratically. Can completely derandomize in polynomial time.
10 Matrix Chernoff Bound Thm [Rudelson 97, Ahlswede-Winter 02, Oliveira 08, Tropp 11 ]. If X 1,, X k are independent mean zero random d d Hermitian matrices with X i 1 then P 1 k i X i ε 2dexp( kε 2 /4)
11 Matrix Chernoff Bound Thm [Rudelson 97, Ahlswede-Winter 02, Oliveira 08, Tropp 11 ]. If X 1,, X k are independent mean zero random d d Hermitian matrices with X i 1 then P 1 k i X i ε 2dexp( kε 2 /4) Factor d is tight because of the diagonal case. Very generic bound (no independence assumptions on the entries). Many applications + martingale extensions (see Tropp).
12 Matrix Expander Chernoff Bound? Conj[Wigderson-Xiao 05]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V C d d be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2dexp( c(1 λ)kε 2 ) Motivated by derandomized Alon-Roichman theorem.
13 Main Theorem Thm. Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V C d d be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2dexp( c(1 λ)kε 2 ) Gives black-box derandomization of any application of matrix Chernoff
14 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp t X i = e tkε Ee tx i i i i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4
15 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Markov
16 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Indep.
17 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Bounded
18 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Mean zero
19 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4
20 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4
21 2. Proof of Expander Chernoff Goal: Show E exp t σ i f v i exp c λ kt 2
22 2. Proof of Expander Chernoff Goal: Show E exp t σ i f v i exp c λ kt 2 Issue: Eexp σ i f v i ς i Eexp(tf(v i )) How to control the mgf without independence?
23 Step 1: Write mgf as quadratic form Ee t σ i k f(v i) = i 0,,i k V P(v 0 = i 0 )P i 0, i 1 P(i k 1, i k ) exp t 1 j k f i j = 1 n i 0,,i k V P i 0, i 1 e tf(i 1) P i k 1, i k e tf(i k) = u, EP k u where u = 1 n,, 1 E = e tf(1) n. i 0 e tf(i 1) i 1 e tf(i 3) i 3 P i 0, i 1 P i 1, i 2 i 2 e tf(n) P i 2, i 3 e tf(i 2)
24 Step 1: Write mgf as quadratic form Ee t σ i k f(v i) = i 0,,i k V P(v 0 = i 0 )P i 0, i 1 P(i k 1, i k ) exp t 1 j k f i j = 1 n i 0,,i k V P i 0, i 1 e tf(i 1) P i k 1, i k e tf(i k) = u, EP k u where u = 1 n,, 1 E = e tf(1) n. i 0 e tf(i 1) i 1 e tf(i 3) i 3 P i 0, i 1 P i 1, i 2 i 2 e tf(n) P i 2, i 3 e tf(i 2)
25 Step 1: Write mgf as quadratic form Ee t σ i k f(v i) = i 0,,i k V P(v 0 = i 0 )P i 0, i 1 P(i k 1, i k ) exp t 1 j k f i j = 1 n i 0,,i k V P i 0, i 1 e tf(i 1) P i k 1, i k e tf(i k) = u, EP k u where u = 1 n,, 1 E = e tf(1) n. i 0 e tf(i 1) i 1 e tf(i 3) i 3 P i 0, i 1 P i 1, i 2 i 2 e tf(n) P i 2, i 3 e tf(i 2)
26 Step 2: Bound quadratic form Goal: Show u, EP k u exp c λ kt 2 Observe: P J λ where J = complete graph with self loops So for small λ, should have u, EP k u u, (EJ) k u
27 Step 2: Bound quadratic form Goal: Show u, EP k u exp c λ kt 2 Observe: P J λ where J = complete graph with self loops So for small λ, should have u, EP k u u, (EJ) k u Approach 1. Use perturbation theory to show EP exp c λ t 2
28 Step 2: Bound quadratic form Goal: Show u, EP k u exp c λ kt 2 Observe: P J λ where J = complete graph with self loops So for small λ, should have u, EP k u u, (EJ) k u Approach 1. Use perturbation theory to show EP exp c λ t 2 Approach 2. (Healy 08) track projection of iterates along u.
29 Simplest case: λ = 0 EPEPEPEPu
30 Simplest case: λ = 0 EPEPEPEPu
31 Simplest case: λ = 0 EPEPEPEPu
32 Simplest case: λ = 0 EPEPEPEPu
33 Simplest case: λ = 0 EPEPEPEPu
34 Observations EPEPEPEPu Observe: P shrinks every vector orthogonal to u by λ. condition. u, Eu = 1 n σ v V e tf(v) = 1 + t 2 by mean zero
35 A small dynamical system m = Q v v = EP j u m = v, u m m
36 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m 0 λ = 0 0 v EPv
37 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m e t λ any λ λt v EPv
38 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m 1 t log 1/λ λt v EPv
39 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m 1 λt Any mass that leaves gets shrunk by λ
40 Analyzing dynamical system gives Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 )
41 Generalization to Matrices? Setup: f: V C d d, random walk v 1,, v k. Goal: ETr exp t i f v i d exp ckt 2 where e A is defined as a power series. Main Issue: exp A + B exp A exp(b) unless A, B = 0 can t express exp(sum) as iterated product.
42 Generalization to Matrices? Setup: f: V C d d, random walk v 1,, v k. Goal: ETr exp t i f v i d exp ckt 2 where e A is defined as a power series. Main Issue: exp A + B exp A exp(b) unless A, B = 0 can t express exp(sum) as iterated product.
43 The Golden-Thompson Inequality Partial Workaround [Golden-Thompson 65]: Tr exp A + B Tr exp A exp B Sufficient for independent case by induction. For expander case, need this for k matrices. False! e A Tr e A+B+C > 0 > Tr(e A e B e C ) e B e C
44 The Golden-Thompson Inequality Partial Workaround [Golden-Thompson 65]: Tr exp A + B Tr exp A exp B Sufficient for independent case by induction. For expander case, need this for k matrices. e A Tr e A+B+C > 0 > Tr(e A e B e C ) e B e C
45 The Golden-Thompson Inequality Partial Workaround [Golden-Thompson 65]: Tr exp A + B Tr exp A exp B Sufficient for independent case by induction. For expander case, need this for k matrices. False! e A Tr e A+B+C > 0 > Tr(e A e B e C ) e B e C
46 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R. 1. Matrix on RHS is always PSD. 2. Average-case inequality: e A i/2 are conjugated by unitaries. 3. Implies Lieb s concavity, triple-matrix, ALT, and more.
47 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R. 1. Matrix on RHS is always PSD. 2. Average-case inequality: e A i/2 are conjugated by unitaries. 3. Implies Lieb s concavity, triple-matrix, ALT, and more.
48 Proof of SBT: Lie-Trotter Formula e A+B+C = lim θ 0 + eθa e θb θc 1/θ e log Tr e A+B+C = lim θ 0 + log G θ 2/ θ For G z e za 2 e zb 2 e zc 2
49 Proof of SBT: Lie-Trotter Formula e A+B+C = lim θ 0 + eθa e θb θc 1/θ e log Tr e A+B+C = lim θ 0 + 2log G θ 2/ θ /θ For G z e za 2 e zb 2 e zc 2
50 Complex Interpolation (Stein-Hirschman) θ log G θ 2/θ
51 Complex Interpolation (Stein-Hirschman) For each θ, find analytic F(z) st: F it = 1 F 1 + it G 1 + it 2 F θ = G θ 2/θ
52 Complex Interpolation (Stein-Hirschman) For each θ, find analytic F(z) st: F it = 1 F 1 + it G 1 + it 2 F θ = G θ 2/θ log F θ නlog F it + නlog F 1 + it
53 Complex Interpolation (Stein-Hirschman) For each θ, find analytic F(z) st: F it = 1 F 1 + it G 1 + it 2 F θ = G θ 2/θ lim θ 0 log F θ නlog F it + නlog F 1 + it
54 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R.
55 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R. Issue. SBT involves integration over unbounded region, bad for Taylor expansion.
56 Bounded Modification of SBT Solution. Prove bounded version of SBT by replacing strip with half-disk. [Thm] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1e ib 2 e A ke ib 2 e A 1e ib 2 e A ke ib 2 θ where β(b) is an explicit probability density on π 2, π 2. Proof. Analytic F(z) + Poisson Kernel + Riemann map.
57 Handling Two-sided Products Issue. Two-sided rather than one-sided products: Tr e tf v 1 e ib 2 e tf v k e ib 2 e tf v 1 e ib 2 e tf v k e ib 2 Solution. Encode as one-sided product by using Tr AXB = A B T vec X : e tf v 1 e ib 2 e tf v 1 e ib 2 e tf v k T e ib 2 e tf v k T e ib 2 vec I d, vec I d
58 Handling Two-sided Products Issue. Two-sided rather than one-sided products: Tr e tf v 1 e ib 2 e tf v k e ib 2 e tf v 1 e ib 2 e tf v k e ib 2 Solution. Encode as one-sided product by using Tr AXB = A B T vec X : e tf v 1 e ib 2 e tf v 1 T e ib 2 e tf v k T e ib 2 e tf v k T e ib 2 vec I d, vec I d
59 Finishing the Proof Carry out a version of Healy s argument with P I d 2 and: E = tf 1 eib e 2 e tf 1 T e ib 2 e tf(n)eib 2 e tf n T e ib 2 i 0 And vec(i d ) u instead of u. This leads to the additional d factor. i 1 P i 1, i 2 P i 0, i 1 i 2 i 3 P i 2, i 3
60 Main Theorem Thm. Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V C d d be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2dexp( c(1 λ)kε 2 )
61 Open Questions Other matrix concentration inequalities (multiplicative, low-rank, moments) Other Banach spaces (Schatten norms) More applications of complex interpolation
Matrix Concentration. Nick Harvey University of British Columbia
Matrix Concentration Nick Harvey University of British Columbia The Problem Given any random nxn, symmetric matrices Y 1,,Y k. Show that i Y i is probably close to E[ i Y i ]. Why? A matrix generalization
More informationStein s Method for Matrix Concentration
Stein s Method for Matrix Concentration Lester Mackey Collaborators: Michael I. Jordan, Richard Y. Chen, Brendan Farrell, and Joel A. Tropp University of California, Berkeley California Institute of Technology
More informationMultivariate trace inequalities. David Sutter, Mario Berta, Marco Tomamichel
Multivariate trace inequalities David Sutter, Mario Berta, Marco Tomamichel What are trace inequalities and why we should care. Main difference between classical and quantum world are complementarity and
More informationA NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER
A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER 1. The method Ashwelde and Winter [1] proposed a new approach to deviation inequalities for sums of independent random matrices. The
More informationMultivariate Trace Inequalities
Multivariate Trace Inequalities Mario Berta arxiv:1604.03023 with Sutter and Tomamichel (to appear in CMP) arxiv:1512.02615 with Fawzi and Tomamichel QMath13 - October 8, 2016 Mario Berta (Caltech) Multivariate
More informationarxiv: v1 [math.pr] 11 Feb 2019
A Short Note on Concentration Inequalities for Random Vectors with SubGaussian Norm arxiv:190.03736v1 math.pr] 11 Feb 019 Chi Jin University of California, Berkeley chijin@cs.berkeley.edu Rong Ge Duke
More informationGraph Sparsification I : Effective Resistance Sampling
Graph Sparsification I : Effective Resistance Sampling Nikhil Srivastava Microsoft Research India Simons Institute, August 26 2014 Graphs G G=(V,E,w) undirected V = n w: E R + Sparsification Approximate
More informationGraph Sparsification II: Rank one updates, Interlacing, and Barriers
Graph Sparsification II: Rank one updates, Interlacing, and Barriers Nikhil Srivastava Simons Institute August 26, 2014 Previous Lecture Definition. H = (V, F, u) is a κ approximation of G = V, E, w if:
More informationU.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018
U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 Lecture 3 In which we show how to find a planted clique in a random graph. 1 Finding a Planted Clique We will analyze
More informationUsing Friendly Tail Bounds for Sums of Random Matrices
Using Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Computing + Mathematical Sciences California Institute of Technology jtropp@cms.caltech.edu Research supported in part by NSF, DARPA,
More informationMatrix concentration inequalities
ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random
More informationGraph Sparsifiers: A Survey
Graph Sparsifiers: A Survey Nick Harvey UBC Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman, Srivastava and Teng Approximating Dense Objects
More informationOn Some Extensions of Bernstein s Inequality for Self-Adjoint Operators
On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators Stanislav Minsker e-mail: sminsker@math.duke.edu Abstract: We present some extensions of Bernstein s inequality for random self-adjoint
More informationLab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018
Lab 8: Measuring Graph Centrality - PageRank Monday, November 5 CompSci 531, Fall 2018 Outline Measuring Graph Centrality: Motivation Random Walks, Markov Chains, and Stationarity Distributions Google
More informationLecture 8 : Eigenvalues and Eigenvectors
CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with
More informationOn the Spectra of General Random Graphs
On the Spectra of General Random Graphs Fan Chung Mary Radcliffe University of California, San Diego La Jolla, CA 92093 Abstract We consider random graphs such that each edge is determined by an independent
More informationLecture 3. Random Fourier measurements
Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our
More informationSpectral Theorem for Self-adjoint Linear Operators
Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;
More informationA Randomness-Efficient Sampler for Matrix-valued Functions and Applications
A Randomness-Efficient Sampler for Matrix-valued Functions and Applications Avi Wigderson Institute for Advanced Study Princeton, NJ 08540 avi@ias.edu David Xiao Princeton University Princeton, NJ 08544
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationSparsification by Effective Resistance Sampling
Spectral raph Theory Lecture 17 Sparsification by Effective Resistance Sampling Daniel A. Spielman November 2, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened
More informationSparsification of Graphs and Matrices
Sparsification of Graphs and Matrices Daniel A. Spielman Yale University joint work with Joshua Batson (MIT) Nikhil Srivastava (MSR) Shang- Hua Teng (USC) HUJI, May 21, 2014 Objective of Sparsification:
More informationDensity of States for Random Band Matrices in d = 2
Density of States for Random Band Matrices in d = 2 via the supersymmetric approach Mareike Lager Institute for applied mathematics University of Bonn Joint work with Margherita Disertori ZiF Summer School
More informationOn Composite Quantum Hypothesis Testing
University of York 7 November 207 On Composite Quantum Hypothesis Testing Mario Berta Department of Computing with Fernando Brandão and Christoph Hirche arxiv:709.07268 Overview Introduction 2 Composite
More informationAssignment 4: Solutions
Math 340: Discrete Structures II Assignment 4: Solutions. Random Walks. Consider a random walk on an connected, non-bipartite, undirected graph G. Show that, in the long run, the walk will traverse each
More informationMath 307 Learning Goals. March 23, 2010
Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationLecture 14: Random Walks, Local Graph Clustering, Linear Programming
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 14: Random Walks, Local Graph Clustering, Linear Programming Lecturer: Shayan Oveis Gharan 3/01/17 Scribe: Laura Vonessen Disclaimer: These
More informationLecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016
Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,
More informationApplied Linear Algebra
Applied Linear Algebra Peter J. Olver School of Mathematics University of Minnesota Minneapolis, MN 55455 olver@math.umn.edu http://www.math.umn.edu/ olver Chehrzad Shakiban Department of Mathematics University
More informationEfron Stein Inequalities for Random Matrices
Efron Stein Inequalities for Random Matrices Daniel Paulin, Lester Mackey and Joel A. Tropp D. Paulin Department of Statistics and Applied Probability National University of Singapore 6 Science Drive,
More informationLecture 8: Linear Algebra Background
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected
More informationConvergence of Random Walks
Graphs and Networks Lecture 9 Convergence of Random Walks Daniel A. Spielman September 30, 010 9.1 Overview We begin by reviewing the basics of spectral theory. We then apply this theory to show that lazy
More informationConcentration of Measures by Bounded Couplings
Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationProximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725
Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:
More informationA Tutorial on Matrix Approximation by Row Sampling
A Tutorial on Matrix Approximation by Row Sampling Rasmus Kyng June 11, 018 Contents 1 Fast Linear Algebra Talk 1.1 Matrix Concentration................................... 1. Algorithms for ɛ-approximation
More information1 Math 241A-B Homework Problem List for F2015 and W2016
1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let
More informationGROUP THEORY PRIMER. New terms: so(2n), so(2n+1), symplectic algebra sp(2n)
GROUP THEORY PRIMER New terms: so(2n), so(2n+1), symplectic algebra sp(2n) 1. Some examples of semi-simple Lie algebras In the previous chapter, we developed the idea of understanding semi-simple Lie algebras
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017
U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 3, 207 Scribed by Keyhan Vakil Lecture 3 In which we complete the study of Independent Set and Max Cut in G n,p random graphs.
More informationMod-φ convergence I: examples and probabilistic estimates
Mod-φ convergence I: examples and probabilistic estimates Valentin Féray (joint work with Pierre-Loïc Méliot and Ashkan Nikeghbali) Institut für Mathematik, Universität Zürich Summer school in Villa Volpi,
More informationGraph Sparsification III: Ramanujan Graphs, Lifts, and Interlacing Families
Graph Sparsification III: Ramanujan Graphs, Lifts, and Interlacing Families Nikhil Srivastava Microsoft Research India Simons Institute, August 27, 2014 The Last Two Lectures Lecture 1. Every weighted
More informationPCA with random noise. Van Ha Vu. Department of Mathematics Yale University
PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical
More informationExpander Construction in VNC 1
Expander Construction in VNC 1 Sam Buss joint work with Valentine Kabanets, Antonina Kolokolova & Michal Koucký Prague Workshop on Bounded Arithmetic November 2-3, 2017 Talk outline I. Combinatorial construction
More informationStein Couplings for Concentration of Measure
Stein Couplings for Concentration of Measure Jay Bartroff, Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] [arxiv:1402.6769] Borchard
More informationNonlinear Optimization
Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More information1 Matrices and vector spaces
Matrices and vector spaces. Which of the following statements about linear vector spaces are true? Where a statement is false, give a counter-example to demonstrate this. (a) Non-singular N N matrices
More informationOptimal compression of approximate Euclidean distances
Optimal compression of approximate Euclidean distances Noga Alon 1 Bo az Klartag 2 Abstract Let X be a set of n points of norm at most 1 in the Euclidean space R k, and suppose ε > 0. An ε-distance sketch
More informationFINITE-DIMENSIONAL LINEAR ALGEBRA
DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITE-DIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationMATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1
The Annals of Probability 014, Vol. 4, No. 3, 906 945 DOI: 10.114/13-AOP89 Institute of Mathematical Statistics, 014 MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1 BY LESTER MACKEY,MICHAEL
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationMath 489AB Exercises for Chapter 2 Fall Section 2.3
Math 489AB Exercises for Chapter 2 Fall 2008 Section 2.3 2.3.3. Let A M n (R). Then the eigenvalues of A are the roots of the characteristic polynomial p A (t). Since A is real, p A (t) is a polynomial
More informationA Hierarchy of Information Quantities for Finite Block Length Analysis of Quantum Tasks
A Hierarchy of Information Quantities for Finite Block Length Analysis of Quantum Tasks Marco Tomamichel, Masahito Hayashi arxiv: 1208.1478 Also discussing results of: Second Order Asymptotics for Quantum
More informationClarkson Inequalities With Several Operators
isid/ms/2003/23 August 14, 2003 http://www.isid.ac.in/ statmath/eprints Clarkson Inequalities With Several Operators Rajendra Bhatia Fuad Kittaneh Indian Statistical Institute, Delhi Centre 7, SJSS Marg,
More informationBounds for all eigenvalues of sums of Hermitian random matrices
20 Chapter 2 Bounds for all eigenvalues of sums of Hermitian random matrices 2.1 Introduction The classical tools of nonasymptotic random matrix theory can sometimes give quite sharp estimates of the extreme
More informationModel Counting for Logical Theories
Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern
More informationGeneralization theory
Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1
More informationLecture 17 November 8, 2012
6.841: Advanced Complexity Theory Fall 2012 Prof. Dana Moshkovitz Lecture 17 November 8, 2012 Scribe: Mark Bun 1 Overview In the previous lecture, we saw an overview of probabilistically checkable proofs,
More information08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms
(February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops
More informationProblem Set 2. Assigned: Mon. November. 23, 2015
Pseudorandomness Prof. Salil Vadhan Problem Set 2 Assigned: Mon. November. 23, 2015 Chi-Ning Chou Index Problem Progress 1 SchwartzZippel lemma 1/1 2 Robustness of the model 1/1 3 Zero error versus 1-sided
More informationCAAM 335 Matrix Analysis
CAAM 335 Matrix Analysis Solutions to Homework 8 Problem (5+5+5=5 points The partial fraction expansion of the resolvent for the matrix B = is given by (si B = s } {{ } =P + s + } {{ } =P + (s (5 points
More informationNotes on uniform convergence
Notes on uniform convergence Erik Wahlén erik.wahlen@math.lu.se January 17, 2012 1 Numerical sequences We begin by recalling some properties of numerical sequences. By a numerical sequence we simply mean
More informationUser-Friendly Tail Bounds for Sums of Random Matrices
DOI 10.1007/s10208-011-9099-z User-Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Received: 16 January 2011 / Accepted: 13 June 2011 The Authors 2011. This article is published with open
More informationRandomized algorithm
Tutorial 4 Joyce 2009-11-24 Outline Solution to Midterm Question 1 Question 2 Question 1 Question 2 Question 3 Question 4 Question 5 Solution to Midterm Solution to Midterm Solution to Midterm Question
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationA sequence of triangle-free pseudorandom graphs
A sequence of triangle-free pseudorandom graphs David Conlon Abstract A construction of Alon yields a sequence of highly pseudorandom triangle-free graphs with edge density significantly higher than one
More informationMath 577 Assignment 7
Math 577 Assignment 7 Thanks for Yu Cao 1. Solution. The linear system being solved is Ax = 0, where A is a (n 1 (n 1 matrix such that 2 1 1 2 1 A =......... 1 2 1 1 2 and x = (U 1, U 2,, U n 1. By the
More informationarxiv:math/ v1 [math.fa] 4 Jan 2007
Tr[ABA] p = Tr[B 1/2 A 2 B 1/2 ] p. On the Araki-Lieb-Thirring inequality arxiv:math/0701129v1 [math.fa] 4 Jan 2007 Koenraad M.R. Audenaert Institute for Mathematical Sciences, Imperial College London
More informationQuantum Physics II (8.05) Fall 2002 Assignment 3
Quantum Physics II (8.05) Fall 00 Assignment Readings The readings below will take you through the material for Problem Sets and 4. Cohen-Tannoudji Ch. II, III. Shankar Ch. 1 continues to be helpful. Sakurai
More informationNovember 18, 2013 ANALYTIC FUNCTIONAL CALCULUS
November 8, 203 ANALYTIC FUNCTIONAL CALCULUS RODICA D. COSTIN Contents. The spectral projection theorem. Functional calculus 2.. The spectral projection theorem for self-adjoint matrices 2.2. The spectral
More informationMATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 14: The Power Function and Native Space Error Estimates Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu
More informationor E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4)
:23 /4/2000 TOPIC Characteristic functions This lecture begins our study of the characteristic function φ X (t) := Ee itx = E cos(tx)+ie sin(tx) (t R) of a real random variable X Characteristic functions
More informationMathematical Methods for Engineers and Scientists 1
K.T. Tang Mathematical Methods for Engineers and Scientists 1 Complex Analysis, Determinants and Matrices With 49 Figures and 2 Tables fyj Springer Part I Complex Analysis 1 Complex Numbers 3 1.1 Our Number
More informationCourse Description - Master in of Mathematics Comprehensive exam& Thesis Tracks
Course Description - Master in of Mathematics Comprehensive exam& Thesis Tracks 1309701 Theory of ordinary differential equations Review of ODEs, existence and uniqueness of solutions for ODEs, existence
More informationIn particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with
Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient
More informationLecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora
princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora Scribe: Today we continue the
More informationGraph Partitioning Algorithms and Laplacian Eigenvalues
Graph Partitioning Algorithms and Laplacian Eigenvalues Luca Trevisan Stanford Based on work with Tsz Chiu Kwok, Lap Chi Lau, James Lee, Yin Tat Lee, and Shayan Oveis Gharan spectral graph theory Use linear
More informationarxiv: v1 [math.na] 1 Sep 2018
On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationExpansion properties of random Cayley graphs and vertex transitive graphs via matrix martingales
Expansion properties of random Cayley graphs and vertex transitive graphs via matrix martingales Demetres Christofides Klas Markström Abstract The Alon-Roichman theorem states that for every ε > 0 there
More informationTOEPLITZ OPERATORS. Toeplitz studied infinite matrices with NW-SE diagonals constant. f e C :
TOEPLITZ OPERATORS EFTON PARK 1. Introduction to Toeplitz Operators Otto Toeplitz lived from 1881-1940 in Goettingen, and it was pretty rough there, so he eventually went to Palestine and eventually contracted
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationMarch 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.
Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)
More informationDimension reduction for semidefinite programming
1 / 22 Dimension reduction for semidefinite programming Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology
More informationTail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).
Tail Inequalities William Hunt Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV William.Hunt@mail.wvu.edu Introduction In this chapter, we are interested
More informationMath 307 Learning Goals
Math 307 Learning Goals May 14, 2018 Chapter 1 Linear Equations 1.1 Solving Linear Equations Write a system of linear equations using matrix notation. Use Gaussian elimination to bring a system of linear
More informationL. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1
L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones definition spectral decomposition quadratic representation log-det barrier 18-1 Introduction This lecture: theoretical properties of the following
More informationQuantum Field Theory III
Quantum Field Theory III Prof. Erick Weinberg January 19, 2011 1 Lecture 1 1.1 Structure We will start with a bit of group theory, and we will talk about spontaneous symmetry broken. Then we will talk
More informationA new method to bound rate of convergence
A new method to bound rate of convergence Arup Bose Indian Statistical Institute, Kolkata, abose@isical.ac.in Sourav Chatterjee University of California, Berkeley Empirical Spectral Distribution Distribution
More informationThe Solovay-Kitaev theorem
The Solovay-Kitaev theorem Maris Ozols December 10, 009 1 Introduction There are several accounts of the Solovay-Kitaev theorem available [K97, NC00, KSV0, DN05]. I chose to base my report on [NC00], since
More informationAnalysis Preliminary Exam Workshop: Hilbert Spaces
Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H
More informationChapter 7. Homogeneous equations with constant coefficients
Chapter 7. Homogeneous equations with constant coefficients It has already been remarked that we can write down a formula for the general solution of any linear second differential equation y + a(t)y +
More informationCentral Limit Theorems for linear statistics for Biorthogonal Ensembles
Central Limit Theorems for linear statistics for Biorthogonal Ensembles Maurice Duits, Stockholm University Based on joint work with Jonathan Breuer (HUJI) Princeton, April 2, 2014 M. Duits (SU) CLT s
More informationSolutions to Problem Set 4
UC Berkeley, CS 174: Combinatorics and Discrete Probability (Fall 010 Solutions to Problem Set 4 1. (MU 5.4 In a lecture hall containing 100 people, you consider whether or not there are three people in
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More information