A Matrix Expander Chernoff Bound

Size: px
Start display at page:

Download "A Matrix Expander Chernoff Bound"

Transcription

1 A Matrix Expander Chernoff Bound Ankit Garg, Yin Tat Lee, Zhao Song, Nikhil Srivastava UC Berkeley

2 Vanilla Chernoff Bound Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4)

3 Vanilla Chernoff Bound Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) Two Extensions: 1.Dependent Random Variables 2.Sums of random matrices

4 Expander Chernoff Bound [AKS 87, G 94] Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0.

5 Expander Chernoff Bound [AKS 87, G 94] Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 ) Implies walk of length k 1 λ 1 concentrates around mean.

6 Intuition for Dependence on Spectral Gap 1 λ = 1 n 2 f = +1 f = 1

7 Intuition for Dependence on Spectral Gap 1 λ = 1 n 2 Typical random walk takes Ω n 2 steps to see both ±1.

8 Derandomization Motivation Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 ) To generate k indep samples from [n] need k log n bits. If G is 3 regular with constant λ, walk needs: log n + k log(3) bits.

9 Derandomization Motivation Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 ) To generate k indep samples from [n] need k log n bits. If G is 3 regular with constant λ, walk needs: log n + k log(3) bits. When k = O(log n) reduces randomness quadratically. Can completely derandomize in polynomial time.

10 Matrix Chernoff Bound Thm [Rudelson 97, Ahlswede-Winter 02, Oliveira 08, Tropp 11 ]. If X 1,, X k are independent mean zero random d d Hermitian matrices with X i 1 then P 1 k i X i ε 2dexp( kε 2 /4)

11 Matrix Chernoff Bound Thm [Rudelson 97, Ahlswede-Winter 02, Oliveira 08, Tropp 11 ]. If X 1,, X k are independent mean zero random d d Hermitian matrices with X i 1 then P 1 k i X i ε 2dexp( kε 2 /4) Factor d is tight because of the diagonal case. Very generic bound (no independence assumptions on the entries). Many applications + martingale extensions (see Tropp).

12 Matrix Expander Chernoff Bound? Conj[Wigderson-Xiao 05]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V C d d be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2dexp( c(1 λ)kε 2 ) Motivated by derandomized Alon-Roichman theorem.

13 Main Theorem Thm. Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V C d d be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2dexp( c(1 λ)kε 2 ) Gives black-box derandomization of any application of matrix Chernoff

14 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp t X i = e tkε Ee tx i i i i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4

15 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Markov

16 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Indep.

17 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Bounded

18 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4 Mean zero

19 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4

20 1. Proof of Chernoff: reduction to mgf Thm [Hoeffding, Chernoff]. If X 1,, X k are independent mean zero random variables with X i 1 then P 1 k i X i ε 2exp( kε 2 /4) P X i kε e tkε E exp i t i X i = e tkε i Ee tx i e tkε 1 + tex i + t 2 k e tkε+kt2 e kε2 4

21 2. Proof of Expander Chernoff Goal: Show E exp t σ i f v i exp c λ kt 2

22 2. Proof of Expander Chernoff Goal: Show E exp t σ i f v i exp c λ kt 2 Issue: Eexp σ i f v i ς i Eexp(tf(v i )) How to control the mgf without independence?

23 Step 1: Write mgf as quadratic form Ee t σ i k f(v i) = i 0,,i k V P(v 0 = i 0 )P i 0, i 1 P(i k 1, i k ) exp t 1 j k f i j = 1 n i 0,,i k V P i 0, i 1 e tf(i 1) P i k 1, i k e tf(i k) = u, EP k u where u = 1 n,, 1 E = e tf(1) n. i 0 e tf(i 1) i 1 e tf(i 3) i 3 P i 0, i 1 P i 1, i 2 i 2 e tf(n) P i 2, i 3 e tf(i 2)

24 Step 1: Write mgf as quadratic form Ee t σ i k f(v i) = i 0,,i k V P(v 0 = i 0 )P i 0, i 1 P(i k 1, i k ) exp t 1 j k f i j = 1 n i 0,,i k V P i 0, i 1 e tf(i 1) P i k 1, i k e tf(i k) = u, EP k u where u = 1 n,, 1 E = e tf(1) n. i 0 e tf(i 1) i 1 e tf(i 3) i 3 P i 0, i 1 P i 1, i 2 i 2 e tf(n) P i 2, i 3 e tf(i 2)

25 Step 1: Write mgf as quadratic form Ee t σ i k f(v i) = i 0,,i k V P(v 0 = i 0 )P i 0, i 1 P(i k 1, i k ) exp t 1 j k f i j = 1 n i 0,,i k V P i 0, i 1 e tf(i 1) P i k 1, i k e tf(i k) = u, EP k u where u = 1 n,, 1 E = e tf(1) n. i 0 e tf(i 1) i 1 e tf(i 3) i 3 P i 0, i 1 P i 1, i 2 i 2 e tf(n) P i 2, i 3 e tf(i 2)

26 Step 2: Bound quadratic form Goal: Show u, EP k u exp c λ kt 2 Observe: P J λ where J = complete graph with self loops So for small λ, should have u, EP k u u, (EJ) k u

27 Step 2: Bound quadratic form Goal: Show u, EP k u exp c λ kt 2 Observe: P J λ where J = complete graph with self loops So for small λ, should have u, EP k u u, (EJ) k u Approach 1. Use perturbation theory to show EP exp c λ t 2

28 Step 2: Bound quadratic form Goal: Show u, EP k u exp c λ kt 2 Observe: P J λ where J = complete graph with self loops So for small λ, should have u, EP k u u, (EJ) k u Approach 1. Use perturbation theory to show EP exp c λ t 2 Approach 2. (Healy 08) track projection of iterates along u.

29 Simplest case: λ = 0 EPEPEPEPu

30 Simplest case: λ = 0 EPEPEPEPu

31 Simplest case: λ = 0 EPEPEPEPu

32 Simplest case: λ = 0 EPEPEPEPu

33 Simplest case: λ = 0 EPEPEPEPu

34 Observations EPEPEPEPu Observe: P shrinks every vector orthogonal to u by λ. condition. u, Eu = 1 n σ v V e tf(v) = 1 + t 2 by mean zero

35 A small dynamical system m = Q v v = EP j u m = v, u m m

36 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m 0 λ = 0 0 v EPv

37 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m e t λ any λ λt v EPv

38 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m 1 t log 1/λ λt v EPv

39 A small dynamical system m = Q v v = EP j u m = v, u e t2 t m m 1 λt Any mass that leaves gets shrunk by λ

40 Analyzing dynamical system gives Thm[Gillman 94]: Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V R be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2exp( c(1 λ)kε 2 )

41 Generalization to Matrices? Setup: f: V C d d, random walk v 1,, v k. Goal: ETr exp t i f v i d exp ckt 2 where e A is defined as a power series. Main Issue: exp A + B exp A exp(b) unless A, B = 0 can t express exp(sum) as iterated product.

42 Generalization to Matrices? Setup: f: V C d d, random walk v 1,, v k. Goal: ETr exp t i f v i d exp ckt 2 where e A is defined as a power series. Main Issue: exp A + B exp A exp(b) unless A, B = 0 can t express exp(sum) as iterated product.

43 The Golden-Thompson Inequality Partial Workaround [Golden-Thompson 65]: Tr exp A + B Tr exp A exp B Sufficient for independent case by induction. For expander case, need this for k matrices. False! e A Tr e A+B+C > 0 > Tr(e A e B e C ) e B e C

44 The Golden-Thompson Inequality Partial Workaround [Golden-Thompson 65]: Tr exp A + B Tr exp A exp B Sufficient for independent case by induction. For expander case, need this for k matrices. e A Tr e A+B+C > 0 > Tr(e A e B e C ) e B e C

45 The Golden-Thompson Inequality Partial Workaround [Golden-Thompson 65]: Tr exp A + B Tr exp A exp B Sufficient for independent case by induction. For expander case, need this for k matrices. False! e A Tr e A+B+C > 0 > Tr(e A e B e C ) e B e C

46 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R. 1. Matrix on RHS is always PSD. 2. Average-case inequality: e A i/2 are conjugated by unitaries. 3. Implies Lieb s concavity, triple-matrix, ALT, and more.

47 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R. 1. Matrix on RHS is always PSD. 2. Average-case inequality: e A i/2 are conjugated by unitaries. 3. Implies Lieb s concavity, triple-matrix, ALT, and more.

48 Proof of SBT: Lie-Trotter Formula e A+B+C = lim θ 0 + eθa e θb θc 1/θ e log Tr e A+B+C = lim θ 0 + log G θ 2/ θ For G z e za 2 e zb 2 e zc 2

49 Proof of SBT: Lie-Trotter Formula e A+B+C = lim θ 0 + eθa e θb θc 1/θ e log Tr e A+B+C = lim θ 0 + 2log G θ 2/ θ /θ For G z e za 2 e zb 2 e zc 2

50 Complex Interpolation (Stein-Hirschman) θ log G θ 2/θ

51 Complex Interpolation (Stein-Hirschman) For each θ, find analytic F(z) st: F it = 1 F 1 + it G 1 + it 2 F θ = G θ 2/θ

52 Complex Interpolation (Stein-Hirschman) For each θ, find analytic F(z) st: F it = 1 F 1 + it G 1 + it 2 F θ = G θ 2/θ log F θ නlog F it + නlog F 1 + it

53 Complex Interpolation (Stein-Hirschman) For each θ, find analytic F(z) st: F it = 1 F 1 + it G 1 + it 2 F θ = G θ 2/θ lim θ 0 log F θ නlog F it + නlog F 1 + it

54 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R.

55 Key Ingredient [Sutter-Berta-Tomamichel 16] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1(1+ib) 2 e A k(1+ib) 2 e A 1(1+ib) 2 e A k(1+ib) 2 where β(b) is an explicit probability density on R. Issue. SBT involves integration over unbounded region, bad for Taylor expansion.

56 Bounded Modification of SBT Solution. Prove bounded version of SBT by replacing strip with half-disk. [Thm] If A 1,, A k are Hermitian, then log Tr e A 1+ +A k dβ(b) log Tr e A 1e ib 2 e A ke ib 2 e A 1e ib 2 e A ke ib 2 θ where β(b) is an explicit probability density on π 2, π 2. Proof. Analytic F(z) + Poisson Kernel + Riemann map.

57 Handling Two-sided Products Issue. Two-sided rather than one-sided products: Tr e tf v 1 e ib 2 e tf v k e ib 2 e tf v 1 e ib 2 e tf v k e ib 2 Solution. Encode as one-sided product by using Tr AXB = A B T vec X : e tf v 1 e ib 2 e tf v 1 e ib 2 e tf v k T e ib 2 e tf v k T e ib 2 vec I d, vec I d

58 Handling Two-sided Products Issue. Two-sided rather than one-sided products: Tr e tf v 1 e ib 2 e tf v k e ib 2 e tf v 1 e ib 2 e tf v k e ib 2 Solution. Encode as one-sided product by using Tr AXB = A B T vec X : e tf v 1 e ib 2 e tf v 1 T e ib 2 e tf v k T e ib 2 e tf v k T e ib 2 vec I d, vec I d

59 Finishing the Proof Carry out a version of Healy s argument with P I d 2 and: E = tf 1 eib e 2 e tf 1 T e ib 2 e tf(n)eib 2 e tf n T e ib 2 i 0 And vec(i d ) u instead of u. This leads to the additional d factor. i 1 P i 1, i 2 P i 0, i 1 i 2 i 3 P i 2, i 3

60 Main Theorem Thm. Suppose G = (V, E) is a regular graph with transition matrix P which has second eigenvalue λ. Let f: V C d d be a function with f v 1 and σ v f v = 0. Then, if v 1,, v k is a stationary random walk: P 1 k i f(v i ) ε 2dexp( c(1 λ)kε 2 )

61 Open Questions Other matrix concentration inequalities (multiplicative, low-rank, moments) Other Banach spaces (Schatten norms) More applications of complex interpolation

Matrix Concentration. Nick Harvey University of British Columbia

Matrix Concentration. Nick Harvey University of British Columbia Matrix Concentration Nick Harvey University of British Columbia The Problem Given any random nxn, symmetric matrices Y 1,,Y k. Show that i Y i is probably close to E[ i Y i ]. Why? A matrix generalization

More information

Stein s Method for Matrix Concentration

Stein s Method for Matrix Concentration Stein s Method for Matrix Concentration Lester Mackey Collaborators: Michael I. Jordan, Richard Y. Chen, Brendan Farrell, and Joel A. Tropp University of California, Berkeley California Institute of Technology

More information

Multivariate trace inequalities. David Sutter, Mario Berta, Marco Tomamichel

Multivariate trace inequalities. David Sutter, Mario Berta, Marco Tomamichel Multivariate trace inequalities David Sutter, Mario Berta, Marco Tomamichel What are trace inequalities and why we should care. Main difference between classical and quantum world are complementarity and

More information

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER 1. The method Ashwelde and Winter [1] proposed a new approach to deviation inequalities for sums of independent random matrices. The

More information

Multivariate Trace Inequalities

Multivariate Trace Inequalities Multivariate Trace Inequalities Mario Berta arxiv:1604.03023 with Sutter and Tomamichel (to appear in CMP) arxiv:1512.02615 with Fawzi and Tomamichel QMath13 - October 8, 2016 Mario Berta (Caltech) Multivariate

More information

arxiv: v1 [math.pr] 11 Feb 2019

arxiv: v1 [math.pr] 11 Feb 2019 A Short Note on Concentration Inequalities for Random Vectors with SubGaussian Norm arxiv:190.03736v1 math.pr] 11 Feb 019 Chi Jin University of California, Berkeley chijin@cs.berkeley.edu Rong Ge Duke

More information

Graph Sparsification I : Effective Resistance Sampling

Graph Sparsification I : Effective Resistance Sampling Graph Sparsification I : Effective Resistance Sampling Nikhil Srivastava Microsoft Research India Simons Institute, August 26 2014 Graphs G G=(V,E,w) undirected V = n w: E R + Sparsification Approximate

More information

Graph Sparsification II: Rank one updates, Interlacing, and Barriers

Graph Sparsification II: Rank one updates, Interlacing, and Barriers Graph Sparsification II: Rank one updates, Interlacing, and Barriers Nikhil Srivastava Simons Institute August 26, 2014 Previous Lecture Definition. H = (V, F, u) is a κ approximation of G = V, E, w if:

More information

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 Lecture 3 In which we show how to find a planted clique in a random graph. 1 Finding a Planted Clique We will analyze

More information

Using Friendly Tail Bounds for Sums of Random Matrices

Using Friendly Tail Bounds for Sums of Random Matrices Using Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Computing + Mathematical Sciences California Institute of Technology jtropp@cms.caltech.edu Research supported in part by NSF, DARPA,

More information

Matrix concentration inequalities

Matrix concentration inequalities ELE 538B: Mathematics of High-Dimensional Data Matrix concentration inequalities Yuxin Chen Princeton University, Fall 2018 Recap: matrix Bernstein inequality Consider a sequence of independent random

More information

Graph Sparsifiers: A Survey

Graph Sparsifiers: A Survey Graph Sparsifiers: A Survey Nick Harvey UBC Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman, Srivastava and Teng Approximating Dense Objects

More information

On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators

On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators On Some Extensions of Bernstein s Inequality for Self-Adjoint Operators Stanislav Minsker e-mail: sminsker@math.duke.edu Abstract: We present some extensions of Bernstein s inequality for random self-adjoint

More information

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018 Lab 8: Measuring Graph Centrality - PageRank Monday, November 5 CompSci 531, Fall 2018 Outline Measuring Graph Centrality: Motivation Random Walks, Markov Chains, and Stationarity Distributions Google

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

On the Spectra of General Random Graphs

On the Spectra of General Random Graphs On the Spectra of General Random Graphs Fan Chung Mary Radcliffe University of California, San Diego La Jolla, CA 92093 Abstract We consider random graphs such that each edge is determined by an independent

More information

Lecture 3. Random Fourier measurements

Lecture 3. Random Fourier measurements Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our

More information

Spectral Theorem for Self-adjoint Linear Operators

Spectral Theorem for Self-adjoint Linear Operators Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;

More information

A Randomness-Efficient Sampler for Matrix-valued Functions and Applications

A Randomness-Efficient Sampler for Matrix-valued Functions and Applications A Randomness-Efficient Sampler for Matrix-valued Functions and Applications Avi Wigderson Institute for Advanced Study Princeton, NJ 08540 avi@ias.edu David Xiao Princeton University Princeton, NJ 08544

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

Sparsification by Effective Resistance Sampling

Sparsification by Effective Resistance Sampling Spectral raph Theory Lecture 17 Sparsification by Effective Resistance Sampling Daniel A. Spielman November 2, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened

More information

Sparsification of Graphs and Matrices

Sparsification of Graphs and Matrices Sparsification of Graphs and Matrices Daniel A. Spielman Yale University joint work with Joshua Batson (MIT) Nikhil Srivastava (MSR) Shang- Hua Teng (USC) HUJI, May 21, 2014 Objective of Sparsification:

More information

Density of States for Random Band Matrices in d = 2

Density of States for Random Band Matrices in d = 2 Density of States for Random Band Matrices in d = 2 via the supersymmetric approach Mareike Lager Institute for applied mathematics University of Bonn Joint work with Margherita Disertori ZiF Summer School

More information

On Composite Quantum Hypothesis Testing

On Composite Quantum Hypothesis Testing University of York 7 November 207 On Composite Quantum Hypothesis Testing Mario Berta Department of Computing with Fernando Brandão and Christoph Hirche arxiv:709.07268 Overview Introduction 2 Composite

More information

Assignment 4: Solutions

Assignment 4: Solutions Math 340: Discrete Structures II Assignment 4: Solutions. Random Walks. Consider a random walk on an connected, non-bipartite, undirected graph G. Show that, in the long run, the walk will traverse each

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

Probability Background

Probability Background Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second

More information

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 14: Random Walks, Local Graph Clustering, Linear Programming Lecturer: Shayan Oveis Gharan 3/01/17 Scribe: Laura Vonessen Disclaimer: These

More information

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,

More information

Applied Linear Algebra

Applied Linear Algebra Applied Linear Algebra Peter J. Olver School of Mathematics University of Minnesota Minneapolis, MN 55455 olver@math.umn.edu http://www.math.umn.edu/ olver Chehrzad Shakiban Department of Mathematics University

More information

Efron Stein Inequalities for Random Matrices

Efron Stein Inequalities for Random Matrices Efron Stein Inequalities for Random Matrices Daniel Paulin, Lester Mackey and Joel A. Tropp D. Paulin Department of Statistics and Applied Probability National University of Singapore 6 Science Drive,

More information

Lecture 8: Linear Algebra Background

Lecture 8: Linear Algebra Background CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected

More information

Convergence of Random Walks

Convergence of Random Walks Graphs and Networks Lecture 9 Convergence of Random Walks Daniel A. Spielman September 30, 010 9.1 Overview We begin by reviewing the basics of spectral theory. We then apply this theory to show that lazy

More information

Concentration of Measures by Bounded Couplings

Concentration of Measures by Bounded Couplings Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725 Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:

More information

A Tutorial on Matrix Approximation by Row Sampling

A Tutorial on Matrix Approximation by Row Sampling A Tutorial on Matrix Approximation by Row Sampling Rasmus Kyng June 11, 018 Contents 1 Fast Linear Algebra Talk 1.1 Matrix Concentration................................... 1. Algorithms for ɛ-approximation

More information

1 Math 241A-B Homework Problem List for F2015 and W2016

1 Math 241A-B Homework Problem List for F2015 and W2016 1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let

More information

GROUP THEORY PRIMER. New terms: so(2n), so(2n+1), symplectic algebra sp(2n)

GROUP THEORY PRIMER. New terms: so(2n), so(2n+1), symplectic algebra sp(2n) GROUP THEORY PRIMER New terms: so(2n), so(2n+1), symplectic algebra sp(2n) 1. Some examples of semi-simple Lie algebras In the previous chapter, we developed the idea of understanding semi-simple Lie algebras

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017 U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 3, 207 Scribed by Keyhan Vakil Lecture 3 In which we complete the study of Independent Set and Max Cut in G n,p random graphs.

More information

Mod-φ convergence I: examples and probabilistic estimates

Mod-φ convergence I: examples and probabilistic estimates Mod-φ convergence I: examples and probabilistic estimates Valentin Féray (joint work with Pierre-Loïc Méliot and Ashkan Nikeghbali) Institut für Mathematik, Universität Zürich Summer school in Villa Volpi,

More information

Graph Sparsification III: Ramanujan Graphs, Lifts, and Interlacing Families

Graph Sparsification III: Ramanujan Graphs, Lifts, and Interlacing Families Graph Sparsification III: Ramanujan Graphs, Lifts, and Interlacing Families Nikhil Srivastava Microsoft Research India Simons Institute, August 27, 2014 The Last Two Lectures Lecture 1. Every weighted

More information

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical

More information

Expander Construction in VNC 1

Expander Construction in VNC 1 Expander Construction in VNC 1 Sam Buss joint work with Valentine Kabanets, Antonina Kolokolova & Michal Koucký Prague Workshop on Bounded Arithmetic November 2-3, 2017 Talk outline I. Combinatorial construction

More information

Stein Couplings for Concentration of Measure

Stein Couplings for Concentration of Measure Stein Couplings for Concentration of Measure Jay Bartroff, Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] [arxiv:1402.6769] Borchard

More information

Nonlinear Optimization

Nonlinear Optimization Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

1 Matrices and vector spaces

1 Matrices and vector spaces Matrices and vector spaces. Which of the following statements about linear vector spaces are true? Where a statement is false, give a counter-example to demonstrate this. (a) Non-singular N N matrices

More information

Optimal compression of approximate Euclidean distances

Optimal compression of approximate Euclidean distances Optimal compression of approximate Euclidean distances Noga Alon 1 Bo az Klartag 2 Abstract Let X be a set of n points of norm at most 1 in the Euclidean space R k, and suppose ε > 0. An ε-distance sketch

More information

FINITE-DIMENSIONAL LINEAR ALGEBRA

FINITE-DIMENSIONAL LINEAR ALGEBRA DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITE-DIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Math 108b: Notes on the Spectral Theorem

Math 108b: Notes on the Spectral Theorem Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator

More information

MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1

MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1 The Annals of Probability 014, Vol. 4, No. 3, 906 945 DOI: 10.114/13-AOP89 Institute of Mathematical Statistics, 014 MATRIX CONCENTRATION INEQUALITIES VIA THE METHOD OF EXCHANGEABLE PAIRS 1 BY LESTER MACKEY,MICHAEL

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Math 489AB Exercises for Chapter 2 Fall Section 2.3 Math 489AB Exercises for Chapter 2 Fall 2008 Section 2.3 2.3.3. Let A M n (R). Then the eigenvalues of A are the roots of the characteristic polynomial p A (t). Since A is real, p A (t) is a polynomial

More information

A Hierarchy of Information Quantities for Finite Block Length Analysis of Quantum Tasks

A Hierarchy of Information Quantities for Finite Block Length Analysis of Quantum Tasks A Hierarchy of Information Quantities for Finite Block Length Analysis of Quantum Tasks Marco Tomamichel, Masahito Hayashi arxiv: 1208.1478 Also discussing results of: Second Order Asymptotics for Quantum

More information

Clarkson Inequalities With Several Operators

Clarkson Inequalities With Several Operators isid/ms/2003/23 August 14, 2003 http://www.isid.ac.in/ statmath/eprints Clarkson Inequalities With Several Operators Rajendra Bhatia Fuad Kittaneh Indian Statistical Institute, Delhi Centre 7, SJSS Marg,

More information

Bounds for all eigenvalues of sums of Hermitian random matrices

Bounds for all eigenvalues of sums of Hermitian random matrices 20 Chapter 2 Bounds for all eigenvalues of sums of Hermitian random matrices 2.1 Introduction The classical tools of nonasymptotic random matrix theory can sometimes give quite sharp estimates of the extreme

More information

Model Counting for Logical Theories

Model Counting for Logical Theories Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern

More information

Generalization theory

Generalization theory Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1

More information

Lecture 17 November 8, 2012

Lecture 17 November 8, 2012 6.841: Advanced Complexity Theory Fall 2012 Prof. Dana Moshkovitz Lecture 17 November 8, 2012 Scribe: Mark Bun 1 Overview In the previous lecture, we saw an overview of probabilistically checkable proofs,

More information

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms (February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops

More information

Problem Set 2. Assigned: Mon. November. 23, 2015

Problem Set 2. Assigned: Mon. November. 23, 2015 Pseudorandomness Prof. Salil Vadhan Problem Set 2 Assigned: Mon. November. 23, 2015 Chi-Ning Chou Index Problem Progress 1 SchwartzZippel lemma 1/1 2 Robustness of the model 1/1 3 Zero error versus 1-sided

More information

CAAM 335 Matrix Analysis

CAAM 335 Matrix Analysis CAAM 335 Matrix Analysis Solutions to Homework 8 Problem (5+5+5=5 points The partial fraction expansion of the resolvent for the matrix B = is given by (si B = s } {{ } =P + s + } {{ } =P + (s (5 points

More information

Notes on uniform convergence

Notes on uniform convergence Notes on uniform convergence Erik Wahlén erik.wahlen@math.lu.se January 17, 2012 1 Numerical sequences We begin by recalling some properties of numerical sequences. By a numerical sequence we simply mean

More information

User-Friendly Tail Bounds for Sums of Random Matrices

User-Friendly Tail Bounds for Sums of Random Matrices DOI 10.1007/s10208-011-9099-z User-Friendly Tail Bounds for Sums of Random Matrices Joel A. Tropp Received: 16 January 2011 / Accepted: 13 June 2011 The Authors 2011. This article is published with open

More information

Randomized algorithm

Randomized algorithm Tutorial 4 Joyce 2009-11-24 Outline Solution to Midterm Question 1 Question 2 Question 1 Question 2 Question 3 Question 4 Question 5 Solution to Midterm Solution to Midterm Solution to Midterm Question

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

A sequence of triangle-free pseudorandom graphs

A sequence of triangle-free pseudorandom graphs A sequence of triangle-free pseudorandom graphs David Conlon Abstract A construction of Alon yields a sequence of highly pseudorandom triangle-free graphs with edge density significantly higher than one

More information

Math 577 Assignment 7

Math 577 Assignment 7 Math 577 Assignment 7 Thanks for Yu Cao 1. Solution. The linear system being solved is Ax = 0, where A is a (n 1 (n 1 matrix such that 2 1 1 2 1 A =......... 1 2 1 1 2 and x = (U 1, U 2,, U n 1. By the

More information

arxiv:math/ v1 [math.fa] 4 Jan 2007

arxiv:math/ v1 [math.fa] 4 Jan 2007 Tr[ABA] p = Tr[B 1/2 A 2 B 1/2 ] p. On the Araki-Lieb-Thirring inequality arxiv:math/0701129v1 [math.fa] 4 Jan 2007 Koenraad M.R. Audenaert Institute for Mathematical Sciences, Imperial College London

More information

Quantum Physics II (8.05) Fall 2002 Assignment 3

Quantum Physics II (8.05) Fall 2002 Assignment 3 Quantum Physics II (8.05) Fall 00 Assignment Readings The readings below will take you through the material for Problem Sets and 4. Cohen-Tannoudji Ch. II, III. Shankar Ch. 1 continues to be helpful. Sakurai

More information

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS November 8, 203 ANALYTIC FUNCTIONAL CALCULUS RODICA D. COSTIN Contents. The spectral projection theorem. Functional calculus 2.. The spectral projection theorem for self-adjoint matrices 2.2. The spectral

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 14: The Power Function and Native Space Error Estimates Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu

More information

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4)

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4) :23 /4/2000 TOPIC Characteristic functions This lecture begins our study of the characteristic function φ X (t) := Ee itx = E cos(tx)+ie sin(tx) (t R) of a real random variable X Characteristic functions

More information

Mathematical Methods for Engineers and Scientists 1

Mathematical Methods for Engineers and Scientists 1 K.T. Tang Mathematical Methods for Engineers and Scientists 1 Complex Analysis, Determinants and Matrices With 49 Figures and 2 Tables fyj Springer Part I Complex Analysis 1 Complex Numbers 3 1.1 Our Number

More information

Course Description - Master in of Mathematics Comprehensive exam& Thesis Tracks

Course Description - Master in of Mathematics Comprehensive exam& Thesis Tracks Course Description - Master in of Mathematics Comprehensive exam& Thesis Tracks 1309701 Theory of ordinary differential equations Review of ODEs, existence and uniqueness of solutions for ODEs, existence

More information

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient

More information

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora Scribe: Today we continue the

More information

Graph Partitioning Algorithms and Laplacian Eigenvalues

Graph Partitioning Algorithms and Laplacian Eigenvalues Graph Partitioning Algorithms and Laplacian Eigenvalues Luca Trevisan Stanford Based on work with Tsz Chiu Kwok, Lap Chi Lau, James Lee, Yin Tat Lee, and Shayan Oveis Gharan spectral graph theory Use linear

More information

arxiv: v1 [math.na] 1 Sep 2018

arxiv: v1 [math.na] 1 Sep 2018 On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Expansion properties of random Cayley graphs and vertex transitive graphs via matrix martingales

Expansion properties of random Cayley graphs and vertex transitive graphs via matrix martingales Expansion properties of random Cayley graphs and vertex transitive graphs via matrix martingales Demetres Christofides Klas Markström Abstract The Alon-Roichman theorem states that for every ε > 0 there

More information

TOEPLITZ OPERATORS. Toeplitz studied infinite matrices with NW-SE diagonals constant. f e C :

TOEPLITZ OPERATORS. Toeplitz studied infinite matrices with NW-SE diagonals constant. f e C : TOEPLITZ OPERATORS EFTON PARK 1. Introduction to Toeplitz Operators Otto Toeplitz lived from 1881-1940 in Goettingen, and it was pretty rough there, so he eventually went to Palestine and eventually contracted

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang. Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)

More information

Dimension reduction for semidefinite programming

Dimension reduction for semidefinite programming 1 / 22 Dimension reduction for semidefinite programming Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology

More information

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials). Tail Inequalities William Hunt Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV William.Hunt@mail.wvu.edu Introduction In this chapter, we are interested

More information

Math 307 Learning Goals

Math 307 Learning Goals Math 307 Learning Goals May 14, 2018 Chapter 1 Linear Equations 1.1 Solving Linear Equations Write a system of linear equations using matrix notation. Use Gaussian elimination to bring a system of linear

More information

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1 L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones definition spectral decomposition quadratic representation log-det barrier 18-1 Introduction This lecture: theoretical properties of the following

More information

Quantum Field Theory III

Quantum Field Theory III Quantum Field Theory III Prof. Erick Weinberg January 19, 2011 1 Lecture 1 1.1 Structure We will start with a bit of group theory, and we will talk about spontaneous symmetry broken. Then we will talk

More information

A new method to bound rate of convergence

A new method to bound rate of convergence A new method to bound rate of convergence Arup Bose Indian Statistical Institute, Kolkata, abose@isical.ac.in Sourav Chatterjee University of California, Berkeley Empirical Spectral Distribution Distribution

More information

The Solovay-Kitaev theorem

The Solovay-Kitaev theorem The Solovay-Kitaev theorem Maris Ozols December 10, 009 1 Introduction There are several accounts of the Solovay-Kitaev theorem available [K97, NC00, KSV0, DN05]. I chose to base my report on [NC00], since

More information

Analysis Preliminary Exam Workshop: Hilbert Spaces

Analysis Preliminary Exam Workshop: Hilbert Spaces Analysis Preliminary Exam Workshop: Hilbert Spaces 1. Hilbert spaces A Hilbert space H is a complete real or complex inner product space. Consider complex Hilbert spaces for definiteness. If (, ) : H H

More information

Chapter 7. Homogeneous equations with constant coefficients

Chapter 7. Homogeneous equations with constant coefficients Chapter 7. Homogeneous equations with constant coefficients It has already been remarked that we can write down a formula for the general solution of any linear second differential equation y + a(t)y +

More information

Central Limit Theorems for linear statistics for Biorthogonal Ensembles

Central Limit Theorems for linear statistics for Biorthogonal Ensembles Central Limit Theorems for linear statistics for Biorthogonal Ensembles Maurice Duits, Stockholm University Based on joint work with Jonathan Breuer (HUJI) Princeton, April 2, 2014 M. Duits (SU) CLT s

More information

Solutions to Problem Set 4

Solutions to Problem Set 4 UC Berkeley, CS 174: Combinatorics and Discrete Probability (Fall 010 Solutions to Problem Set 4 1. (MU 5.4 In a lecture hall containing 100 people, you consider whether or not there are three people in

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information