Quadrature for the Finite Free Convolution

Similar documents
Ramanujan Graphs of Every Size

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min

Matching Polynomials of Graphs

Bipartite Ramanujan Graphs of Every Degree

Effective Resistance and Schur Complements

Strongly Regular Graphs, part 1

Rings, Paths, and Paley Graphs

Review of Linear Algebra Definitions, Change of Basis, Trace, Spectral Theorem

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

Physical Metaphors for Graphs

Rings, Paths, and Cayley Graphs

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

Symmetries and Polynomials

Lecture 6: Lies, Inner Product Spaces, and Symmetric Matrices

Iterative solvers for linear equations

Eigenvalue PDFs. Peter Forrester, M&S, University of Melbourne

Math 108b: Notes on the Spectral Theorem

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

A Brief Outline of Math 355

4 Group representations

Math Linear Algebra Final Exam Review Sheet

Quantum Computing Lecture 2. Review of Linear Algebra

Fiedler s Theorems on Nodal Domains

Fiedler s Theorems on Nodal Domains

Eigenvectors and Hermitian Operators

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Spectral Graph Theory and You: Matrix Tree Theorem and Centrality Metrics

ORIE 6334 Spectral Graph Theory September 8, Lecture 6. In order to do the first proof, we need to use the following fact.

Recall the convention that, for us, all vectors are column vectors.

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

Iterative solvers for linear equations

1. General Vector Spaces

There are two things that are particularly nice about the first basis

88 CHAPTER 3. SYMMETRIES

Geometry of the Special Unitary Group

Topics in Representation Theory: Fourier Analysis and the Peter Weyl Theorem

Linear Algebra: Matrix Eigenvalue Problems

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Algebra I Fall 2007

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Review of Linear Algebra

1 Planar rotations. Math Abstract Linear Algebra Fall 2011, section E1 Orthogonal matrices and rotations

Math 416, Spring 2010 Gram-Schmidt, the QR-factorization, Orthogonal Matrices March 4, 2010 GRAM-SCHMIDT, THE QR-FACTORIZATION, ORTHOGONAL MATRICES

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Diagonalization. MATH 1502 Calculus II Notes. November 4, 2008

Iterative solvers for linear equations

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Designing Information Devices and Systems I Spring 2016 Official Lecture Notes Note 21

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Algebra Workshops 10 and 11

Linear Algebra. Workbook

Lecture 1: Systems of linear equations and their solutions

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

Spectral Graph Theory Lecture 3. Fundamental Graphs. Daniel A. Spielman September 5, 2018

EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants

Diagonalization of Matrices

FFTs in Graphics and Vision. Homogenous Polynomials and Irreducible Representations

Cheat Sheet for MATH461

Chapter 6: Orthogonality

Dot Products, Transposes, and Orthogonal Projections

Spectral Partitiong in a Stochastic Block Model

Apprentice Program: Linear Algebra

Lecture 2: Eigenvalues and their Uses

I = i 0,

Sparsification by Effective Resistance Sampling

Topics in linear algebra

Exercises * on Principal Component Analysis

MAA704, Perron-Frobenius theory and Markov chains.

4.2. ORTHOGONALITY 161

(x, y) = d(x, y) = x y.

Math 290-2: Linear Algebra & Multivariable Calculus Northwestern University, Lecture Notes

CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

The following definition is fundamental.

Math Matrix Algebra

Functional Analysis Review

1 Linear Algebra Problems

Linear Algebra 2 Spectral Notes

Symmetric and anti symmetric matrices

MATH 583A REVIEW SESSION #1

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Linear Algebra- Final Exam Review

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Mathematical Methods wk 2: Linear Operators

Q N id β. 2. Let I and J be ideals in a commutative ring A. Give a simple description of

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

The Spectral Theorem for normal linear maps

REPRESENTATION THEORY WEEK 7

INNER PRODUCT SPACE. Definition 1

Since G is a compact Lie group, we can apply Schur orthogonality to see that G χ π (g) 2 dg =

CHAPTER 6. Representations of compact groups

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

Math 291-2: Lecture Notes Northwestern University, Winter 2016

Math Lecture 26 : The Properties of Determinants

Specral Graph Theory and its Applications September 7, Lecture 2 L G1,2 = 1 1.

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

Transcription:

Spectral Graph Theory Lecture 23 Quadrature for the Finite Free Convolution Daniel A. Spielman November 30, 205 Disclaimer These notes are not necessarily an accurate representation of what happened in class. The notes written before class say what I think I should say. I sometimes edit the notes after class to make them way what I wish I had said. There may be small mistakes, so I recommend that you check any mathematically precise statement before using it in your own work. These notes were last revised on December, 205. 23. Overview The material in today s lecture comes from [?] and [?]. My goal today is to prove simple analogs of the main quadrature results, and then give some indication of how the other quadrature statements are proved. I will also try to explain what led us to believe that these results should be true. Recall that last lecture we considered the expected characteristic polynomial of a random matrix of the form A + ΠBΠ T, where A and B are symmetric. We do not know a nice expression for this expected polynomial for general A and B. However, we will see that there is a very nice expression when A and B are Laplacian matrices or the adjacency matrices of regular graphs. 23.2 The Finite Free Convolution In Free Probability [?], one studies operations on matrices in a large dimensional limit. These matrices are determined by the moments of their spectrum, and thus the operations are independent of the eigenvectors of the matrices. We consider a finite dimensional analog. For n-dimensional symmetric matrices A and B, we consider the expected characteristic polynomial E χ x (A + QBQ T ), where O(n) is the group of n-by-n orthonormal matrices, and Q is a random orthonormal matrix chosen according to the Haar measure. In case you are not familiar with Haar measure, I ll quickly explain the idea. It captures our most natural idea of a random orthnormal matrix. For example, if A is a Gaussian random symmetric matrix, and V is its matrix of eigenvectors, then V 23-

Lecture 23: November 30, 205 23-2 is a random orthonormal matrix chosen according to Haar measure. Formally, it is the measure that is invariant under group operations, which in this case are multiplication by orthnormal matrices. That is, the Haar measure is the measure under which for every S O(n) and P O(n), S has the same measure as {QP : Q S}. This expected characteristic polynomial does not depend on the eigenvectors of A and B, and thus can be written as a function of the characteristic polynomials of these matrices. To see this, write A = V DV T and B = U C U T where U and V are the orthnormal eigenvectors matrices and C and D are the diagonal matrices of eigenvalues. We have χ x (V DV T + QU C U T Q T ) = χ x (D + V T QU C U T Q T V ) = χ x (D + (V T QU )C (V T QU ) T ). If Q is distributed according to the Haar measure on O(n), then so is V T QU. If p(x) and q(x) are the characteristic polynomials of A and B, then we define their finite free convolution to be the polynomial p(x) n q(x) def = E χ x (A + QBQ T ). In today s lecture, we will establish the following formula for the finite free convolution. Theorem 23.2.. Let Then, p(x) = x n i ( ) i a i and q(x) = p(x) n q(x) = x n k ( ) k k=0 i+j=k x n i ( ) i b i. (n i)!(n j)! n!(n i j)! a ib j. (23.) This convolution was studied by Walsh [?], who proved that when p and q are real rooted, so is p n q. Our interest in the finite free convolution comes from the following theorem, whose proof we will also sketch today. Theorem 23.2.2. Let A and B be symmetric matrices with constant row sums. If A = a and B = b, we may write their characteristic polynomials as χ x (A) = (x a)p(x) and χ x (B) = (x b)q(x). We then have EΠ S n χ x (A + ΠBΠ T ) = (x (a + b))(p(x) n q(x)). We know that is an eigenvector of eigenvalue a + b of A + ΠBΠ T for every permutation matrix Π. Once we work orthogonal to this vector, we discover the finite free convolution. We describe this theorem as a quadrature result, because it obtains an integral over a continuous space as a sum over a finite number of points.

Lecture 23: November 30, 205 23-3 Before going in to the proof of the theorem, I would like to explain why one might think something like this could be true. The first answer is that it was a lucky guess. We hoped that this expectation would have a nice formula. The nicest possible formula would be a bi-linear map: a function that is linear in p when q is held fixed, and vice versa. So, we computed some examples by holding B and q fixed and varying A. We then observed that the coefficients of the resulting expected polynomial are in fact a linear functions of the coefficients of p. Once we knew this, it didn t take too much work to guess the formula. I now describe the main quadrature result we will prove today. Let B(n) be the nth hyperoctahedral group. This is the group of symmetries of the generalized octahedron in n dimensions. It may be described as the set of matrices that can be written in the form DΠ, where D is a diagonal matrix of ± entries and Π is a permutation. It looks like the family of permutation matrices, except that both and are allowed as nonzero entries. B(n) is a subgroup of O(n). Theorem 23.2.3. For all symmetric matrices A and B, E χ x (A + QBQ T ) = E P B(n) χ x (A + PBP T ). We will use this result to prove Theorem 23.2.. The proof of Theorem 23.2.2 is similar to the proof of Theorem 23.2.3. So, we will prove Theorem 23.2.3 and then explain the major differences. 23.3 Quadrature In general, quadrature formulas allow one to evaluate integrals of a family of functions over a fixed continuous domain by summing the values of those functions at a fixed number of points. There is an intimate connection between families of orthogonal polynomials and quadrature formulae that we unfortunately do not have time to discuss. The best known quadrature formula allows us to evalue the integral of a polynomial around the unit circle in the complex plane. For a polynomial p(x) of degree less than n, 2π where ω = e 2πi/n is a primitive nth root of unity. p(e iθ )dθ = n p(ω k ), θ=0 n We may prove this result by establishing it separately for each monomial. For p(x) = x k with k 0, 2π θ=0 p(e iθ )dθ = 2π θ=0 k=0 e iθk dθ = 0. And, for k < n, the corresponding sum is the sum of nth roots of unity distributed symmetrically about the unit circle. So, n ω jk = 0.

Lecture 23: November 30, 205 23-4 We used this fact in the start of the semester when we computed the eigenvectors of the ring graph and observed that all but the dominant are orthogonal to the all-s vector. On the other hand, for p(x) = both the integral and sum are. We will use an alternative approach to quadrature on groups, encapsulted by the following lemma. Lemma 23.3.. For every n and function p(x) = k <n c kx k, and every θ [0, 2π], p(e i(2πj/n+θ) ) = p(e i(2πj/n) ). This identity implies the quadrature formula above, and has the advantage that it can be experimentally confirmed by evaluating both sums for a random θ. Proof. We again evaluate the sums monomial-by-monomial. For p(x) = x k, with k < n, we have (e i(2πj/n+θ) ) k = e iθk (e i(2πj/n) ) k. For k 0, the latter sum is zero. For k = 0, e iθk =. 23.4 Quadrature by Invariance For symmetric matrices A and B, define the function f A,B (Q) = det(a + QBQ T ). We will derive Theorem 23.2.3 from the following theorem. Theorem 23.4.. For all Q O(n), E P B(n) f(p) = E P B(n) f(qp). Proof of Theorem 23.2.3. First, observe that it suffices to consider determinants. For every P B(n), we have det(a + QBQ T ) = f(q) = f(qp). So, E P B(n) f(qp) = f(q). On the other hand, as B(n) is discrete we can reverse the order of integration to obtain f(q) = E P B(n) f(qp) = E P B(n) f(p) = E P B(n) f(p), where the second equality follows from Theorem 23.4..

Lecture 23: November 30, 205 23-5 23.5 Structure of the Orthogonal Group To prove Theorem 23.4., we need to know a little more about the orthogonal group. We divide the orthonormal matrices into two types, those of determinant and those of determinant. The orthonormal matrices of determinant form the special orthogonal group, SO(n), and every matrix in O(n) may be written in the form DQ where Q SO(n) and D is a diagonal matrix in which the first entry is ± and all others are. Every matrix in SO(n) may be expressed as a product of 2-by-2 rotation matrices. That is, for every Q SO(n) there are matrices Q i,j for i < j n so that Q i,j is a rotation in the span of δ i and δ j and so that Q = Q,2 Q,3 Q,n Q 2,3 Q 2,n Q n,n. If you learned the QR-factorization of a matrix, then you learned an algorithm for computing this decomposition. These facts about the structure of O(n) tell us that it suffices to prove Theorem 23.4. for the special cases in which Q = diag(,,,..., ) and when Q is rotation of the plane spanned by δ i and δ j. As the diagonal matrix is contained in B(n), the result is immediate in that case. For simplicity, consider the case i = and j = 2, and let R θ denote the rotation by angle θ in the first two coordinates: cos θ sin θ 0 def R θ = sin θ cos θ 0. 0 0 I n 2 The hyperoctahedral group B(n) contains the matrices R θ for θ {0, π/2, π, 3π/2}. As B(n) is a group, for these θ we know E P B(n) f(p) = E P B(n) f(r θ P), as the set of matrices in the expectations are identical. This identity implies 4 3 E P B(n) f A,B (R 2πj/4 P) = E P B(n) f(p). We will prove the following lemma, and then show it implies Theorem 23.4.. Lemma 23.5.. For every symmetric A and B, and every θ 4 3 f A,B (R θ+2πj/4 ) = 4 This lemma implies that for every Q,2, 3 f A,B (R 2πj/4 ). E P B(n) f(p) = E P B(n) f(q,2 P). This, in turn, implies Theorem 23.4. and thus Theorem 23.2.3. We can use Lemma 23.3. to derive Lemma 23.5. follows from the following.

Lecture 23: November 30, 205 23-6 Lemma 23.5.2. For every symmetric A and B, there exist c 2, c, c 0, c, c 2 so that f A,B (R θ ) = 2 k= 2 c k (e iθ ) k. Proof. We need to express f(r θ ) as a function of e iθ. To this end, recall that cos θ = 2 (eiθ + e iθ ) and sin θ = i 2 (eiθ e iθ ). From these identities, we see that all two-by-two rotation matrices can be simultaneously diagonalized by writing ( ) ( ) cos θ sin θ e iθ 0 = U sin θ cos θ 0 e iθ U, where ( ) U =, i i and we recall that U is the conjugate transpose: ( ) U i =. i Let D θ be the digaonal matrix having its first two entries e iθ and e iθ, and the rest, and let U n be the matrix with U in its upper 2-by-2 block and s on the diagonal beneath. So, Now, examine R θ = U n D θ U n. f A,B (R θ ) = det(a + R θ BR θ ) = det(a + U n D θ U nbu n D θ U n) = det(u nau n + D θ U nbu n D θ ) = det(u nau n D θ + D θ U nbu n ). The term e iθ only appears in the first row and column of this matrix, and the term e iθ only appears in the second row and column. As a determinant can be expressed as a sum of products of matrix entries with one in each row and column, it is immediate that this determinant can be expressed in terms of e kiθ for k 4. As each such product can have at most 2 terms of the form e iθ and at most two of the form e iθ, we have k 2. The difference between Theorem 23.2.3 and Theorem 23.2.2 is that the first involves a sum over the isometries of hyperoctahedron, while the second involves a sum over the symmetries of the regular n-simplex in n dimensions. The proof of the appropriate quadrature theorem for the symmetries of the regular simplex is very similar to the proof we just saw, except that rotations of the plane through δ i and δ j are replaced by rotations of the plane parallel to the affine subspace spanned by triples of vertices of the simplex.

Lecture 23: November 30, 205 23-7 23.6 The Formula To establish the formula in Theorem 23.2., we observe that it suffices to compute the formula for diagonal matrices, and that Theorem 23.2.3 makes this simple. Every matrix in B(n) can be written as a product ΠD where D is a ± diagonal matrix. If B is the diagonal matrix with entries µ,..., µ n, then ΠDBDΠ T = ΠBΠ T, which is the diagonal matrix with entries µ π(),..., µ π(n), where π is the permutation corresponding to Π. Let A be diagonal with entries λ,..., λ n. For a subset S of {,..., n}, define λ S = i S λ i. We then have Let p n q = a i = λ S. S =i x n k ( ) k c k. k=0 We first compute the expected determinant, c n. c n = (λ h + µ n! π(h) ) = n! π h π S =i λ S h:π(h) S As opposed to expanding this out, let s just figure out how often the product λ S µ T appears. We must have T = n S, and then this term appears for each permutation such that π(t ) S =. This happens / ( n i) fraction of the time, giving the formula c n = ( n ) λ S i S =i ( n i j ( n j T =n i µ T = ( n i)a i b n i = µ h. i!(n i)! a i b n i. n! For general c k and i + j = k, we see that λ S and µ T appear whenever µ(t ) is disjoint from S. The probability of this happening is ) (n i)!(n j)!j! (n i)!(n j)! ) = = n!(n i j)!j! n!(n i j)!, and so c k = i+j=k a i b j (n i)!(n j)! n!(n i j)!. 23.7 Question For which discrete subgroups of O(n) does a result like Theorem 23.2.3 hold? Can it hold for a substantially smaller subgroup than the symmetries of the simplex (which has size (n + )! in n dimensions).