A Robust Form of Kruskal s Identifiability Theorem
|
|
- Shannon Wheeler
- 5 years ago
- Views:
Transcription
1 A Robust Form of Kruskal s Identifiability Theorem Aditya Bhaskara (Google NYC) Joint work with Moses Charikar (Princeton University) Aravindan Vijayaraghavan (NYU Northwestern)
2 Background: understanding data Data from various sources: QUESTION ( simple explanation ): can we think of data as being generated from a model with a small number of parameters?
3 Background: understanding data Data from various sources: QUESTION ( simple explanation ): can we think of data as being generated from a model with a small number of parameters? Very successful; e.g. topic models for documents, hidden markov models for speech,...
4 Topic model for documents Large collection of documents Carricature model: each document is about a topic, and each topic is a distribution over n words;
5 Topic model for documents Large collection of documents Carricature model: each document is about a topic, and each topic is a distribution over n words; e.g. aardvark apple ball lion... zebra sports wildlife
6 Topic model for documents Large collection of documents Carricature model: each document is about a topic, and each topic is a distribution over n words; e.g. aardvark apple ball lion... zebra sports wildlife Model parameters: probability that document is on topic i : w i (sum to 1) word probability vector for topic i : p i R n
7 Topic model for documents Generating a document: say r-word document Topic w 1 w 2 w R word p 2 word p 2 word p 2 r words
8 Topic model for documents Generating a document: say r-word document Topic w 1 w 2 w R word p 2 word p 2 word p 2 r words ASSUMPTION: picking a document from corpus at random, randomly sampling r words picking r word document as above
9 Topic model for documents Generating a document: say r-word document Topic w 1 w 2 w R word p 2 word p 2 word p 2 r words ASSUMPTION: picking a document from corpus at random, randomly sampling r words picking r word document as above Goal: find the {w r, p r }
10 What about tensors? Experiment: Pick three random words from random document QUESTION: What is Pr[word 1 = i, word 2 = j, word 3 = k]?
11 What about tensors? Experiment: Pick three random words from random document QUESTION: What is Pr[word 1 = i, word 2 = j, word 3 = k]? Precisely equal to R r=1 w rp r (i)p r (j)p r (k).
12 What about tensors? Experiment: Pick three random words from random document QUESTION: What is Pr[word 1 = i, word 2 = j, word 3 = k]? Precisely equal to R r=1 w rp r (i)p r (j)p r (k). (i, j, k) = R r=1 w r(p r p r p r ). Finding parameters tensor decomposition!
13 Recipe for tensor methods in mixture models Algebraic statistics literature: [Allman, Mathias, Rhodes 09], Estimate a tensor whose decomposition allows reading off the model parameters
14 Recipe for tensor methods in mixture models Algebraic statistics literature: [Allman, Mathias, Rhodes 09], Estimate a tensor whose decomposition allows reading off the model parameters 2. Use tensor decomposition
15 Recipe for tensor methods in mixture models Algebraic statistics literature: [Allman, Mathias, Rhodes 09], Estimate a tensor whose decomposition allows reading off the model parameters 2. Use tensor decomposition Many applications: mixtures of gaussians, hidden markov models, communities, crabs,...
16 Are we done?
17 Two caveats Efficiency We need polynomial time algorithms for decomposition! Given T = R r=1 p r p r p r, can we find {p r } R r=1?
18 Two caveats Efficiency We need polynomial time algorithms for decomposition! Given T = R r=1 p r p r p r, can we find {p r } R r=1? Robustness Algorithm needs to work with noisy tensor In applications, we estimate the tensor from samples. N samples = 1/ N error per entry, typically.
19 Two caveats Efficiency We need polynomial time algorithms for decomposition! Given T = R r=1 p r p r p r, can we find {p r } R r=1? Robustness Algorithm needs to work with noisy tensor In applications, we estimate the tensor from samples. N samples = 1/ N error per entry, typically. GOAL: Given, target accuracy ε, and T = R p r p r p r + N, with N < ε/poly(n), r=1 recover {p r } up to error ε.
20 A success story: the full rank case R Given T = p r p r p r + N, r=1 N < ε/poly(n) Define P : matrix (n R) with columns {p r }.
21 A success story: the full rank case R Given T = p r p r p r + N, r=1 N < ε/poly(n) Define P : matrix (n R) with columns {p r }. Theorem If σ R (P ) > 1/poly (n), then can find {p r } up to error ε in polynomial time.
22 A success story: the full rank case R Given T = p r p r p r + N, r=1 N < ε/poly(n) Define P : matrix (n R) with columns {p r }. Theorem If σ R (P ) > 1/poly (n), then can find {p r } up to error ε in polynomial time. Discovered many times: [Harshman 72], [Leurgans, et al. 93], [Chang 96], [Anandkumar, et al. 10], [Goyal et al. 13],...
23 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n?
24 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n? This talk: easier problem of identifiability. Theorem (Informal theorem.) If the Kruskal rank condition holds, then it is possible to recover the decomposition up to error ε.
25 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n? This talk: easier problem of identifiability. Theorem (Informal theorem.) If the Kruskal rank condition holds, then it is possible to recover the decomposition up to error ε. Not efficient ; open problem to do it efficiently
26 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n? This talk: easier problem of identifiability. Theorem (Informal theorem.) If the Kruskal rank condition holds, then it is possible to recover the decomposition up to error ε. Not efficient ; open problem to do it efficiently Can be done if the components are nearly orthogonal [Anandkumar et al. 14]
27 Kruskal s theorem (1977) R T = a i b i c i r=1 GOAL: Find conditions under which the decomposition unique.
28 Kruskal s theorem (1977) T = R a i b i c i r=1 GOAL: Find conditions under which the decomposition unique. Kruskal rank For matrix A(n R), the rank is the largest integer k(a) s.t. every k(a) columns of A are linearly independent.
29 Kruskal s theorem (1977) T = R a i b i c i r=1 GOAL: Find conditions under which the decomposition unique. Kruskal rank For matrix A(n R), the rank is the largest integer k(a) s.t. every k(a) columns of A are linearly independent. Note: Much stronger than the usual notion of rank; reminiscent of restricted isometry
30 Kruskal s theorem (1977) T = R a i b i c i r=1 GOAL: Find conditions under which the decomposition unique. Kruskal rank For matrix A(n R), the rank is the largest integer k(a) s.t. every k(a) columns of A are linearly independent. Note: Much stronger than the usual notion of rank; reminiscent of restricted isometry Theorem (Kruskal) Suppose T is defined as above, and A, B, C are n R matrices with columns a r, b r, c r, respectively. Then a sufficient condition for uniqueness of decomposition is k(a) + k(b) + k(c) 2R + 2
31 Our result R T = a i b i c i + N r=1 GOAL: Find conditions under which decomposition is unique.
32 Our result T = R a i b i c i + N r=1 GOAL: Find conditions under which decomposition is unique. Robust Kruskal rank For matrix A(n R) and parameter τ > 0, the rank is the largest integer k τ (A) s.t. every n k τ (A) submatrix has condition number < τ. Note: Recall that condition number of B is σ max(b)/σ min (B).
33 Our result T = R a i b i c i + N r=1 GOAL: Find conditions under which decomposition is unique. Robust Kruskal rank For matrix A(n R) and parameter τ > 0, the rank is the largest integer k τ (A) s.t. every n k τ (A) submatrix has condition number < τ. Note: Recall that condition number of B is σ max(b)/σ min (B). Theorem (Rough) Let T be as above, and A, B, C be n R matrices as before. Then the decomposition is robustly unique if k τ (A) + k τ (B) + k τ (C) 2R + 2
34 Our result T T rank k-tensors We show: For any ε > 0, there is an ε = ε/poly(n), such that T = ε T implies the decompositions are ε-close, up to a permutation.
35 Our result T = R a i b i c i ; T = r=1 Theorem (Formal) R a i b i c i Suppose T, T are defined as above, and A, B, C, A, B, C are n R matrices. Further, suppose for some τ > 0, that r=1 k τ (A) + k τ (B) + k τ (C) 2R + 2. Then for any ε > 0, there is an ε = ε/poly(n, τ) such that if T T < ε, then there exist diagonal matrices Γ A, Γ B, Γ C, and a permutation Π such that Γ A Γ B Γ C = I, and A = ε Γ A ΠA, B = ε Γ B ΠB, C = ε Γ C ΠC.
36 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths)
37 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths) Implies that no tensor close to T has rank < R
38 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths) Implies that no tensor close to T has rank < R Naturally generalizes to higher order tensors (similar to De Lathauwer s extension of Kruskal s theorem)
39 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths) Implies that no tensor close to T has rank < R Naturally generalizes to higher order tensors (similar to De Lathauwer s extension of Kruskal s theorem) Main difficulty in proof: handling 1/poly(n) noise (If ε = ε/exp(n), much easier to show that decompositions are ε-close)
40 Proof overview Suppose A, B, C, A, B, C are n R, column lengths ρ, and k τ (A) = k τ (B) = k τ (C) = n ; R = 4n/3. (I.e., any n columns of A, B, C are well conditioned (thus lin.ind.))
41 Proof overview Suppose A, B, C, A, B, C are n R, column lengths ρ, and k τ (A) = k τ (B) = k τ (C) = n ; R = 4n/3. (I.e., any n columns of A, B, C are well conditioned (thus lin.ind.)) 1. Show that A is a scaled permutation of A, B of B, C of C 2. Show that permutations are equal, and scalings multiply to identity
42 Proof overview Suppose A, B, C, A, B, C are n R, column lengths ρ, and k τ (A) = k τ (B) = k τ (C) = n ; R = 4n/3. (I.e., any n columns of A, B, C are well conditioned (thus lin.ind.)) 1. Show that A is a scaled permutation of A, B of B, C of C 2. Show that permutations are equal, and scalings multiply to identity Crux: first part permutation lemma
43 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel)
44 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel) Key idea. Look at the spaces spanned by subsets of columns of A, A. S Definition: A S := span of columns indexed by S [R] A
45 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel) Key idea. Look at the spaces spanned by subsets of columns of A, A. S Definition: A S := span of columns indexed by S [R] A Informal: Suppose for all S of size (n 1), A S contains at least S columns of A. Then the same is true for all S.
46 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel) Key idea. Look at the spaces spanned by subsets of columns of A, A. S Definition: A S := span of columns indexed by S [R] A Informal: Suppose for all S of size (n 1), A S contains at least S columns of A. Then the same is true for all S. (If true for S = 1, then every column of A is parallel to some column of A)
47 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r).
48 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S
49 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S 2. Only n/3 + 1 terms remain on RHS = at most this number must remain on LHS! (by Kruskal condition)
50 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S 2. Only n/3 + 1 terms remain on RHS = at most this number must remain on LHS! (by Kruskal condition) 3. Implies at least (n 1) cols {a r } belong to A S
51 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S 2. Only n/3 + 1 terms remain on RHS = at most this number must remain on LHS! (by Kruskal condition) 3. Implies at least (n 1) cols {a r } belong to A S
52 Downward induction Show for S = (n 2),..., 1:
53 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such..
54 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1)
55 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S
56 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S 1. Say we have: a = r S α ra r + α i a i = r S β ra r + β j a j
57 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S 1. Say we have: a = r S α ra r + α i a i = r S β ra r + β j a j 2. Since a i, a j are not parallel, this means α i = β j = 0
58 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T
59 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family
60 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family
61 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2)
62 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i.
63 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i.
64 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i. These don t intersect for different i (sunflower!)
65 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i. These don t intersect for different i (sunflower!) Thus the #(cols) of A is > (n 3) + 2 (n/3 + 2) > (4n/3)..
66 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i. These don t intersect for different i (sunflower!) Thus the #(cols) of A is > (n 3) + 2 (n/3 + 2) > (4n/3)..
67 Making things robust permutation lemma Sufficient conditions for A, A having approximately same columns (up to permutation)
68 Making things robust permutation lemma Sufficient conditions for A, A having approximately same columns (up to permutation) S Key idea. Look at the spaces spanned by subsets of columns of A, A. A Informal: Suppose for all S of size (n 1), A S contains at least S columns of A, up to error ε. Then the same is true for all S, with error ε. (If true for singletons, then every col of A is ε -close to a scaling of a col of A)
69 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n.
70 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n. Continuing this way, we only get the claim for S = 1 with error ε exp(n) Means the initial error should have been exponentially small..
71 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n. Continuing this way, we only get the claim for S = 1 with error ε exp(n) Means the initial error should have been exponentially small.. The fix More combinatorial invariant (different definition of sets T ) Stronger sunflower argument Better base case
72 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n. Continuing this way, we only get the claim for S = 1 with error ε exp(n) Means the initial error should have been exponentially small.. The fix More combinatorial invariant (different definition of sets T ) Stronger sunflower argument Better base case
73 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n?
74 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13])
75 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13]) (b) Recent work assuming incoherence [Anandkumar, et al. 14]
76 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13]) (b) Recent work assuming incoherence [Anandkumar, et al. 14] Can we show robust uniqueness under more algebraic conditions?
77 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13]) (b) Recent work assuming incoherence [Anandkumar, et al. 14] Can we show robust uniqueness under more algebraic conditions? Can we show that a generic n n n tensor has a stable unique decomposition up to rank n 2 /4?
78 Thank you! Questions?
Orthogonal tensor decomposition
Orthogonal tensor decomposition Daniel Hsu Columbia University Largely based on 2012 arxiv report Tensor decompositions for learning latent variable models, with Anandkumar, Ge, Kakade, and Telgarsky.
More informationTENSOR DECOMPOSITIONS AND THEIR APPLICATIONS
TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY SPEARMAN S HYPOTHESIS Charles Spearman (1904): There are two types of intelligence, educ%ve and reproduc%ve
More informationA concise proof of Kruskal s theorem on tensor decomposition
A concise proof of Kruskal s theorem on tensor decomposition John A. Rhodes 1 Department of Mathematics and Statistics University of Alaska Fairbanks PO Box 756660 Fairbanks, AK 99775 Abstract A theorem
More informationECE 598: Representation Learning: Algorithms and Models Fall 2017
ECE 598: Representation Learning: Algorithms and Models Fall 2017 Lecture 1: Tensor Methods in Machine Learning Lecturer: Pramod Viswanathan Scribe: Bharath V Raghavan, Oct 3, 2017 11 Introduction Tensors
More informationarxiv: v1 [math.ra] 13 Jan 2009
A CONCISE PROOF OF KRUSKAL S THEOREM ON TENSOR DECOMPOSITION arxiv:0901.1796v1 [math.ra] 13 Jan 2009 JOHN A. RHODES Abstract. A theorem of J. Kruskal from 1977, motivated by a latent-class statistical
More informationAnnouncements Wednesday, November 01
Announcements Wednesday, November 01 WeBWorK 3.1, 3.2 are due today at 11:59pm. The quiz on Friday covers 3.1, 3.2. My office is Skiles 244. Rabinoffice hours are Monday, 1 3pm and Tuesday, 9 11am. Section
More informationLecture 2: Linear operators
Lecture 2: Linear operators Rajat Mittal IIT Kanpur The mathematical formulation of Quantum computing requires vector spaces and linear operators So, we need to be comfortable with linear algebra to study
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationAnnouncements Wednesday, November 01
Announcements Wednesday, November 01 WeBWorK 3.1, 3.2 are due today at 11:59pm. The quiz on Friday covers 3.1, 3.2. My office is Skiles 244. Rabinoffice hours are Monday, 1 3pm and Tuesday, 9 11am. Section
More informationNotes on the Matrix-Tree theorem and Cayley s tree enumerator
Notes on the Matrix-Tree theorem and Cayley s tree enumerator 1 Cayley s tree enumerator Recall that the degree of a vertex in a tree (or in any graph) is the number of edges emanating from it We will
More informationTensors: a geometric view Open lecture November 24, 2014 Simons Institute, Berkeley. Giorgio Ottaviani, Università di Firenze
Open lecture November 24, 2014 Simons Institute, Berkeley Plan of the talk Introduction to Tensors A few relevant applications Tensors as multidimensional matrices A (complex) a b c tensor is an element
More informationMath 1553, Introduction to Linear Algebra
Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level
More informationDefinition 2.3. We define addition and multiplication of matrices as follows.
14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row
More informationx y B =. v u Note that the determinant of B is xu + yv = 1. Thus B is invertible, with inverse u y v x On the other hand, d BA = va + ub 2
5. Finitely Generated Modules over a PID We want to give a complete classification of finitely generated modules over a PID. ecall that a finitely generated module is a quotient of n, a free module. Let
More informationProbabilistic construction of t-designs over finite fields
Probabilistic construction of t-designs over finite fields Shachar Lovett (UCSD) Based on joint works with Arman Fazeli (UCSD), Greg Kuperberg (UC Davis), Ron Peled (Tel Aviv) and Alex Vardy (UCSD) Gent
More informationc i r i i=1 r 1 = [1, 2] r 2 = [0, 1] r 3 = [3, 4].
Lecture Notes: Rank of a Matrix Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk 1 Linear Independence Definition 1. Let r 1, r 2,..., r m
More informationMAT 1302B Mathematical Methods II
MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture
More informationLINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS
LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F has characteristic zero. The following are facts (in
More informationFinitely Generated Modules over a PID, I
Finitely Generated Modules over a PID, I A will throughout be a fixed PID. We will develop the structure theory for finitely generated A-modules. Lemma 1 Any submodule M F of a free A-module is itself
More informationLecture 4 : Quest for Structure in Counting Problems
CS6840: Advanced Complexity Theory Jan 10, 2012 Lecture 4 : Quest for Structure in Counting Problems Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K. Theme: Between P and PSPACE. Lecture Plan:Counting problems
More informationREGULI AND APPLICATIONS, ZARANKIEWICZ PROBLEM
REGULI AN APPLICATIONS, ZARANKIEWICZ PROBLEM One of our long-term goals is to pursue some questions of incidence geometry in R 3. We recall one question to direct our focus during this lecture. Question
More information1 Last time: inverses
MATH Linear algebra (Fall 8) Lecture 8 Last time: inverses The following all mean the same thing for a function f : X Y : f is invertible f is one-to-one and onto 3 For each b Y there is exactly one a
More informationList Decoding of Noisy Reed-Muller-like Codes
List Decoding of Noisy Reed-Muller-like Codes Martin J. Strauss University of Michigan Joint work with A. Robert Calderbank (Princeton) Anna C. Gilbert (Michigan) Joel Lepak (Michigan) Euclidean List Decoding
More informationBackground on Chevalley Groups Constructed from a Root System
Background on Chevalley Groups Constructed from a Root System Paul Tokorcheck Department of Mathematics University of California, Santa Cruz 10 October 2011 Abstract In 1955, Claude Chevalley described
More informationChapter 2: Matrix Algebra
Chapter 2: Matrix Algebra (Last Updated: October 12, 2016) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). Write A = 1. Matrix operations [a 1 a n. Then entry
More information/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 10/31/16
60.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 0/3/6 6. Introduction We talked a lot the last lecture about greedy algorithms. While both Prim
More informationFAMILIES OF IRREDUCIBLE REPRESENTATIONS OF S 2 S 3
FAMILIES OF IRREDUCIBLE REPRESENTATIONS OF S S 3 JAY TAYLOR We would like to consider the representation theory of the Weyl group of type B 3, which is isomorphic to the wreath product S S 3 = (S S S )
More informationLinear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra
MTH6140 Linear Algebra II Notes 2 21st October 2010 2 Matrices You have certainly seen matrices before; indeed, we met some in the first chapter of the notes Here we revise matrix algebra, consider row
More informationCS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method
CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method Tim Roughgarden & Gregory Valiant April 15, 015 This lecture began with an extended recap of Lecture 7. Recall that
More informationLecture 4: Codes based on Concatenation
Lecture 4: Codes based on Concatenation Error-Correcting Codes (Spring 206) Rutgers University Swastik Kopparty Scribe: Aditya Potukuchi and Meng-Tsung Tsai Overview In the last lecture, we studied codes
More informationQuantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002
Quantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002 1 QMA - the quantum analog to MA (and NP). Definition 1 QMA. The complexity class QMA is the class of all languages
More informationHigher-order Fourier analysis of F n p and the complexity of systems of linear forms
Higher-order Fourier analysis of F n p and the complexity of systems of linear forms Hamed Hatami School of Computer Science, McGill University, Montréal, Canada hatami@cs.mcgill.ca Shachar Lovett School
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationContainment restrictions
Containment restrictions Tibor Szabó Extremal Combinatorics, FU Berlin, WiSe 207 8 In this chapter we switch from studying constraints on the set operation intersection, to constraints on the set relation
More informationRESEARCH ARTICLE. An extension of the polytope of doubly stochastic matrices
Linear and Multilinear Algebra Vol. 00, No. 00, Month 200x, 1 15 RESEARCH ARTICLE An extension of the polytope of doubly stochastic matrices Richard A. Brualdi a and Geir Dahl b a Department of Mathematics,
More informationTrades in complex Hadamard matrices
Trades in complex Hadamard matrices Padraig Ó Catháin Ian M. Wanless School of Mathematical Sciences, Monash University, VIC 3800, Australia. February 9, 2015 Abstract A trade in a complex Hadamard matrix
More informationMath 18.6, Spring 213 Problem Set #6 April 5, 213 Problem 1 ( 5.2, 4). Identify all the nonzero terms in the big formula for the determinants of the following matrices: 1 1 1 2 A = 1 1 1 1 1 1, B = 3 4
More information1 Algorithms for Permutation Groups
AM 106/206: Applied Algebra Madhu Sudan Lecture Notes 9 October 3, 2016 References: Based on text by Akos Seress on Permutation Group Algorithms. Algorithm due to Sims. 1 Algorithms for Permutation Groups
More informationMath 730 Homework 6. Austin Mohr. October 14, 2009
Math 730 Homework 6 Austin Mohr October 14, 2009 1 Problem 3A2 Proposition 1.1. If A X, then the family τ of all subsets of X which contain A, together with the empty set φ, is a topology on X. Proof.
More informationMATH JORDAN FORM
MATH 53 JORDAN FORM Let A,, A k be square matrices of size n,, n k, respectively with entries in a field F We define the matrix A A k of size n = n + + n k as the block matrix A 0 0 0 0 A 0 0 0 0 A k It
More informationMath 343 Midterm I Fall 2006 sections 002 and 003 Instructor: Scott Glasgow
Math 343 Midterm I Fall 006 sections 00 003 Instructor: Scott Glasgow 1 Assuming A B are invertible matrices of the same size, prove that ( ) 1 1 AB B A 1 = (11) B A 1 1 is the inverse of AB if only if
More informationFundamental theorem of modules over a PID and applications
Fundamental theorem of modules over a PID and applications Travis Schedler, WOMP 2007 September 11, 2007 01 The fundamental theorem of modules over PIDs A PID (Principal Ideal Domain) is an integral domain
More informationComputational Lower Bounds for Statistical Estimation Problems
Computational Lower Bounds for Statistical Estimation Problems Ilias Diakonikolas (USC) (joint with Daniel Kane (UCSD) and Alistair Stewart (USC)) Workshop on Local Algorithms, MIT, June 2018 THIS TALK
More informationAlgorithms for Tensor Decomposition via Numerical Homotopy
Algorithms for Tensor Decomposition via Numerical Homotopy Session on Algebraic Geometry of Tensor Decompositions August 3, 2013 This talk involves joint work and discussions with: Hirotachi Abo Dan Bates
More informationLinear Systems. Math A Bianca Santoro. September 23, 2016
Linear Systems Math A4600 - Bianca Santoro September 3, 06 Goal: Understand how to solve Ax = b. Toy Model: Let s study the following system There are two nice ways of thinking about this system: x + y
More informationLecture 9: Low Rank Approximation
CSE 521: Design and Analysis of Algorithms I Fall 2018 Lecture 9: Low Rank Approximation Lecturer: Shayan Oveis Gharan February 8th Scribe: Jun Qi Disclaimer: These notes have not been subjected to the
More informationCitation Osaka Journal of Mathematics. 43(2)
TitleIrreducible representations of the Author(s) Kosuda, Masashi Citation Osaka Journal of Mathematics. 43(2) Issue 2006-06 Date Text Version publisher URL http://hdl.handle.net/094/0396 DOI Rights Osaka
More informationLinear Algebra, Summer 2011, pt. 2
Linear Algebra, Summer 2, pt. 2 June 8, 2 Contents Inverses. 2 Vector Spaces. 3 2. Examples of vector spaces..................... 3 2.2 The column space......................... 6 2.3 The null space...........................
More informationREVIEW FOR EXAM II. The exam covers sections , the part of 3.7 on Markov chains, and
REVIEW FOR EXAM II The exam covers sections 3.4 3.6, the part of 3.7 on Markov chains, and 4.1 4.3. 1. The LU factorization: An n n matrix A has an LU factorization if A = LU, where L is lower triangular
More informationCS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery
CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery Tim Roughgarden & Gregory Valiant May 3, 2017 Last lecture discussed singular value decomposition (SVD), and we
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationMTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics
MTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics Ulrich Meierfrankenfeld Department of Mathematics Michigan State University East Lansing MI 48824 meier@math.msu.edu
More informationRepresentations of Sp(6,R) and SU(3) carried by homogeneous polynomials
Representations of Sp(6,R) and SU(3) carried by homogeneous polynomials Govindan Rangarajan a) Department of Mathematics and Centre for Theoretical Studies, Indian Institute of Science, Bangalore 560 012,
More informationA Simple Proof of Fiedler's Conjecture Concerning Orthogonal Matrices
University of Wyoming Wyoming Scholars Repository Mathematics Faculty Publications Mathematics 9-1-1997 A Simple Proof of Fiedler's Conjecture Concerning Orthogonal Matrices Bryan L. Shader University
More informationSingular Value Decomposition
Singular Value Decomposition (Com S 477/577 Notes Yan-Bin Jia Sep, 7 Introduction Now comes a highlight of linear algebra. Any real m n matrix can be factored as A = UΣV T where U is an m m orthogonal
More informationDuke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014
Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document
More informationsublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU)
sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU) 0 overview Our Contributions: 1 overview Our Contributions: A near optimal low-rank
More information12x + 18y = 30? ax + by = m
Math 2201, Further Linear Algebra: a practical summary. February, 2009 There are just a few themes that were covered in the course. I. Algebra of integers and polynomials. II. Structure theory of one endomorphism.
More informationQuivers of Period 2. Mariya Sardarli Max Wimberley Heyi Zhu. November 26, 2014
Quivers of Period 2 Mariya Sardarli Max Wimberley Heyi Zhu ovember 26, 2014 Abstract A quiver with vertices labeled from 1,..., n is said to have period 2 if the quiver obtained by mutating at 1 and then
More informationCS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works
CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The
More informationMore on Bracket Algebra
7 More on Bracket Algebra Algebra is generous; she often gives more than is asked of her. D Alembert The last chapter demonstrated, that determinants (and in particular multihomogeneous bracket polynomials)
More informationThe Cyclic Decomposition of a Nilpotent Operator
The Cyclic Decomposition of a Nilpotent Operator 1 Introduction. J.H. Shapiro Suppose T is a linear transformation on a vector space V. Recall Exercise #3 of Chapter 8 of our text, which we restate here
More informationVII Selected Topics. 28 Matrix Operations
VII Selected Topics Matrix Operations Linear Programming Number Theoretic Algorithms Polynomials and the FFT Approximation Algorithms 28 Matrix Operations We focus on how to multiply matrices and solve
More informationChapter 2. Error Correcting Codes. 2.1 Basic Notions
Chapter 2 Error Correcting Codes The identification number schemes we discussed in the previous chapter give us the ability to determine if an error has been made in recording or transmitting information.
More information1. Introduction to commutative rings and fields
1. Introduction to commutative rings and fields Very informally speaking, a commutative ring is a set in which we can add, subtract and multiply elements so that the usual laws hold. A field is a commutative
More informationSingular Value Decomposition
Chapter 5 Singular Value Decomposition We now reach an important Chapter in this course concerned with the Singular Value Decomposition of a matrix A. SVD, as it is commonly referred to, is one of the
More informationMath 205, Summer I, Week 3a (continued): Chapter 4, Sections 5 and 6. Week 3b. Chapter 4, [Sections 7], 8 and 9
Math 205, Summer I, 2016 Week 3a (continued): Chapter 4, Sections 5 and 6. Week 3b Chapter 4, [Sections 7], 8 and 9 4.5 Linear Dependence, Linear Independence 4.6 Bases and Dimension 4.7 Change of Basis,
More information1 Review Session. 1.1 Lecture 2
1 Review Session Note: The following lists give an overview of the material that was covered in the lectures and sections. Your TF will go through these lists. If anything is unclear or you have questions
More informationc 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0
LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =
More informationTangent spaces, normals and extrema
Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent
More informationLecture 1: Systems of linear equations and their solutions
Lecture 1: Systems of linear equations and their solutions Course overview Topics to be covered this semester: Systems of linear equations and Gaussian elimination: Solving linear equations and applications
More informationStochastic Realization of Binary Exchangeable Processes
Stochastic Realization of Binary Exchangeable Processes Lorenzo Finesso and Cecilia Prosdocimi Abstract A discrete time stochastic process is called exchangeable if its n-dimensional distributions are,
More informationFirst we introduce the sets that are going to serve as the generalizations of the scalars.
Contents 1 Fields...................................... 2 2 Vector spaces.................................. 4 3 Matrices..................................... 7 4 Linear systems and matrices..........................
More informationLecture 18: The Rank of a Matrix and Consistency of Linear Systems
Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Winfried Just Department of Mathematics, Ohio University February 28, 218 Review: The linear span Definition Let { v 1, v 2,..., v n }
More informationBALANCING GAUSSIAN VECTORS. 1. Introduction
BALANCING GAUSSIAN VECTORS KEVIN P. COSTELLO Abstract. Let x 1,... x n be independent normally distributed vectors on R d. We determine the distribution function of the minimum norm of the 2 n vectors
More informationTopics in linear algebra
Chapter 6 Topics in linear algebra 6.1 Change of basis I want to remind you of one of the basic ideas in linear algebra: change of basis. Let F be a field, V and W be finite dimensional vector spaces over
More informationA NOTE ON THE JORDAN CANONICAL FORM
A NOTE ON THE JORDAN CANONICAL FORM H. Azad Department of Mathematics and Statistics King Fahd University of Petroleum & Minerals Dhahran, Saudi Arabia hassanaz@kfupm.edu.sa Abstract A proof of the Jordan
More informationMin-Rank Conjecture for Log-Depth Circuits
Min-Rank Conjecture for Log-Depth Circuits Stasys Jukna a,,1, Georg Schnitger b,1 a Institute of Mathematics and Computer Science, Akademijos 4, LT-80663 Vilnius, Lithuania b University of Frankfurt, Institut
More informationMATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:
MATH 51H Section 4 October 16, 2015 1 Continuity Recall what it means for a function between metric spaces to be continuous: Definition. Let (X, d X ), (Y, d Y ) be metric spaces. A function f : X Y is
More informationBy allowing randomization in the verification process, we obtain a class known as MA.
Lecture 2 Tel Aviv University, Spring 2006 Quantum Computation Witness-preserving Amplification of QMA Lecturer: Oded Regev Scribe: N. Aharon In the previous class, we have defined the class QMA, which
More informationAn algebraic perspective on integer sparse recovery
An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:
More informationConditions for Robust Principal Component Analysis
Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and
More informationSemidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5
Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize
More informationIntroduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011
Compressed Sensing Huichao Xue CS3750 Fall 2011 Table of Contents Introduction From News Reports Abstract Definition How it works A review of L 1 norm The Algorithm Backgrounds for underdetermined linear
More informationMODEL ANSWERS TO THE FIRST HOMEWORK
MODEL ANSWERS TO THE FIRST HOMEWORK 1. Chapter 4, 1: 2. Suppose that F is a field and that a and b are in F. Suppose that a b = 0, and that b 0. Let c be the inverse of b. Multiplying the equation above
More informationAlgebraic Methods in Combinatorics
Algebraic Methods in Combinatorics Po-Shen Loh 27 June 2008 1 Warm-up 1. (A result of Bourbaki on finite geometries, from Răzvan) Let X be a finite set, and let F be a family of distinct proper subsets
More informationDeterminants of Partition Matrices
journal of number theory 56, 283297 (1996) article no. 0018 Determinants of Partition Matrices Georg Martin Reinhart Wellesley College Communicated by A. Hildebrand Received February 14, 1994; revised
More informationIDEALS DEFINING UNIONS OF MATRIX SCHUBERT VARIETIES
IDEALS DEFINING UNIONS OF MATRIX SCHUBERT VARIETIES A. S. BERTIGER Abstract. This note computes a Gröbner basis for the ideal defining a union of matrix Schubert varieties. Moreover, the theorem presented
More information12. Perturbed Matrices
MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,
More informationLecture 4: Proof of Shannon s theorem and an explicit code
CSE 533: Error-Correcting Codes (Autumn 006 Lecture 4: Proof of Shannon s theorem and an explicit code October 11, 006 Lecturer: Venkatesan Guruswami Scribe: Atri Rudra 1 Overview Last lecture we stated
More informationCentral Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J
Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J Frank Curtis, John Drew, Chi-Kwong Li, and Daniel Pragel September 25, 2003 Abstract We study central groupoids, central
More informationMODEL ANSWERS TO THE FIRST QUIZ. 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function
MODEL ANSWERS TO THE FIRST QUIZ 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function A: I J F, where I is the set of integers between 1 and m and J is
More informationRepresentations of moderate growth Paul Garrett 1. Constructing norms on groups
(December 31, 2004) Representations of moderate growth Paul Garrett Representations of reductive real Lie groups on Banach spaces, and on the smooth vectors in Banach space representations,
More informationACI-matrices all of whose completions have the same rank
ACI-matrices all of whose completions have the same rank Zejun Huang, Xingzhi Zhan Department of Mathematics East China Normal University Shanghai 200241, China Abstract We characterize the ACI-matrices
More information0.2 Vector spaces. J.A.Beachy 1
J.A.Beachy 1 0.2 Vector spaces I m going to begin this section at a rather basic level, giving the definitions of a field and of a vector space in much that same detail as you would have met them in a
More informationseries. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques..
1 Use computational techniques and algebraic skills essential for success in an academic, personal, or workplace setting. (Computational and Algebraic Skills) MAT 203 MAT 204 MAT 205 MAT 206 Calculus I
More informationLecture 1 and 2: Random Spanning Trees
Recent Advances in Approximation Algorithms Spring 2015 Lecture 1 and 2: Random Spanning Trees Lecturer: Shayan Oveis Gharan March 31st Disclaimer: These notes have not been subjected to the usual scrutiny
More informationSAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra
1.1. Introduction SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that
More informationSupport weight enumerators and coset weight distributions of isodual codes
Support weight enumerators and coset weight distributions of isodual codes Olgica Milenkovic Department of Electrical and Computer Engineering University of Colorado, Boulder March 31, 2003 Abstract In
More informationVolodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang. May 11, 2015
Tensor Factorization via Matrix Factorization Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang Stanford University May 11, 2015 Kuleshov, Chaganty, Liang (Stanford University) Tensor Factorization
More information