A Robust Form of Kruskal s Identifiability Theorem

Size: px
Start display at page:

Download "A Robust Form of Kruskal s Identifiability Theorem"

Transcription

1 A Robust Form of Kruskal s Identifiability Theorem Aditya Bhaskara (Google NYC) Joint work with Moses Charikar (Princeton University) Aravindan Vijayaraghavan (NYU Northwestern)

2 Background: understanding data Data from various sources: QUESTION ( simple explanation ): can we think of data as being generated from a model with a small number of parameters?

3 Background: understanding data Data from various sources: QUESTION ( simple explanation ): can we think of data as being generated from a model with a small number of parameters? Very successful; e.g. topic models for documents, hidden markov models for speech,...

4 Topic model for documents Large collection of documents Carricature model: each document is about a topic, and each topic is a distribution over n words;

5 Topic model for documents Large collection of documents Carricature model: each document is about a topic, and each topic is a distribution over n words; e.g. aardvark apple ball lion... zebra sports wildlife

6 Topic model for documents Large collection of documents Carricature model: each document is about a topic, and each topic is a distribution over n words; e.g. aardvark apple ball lion... zebra sports wildlife Model parameters: probability that document is on topic i : w i (sum to 1) word probability vector for topic i : p i R n

7 Topic model for documents Generating a document: say r-word document Topic w 1 w 2 w R word p 2 word p 2 word p 2 r words

8 Topic model for documents Generating a document: say r-word document Topic w 1 w 2 w R word p 2 word p 2 word p 2 r words ASSUMPTION: picking a document from corpus at random, randomly sampling r words picking r word document as above

9 Topic model for documents Generating a document: say r-word document Topic w 1 w 2 w R word p 2 word p 2 word p 2 r words ASSUMPTION: picking a document from corpus at random, randomly sampling r words picking r word document as above Goal: find the {w r, p r }

10 What about tensors? Experiment: Pick three random words from random document QUESTION: What is Pr[word 1 = i, word 2 = j, word 3 = k]?

11 What about tensors? Experiment: Pick three random words from random document QUESTION: What is Pr[word 1 = i, word 2 = j, word 3 = k]? Precisely equal to R r=1 w rp r (i)p r (j)p r (k).

12 What about tensors? Experiment: Pick three random words from random document QUESTION: What is Pr[word 1 = i, word 2 = j, word 3 = k]? Precisely equal to R r=1 w rp r (i)p r (j)p r (k). (i, j, k) = R r=1 w r(p r p r p r ). Finding parameters tensor decomposition!

13 Recipe for tensor methods in mixture models Algebraic statistics literature: [Allman, Mathias, Rhodes 09], Estimate a tensor whose decomposition allows reading off the model parameters

14 Recipe for tensor methods in mixture models Algebraic statistics literature: [Allman, Mathias, Rhodes 09], Estimate a tensor whose decomposition allows reading off the model parameters 2. Use tensor decomposition

15 Recipe for tensor methods in mixture models Algebraic statistics literature: [Allman, Mathias, Rhodes 09], Estimate a tensor whose decomposition allows reading off the model parameters 2. Use tensor decomposition Many applications: mixtures of gaussians, hidden markov models, communities, crabs,...

16 Are we done?

17 Two caveats Efficiency We need polynomial time algorithms for decomposition! Given T = R r=1 p r p r p r, can we find {p r } R r=1?

18 Two caveats Efficiency We need polynomial time algorithms for decomposition! Given T = R r=1 p r p r p r, can we find {p r } R r=1? Robustness Algorithm needs to work with noisy tensor In applications, we estimate the tensor from samples. N samples = 1/ N error per entry, typically.

19 Two caveats Efficiency We need polynomial time algorithms for decomposition! Given T = R r=1 p r p r p r, can we find {p r } R r=1? Robustness Algorithm needs to work with noisy tensor In applications, we estimate the tensor from samples. N samples = 1/ N error per entry, typically. GOAL: Given, target accuracy ε, and T = R p r p r p r + N, with N < ε/poly(n), r=1 recover {p r } up to error ε.

20 A success story: the full rank case R Given T = p r p r p r + N, r=1 N < ε/poly(n) Define P : matrix (n R) with columns {p r }.

21 A success story: the full rank case R Given T = p r p r p r + N, r=1 N < ε/poly(n) Define P : matrix (n R) with columns {p r }. Theorem If σ R (P ) > 1/poly (n), then can find {p r } up to error ε in polynomial time.

22 A success story: the full rank case R Given T = p r p r p r + N, r=1 N < ε/poly(n) Define P : matrix (n R) with columns {p r }. Theorem If σ R (P ) > 1/poly (n), then can find {p r } up to error ε in polynomial time. Discovered many times: [Harshman 72], [Leurgans, et al. 93], [Chang 96], [Anandkumar, et al. 10], [Goyal et al. 13],...

23 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n?

24 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n? This talk: easier problem of identifiability. Theorem (Informal theorem.) If the Kruskal rank condition holds, then it is possible to recover the decomposition up to error ε.

25 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n? This talk: easier problem of identifiability. Theorem (Informal theorem.) If the Kruskal rank condition holds, then it is possible to recover the decomposition up to error ε. Not efficient ; open problem to do it efficiently

26 Beyond non degeneracy Question: Can we approximately recover parameters under weaker conditions? What about the case R > n? This talk: easier problem of identifiability. Theorem (Informal theorem.) If the Kruskal rank condition holds, then it is possible to recover the decomposition up to error ε. Not efficient ; open problem to do it efficiently Can be done if the components are nearly orthogonal [Anandkumar et al. 14]

27 Kruskal s theorem (1977) R T = a i b i c i r=1 GOAL: Find conditions under which the decomposition unique.

28 Kruskal s theorem (1977) T = R a i b i c i r=1 GOAL: Find conditions under which the decomposition unique. Kruskal rank For matrix A(n R), the rank is the largest integer k(a) s.t. every k(a) columns of A are linearly independent.

29 Kruskal s theorem (1977) T = R a i b i c i r=1 GOAL: Find conditions under which the decomposition unique. Kruskal rank For matrix A(n R), the rank is the largest integer k(a) s.t. every k(a) columns of A are linearly independent. Note: Much stronger than the usual notion of rank; reminiscent of restricted isometry

30 Kruskal s theorem (1977) T = R a i b i c i r=1 GOAL: Find conditions under which the decomposition unique. Kruskal rank For matrix A(n R), the rank is the largest integer k(a) s.t. every k(a) columns of A are linearly independent. Note: Much stronger than the usual notion of rank; reminiscent of restricted isometry Theorem (Kruskal) Suppose T is defined as above, and A, B, C are n R matrices with columns a r, b r, c r, respectively. Then a sufficient condition for uniqueness of decomposition is k(a) + k(b) + k(c) 2R + 2

31 Our result R T = a i b i c i + N r=1 GOAL: Find conditions under which decomposition is unique.

32 Our result T = R a i b i c i + N r=1 GOAL: Find conditions under which decomposition is unique. Robust Kruskal rank For matrix A(n R) and parameter τ > 0, the rank is the largest integer k τ (A) s.t. every n k τ (A) submatrix has condition number < τ. Note: Recall that condition number of B is σ max(b)/σ min (B).

33 Our result T = R a i b i c i + N r=1 GOAL: Find conditions under which decomposition is unique. Robust Kruskal rank For matrix A(n R) and parameter τ > 0, the rank is the largest integer k τ (A) s.t. every n k τ (A) submatrix has condition number < τ. Note: Recall that condition number of B is σ max(b)/σ min (B). Theorem (Rough) Let T be as above, and A, B, C be n R matrices as before. Then the decomposition is robustly unique if k τ (A) + k τ (B) + k τ (C) 2R + 2

34 Our result T T rank k-tensors We show: For any ε > 0, there is an ε = ε/poly(n), such that T = ε T implies the decompositions are ε-close, up to a permutation.

35 Our result T = R a i b i c i ; T = r=1 Theorem (Formal) R a i b i c i Suppose T, T are defined as above, and A, B, C, A, B, C are n R matrices. Further, suppose for some τ > 0, that r=1 k τ (A) + k τ (B) + k τ (C) 2R + 2. Then for any ε > 0, there is an ε = ε/poly(n, τ) such that if T T < ε, then there exist diagonal matrices Γ A, Γ B, Γ C, and a permutation Π such that Γ A Γ B Γ C = I, and A = ε Γ A ΠA, B = ε Γ B ΠB, C = ε Γ C ΠC.

36 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths)

37 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths) Implies that no tensor close to T has rank < R

38 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths) Implies that no tensor close to T has rank < R Naturally generalizes to higher order tensors (similar to De Lathauwer s extension of Kruskal s theorem)

39 Remarks Conditions only about k τ (A), k τ (B), k τ (C); nothing is needed about A, B, C (except an upper bound on column lengths) Implies that no tensor close to T has rank < R Naturally generalizes to higher order tensors (similar to De Lathauwer s extension of Kruskal s theorem) Main difficulty in proof: handling 1/poly(n) noise (If ε = ε/exp(n), much easier to show that decompositions are ε-close)

40 Proof overview Suppose A, B, C, A, B, C are n R, column lengths ρ, and k τ (A) = k τ (B) = k τ (C) = n ; R = 4n/3. (I.e., any n columns of A, B, C are well conditioned (thus lin.ind.))

41 Proof overview Suppose A, B, C, A, B, C are n R, column lengths ρ, and k τ (A) = k τ (B) = k τ (C) = n ; R = 4n/3. (I.e., any n columns of A, B, C are well conditioned (thus lin.ind.)) 1. Show that A is a scaled permutation of A, B of B, C of C 2. Show that permutations are equal, and scalings multiply to identity

42 Proof overview Suppose A, B, C, A, B, C are n R, column lengths ρ, and k τ (A) = k τ (B) = k τ (C) = n ; R = 4n/3. (I.e., any n columns of A, B, C are well conditioned (thus lin.ind.)) 1. Show that A is a scaled permutation of A, B of B, C of C 2. Show that permutations are equal, and scalings multiply to identity Crux: first part permutation lemma

43 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel)

44 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel) Key idea. Look at the spaces spanned by subsets of columns of A, A. S Definition: A S := span of columns indexed by S [R] A

45 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel) Key idea. Look at the spaces spanned by subsets of columns of A, A. S Definition: A S := span of columns indexed by S [R] A Informal: Suppose for all S of size (n 1), A S contains at least S columns of A. Then the same is true for all S.

46 Permutation lemma Sufficient conditions for A, A having same columns (up to permutation) (Suppose for now, that no two cols of A are parallel) Key idea. Look at the spaces spanned by subsets of columns of A, A. S Definition: A S := span of columns indexed by S [R] A Informal: Suppose for all S of size (n 1), A S contains at least S columns of A. Then the same is true for all S. (If true for S = 1, then every column of A is parallel to some column of A)

47 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r).

48 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S

49 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S 2. Only n/3 + 1 terms remain on RHS = at most this number must remain on LHS! (by Kruskal condition)

50 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S 2. Only n/3 + 1 terms remain on RHS = at most this number must remain on LHS! (by Kruskal condition) 3. Implies at least (n 1) cols {a r } belong to A S

51 Downward induction: base case S = (n 1) S A A a 1 b 1 c 1 + a 2 b 2 c 2 + a 1 b 1 c 1 + a 2 b 2 c For any w, 4n/3 r=1 w, ar (br cr) 4n/3 r=1 w, a r (b r c r). 1. Pick w to be orthogonal to A S 2. Only n/3 + 1 terms remain on RHS = at most this number must remain on LHS! (by Kruskal condition) 3. Implies at least (n 1) cols {a r } belong to A S

52 Downward induction Show for S = (n 2),..., 1:

53 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such..

54 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1)

55 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S

56 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S 1. Say we have: a = r S α ra r + α i a i = r S β ra r + β j a j

57 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S 1. Say we have: a = r S α ra r + α i a i = r S β ra r + β j a j 2. Since a i, a j are not parallel, this means α i = β j = 0

58 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T

59 Downward induction Show for S = (n 2),..., 1: start with some S of size (n 2) define S i = S {i}, for various i S (n/3) + 2 such.. let T i be indices of cols of A that lie in A S i ; T i (n 1) Main idea. Any col in T i T j must be contained in A S Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family

60 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family

61 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2)

62 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i.

63 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i.

64 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i. These don t intersect for different i (sunflower!)

65 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i. These don t intersect for different i (sunflower!) Thus the #(cols) of A is > (n 3) + 2 (n/3 + 2) > (4n/3)..

66 Downward induction Thus if T is the set of cols contained in A S, then for any i, j, T i T j = T Thus T i form a sunflower family Claim. Implies that the core has size (n 2) Proof. Basic counting argument. We know T i n 1 for all i. Say T = n 3, then T i \ T 2 for all i. These don t intersect for different i (sunflower!) Thus the #(cols) of A is > (n 3) + 2 (n/3 + 2) > (4n/3)..

67 Making things robust permutation lemma Sufficient conditions for A, A having approximately same columns (up to permutation)

68 Making things robust permutation lemma Sufficient conditions for A, A having approximately same columns (up to permutation) S Key idea. Look at the spaces spanned by subsets of columns of A, A. A Informal: Suppose for all S of size (n 1), A S contains at least S columns of A, up to error ε. Then the same is true for all S, with error ε. (If true for singletons, then every col of A is ε -close to a scaling of a col of A)

69 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n.

70 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n. Continuing this way, we only get the claim for S = 1 with error ε exp(n) Means the initial error should have been exponentially small..

71 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n. Continuing this way, we only get the claim for S = 1 with error ε exp(n) Means the initial error should have been exponentially small.. The fix More combinatorial invariant (different definition of sets T ) Stronger sunflower argument Better base case

72 Issue with inductive proof Suppose we have claim for S = (n 1) with error ε; then can prove claim for S = (n 2) with error only ε n. Continuing this way, we only get the claim for S = 1 with error ε exp(n) Means the initial error should have been exponentially small.. The fix More combinatorial invariant (different definition of sets T ) Stronger sunflower argument Better base case

73 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n?

74 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13])

75 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13]) (b) Recent work assuming incoherence [Anandkumar, et al. 14]

76 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13]) (b) Recent work assuming incoherence [Anandkumar, et al. 14] Can we show robust uniqueness under more algebraic conditions?

77 Directions Are there polynomial time algorithms under Kruskal s conditions? What about other tensor based algorithms in the case R > n? (a) Higher order tensors: [Cardoso 92], next talk (also [Goyal et al. 13]) (b) Recent work assuming incoherence [Anandkumar, et al. 14] Can we show robust uniqueness under more algebraic conditions? Can we show that a generic n n n tensor has a stable unique decomposition up to rank n 2 /4?

78 Thank you! Questions?

Orthogonal tensor decomposition

Orthogonal tensor decomposition Orthogonal tensor decomposition Daniel Hsu Columbia University Largely based on 2012 arxiv report Tensor decompositions for learning latent variable models, with Anandkumar, Ge, Kakade, and Telgarsky.

More information

TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS

TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS TENSOR DECOMPOSITIONS AND THEIR APPLICATIONS ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY SPEARMAN S HYPOTHESIS Charles Spearman (1904): There are two types of intelligence, educ%ve and reproduc%ve

More information

A concise proof of Kruskal s theorem on tensor decomposition

A concise proof of Kruskal s theorem on tensor decomposition A concise proof of Kruskal s theorem on tensor decomposition John A. Rhodes 1 Department of Mathematics and Statistics University of Alaska Fairbanks PO Box 756660 Fairbanks, AK 99775 Abstract A theorem

More information

ECE 598: Representation Learning: Algorithms and Models Fall 2017

ECE 598: Representation Learning: Algorithms and Models Fall 2017 ECE 598: Representation Learning: Algorithms and Models Fall 2017 Lecture 1: Tensor Methods in Machine Learning Lecturer: Pramod Viswanathan Scribe: Bharath V Raghavan, Oct 3, 2017 11 Introduction Tensors

More information

arxiv: v1 [math.ra] 13 Jan 2009

arxiv: v1 [math.ra] 13 Jan 2009 A CONCISE PROOF OF KRUSKAL S THEOREM ON TENSOR DECOMPOSITION arxiv:0901.1796v1 [math.ra] 13 Jan 2009 JOHN A. RHODES Abstract. A theorem of J. Kruskal from 1977, motivated by a latent-class statistical

More information

Announcements Wednesday, November 01

Announcements Wednesday, November 01 Announcements Wednesday, November 01 WeBWorK 3.1, 3.2 are due today at 11:59pm. The quiz on Friday covers 3.1, 3.2. My office is Skiles 244. Rabinoffice hours are Monday, 1 3pm and Tuesday, 9 11am. Section

More information

Lecture 2: Linear operators

Lecture 2: Linear operators Lecture 2: Linear operators Rajat Mittal IIT Kanpur The mathematical formulation of Quantum computing requires vector spaces and linear operators So, we need to be comfortable with linear algebra to study

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Announcements Wednesday, November 01

Announcements Wednesday, November 01 Announcements Wednesday, November 01 WeBWorK 3.1, 3.2 are due today at 11:59pm. The quiz on Friday covers 3.1, 3.2. My office is Skiles 244. Rabinoffice hours are Monday, 1 3pm and Tuesday, 9 11am. Section

More information

Notes on the Matrix-Tree theorem and Cayley s tree enumerator

Notes on the Matrix-Tree theorem and Cayley s tree enumerator Notes on the Matrix-Tree theorem and Cayley s tree enumerator 1 Cayley s tree enumerator Recall that the degree of a vertex in a tree (or in any graph) is the number of edges emanating from it We will

More information

Tensors: a geometric view Open lecture November 24, 2014 Simons Institute, Berkeley. Giorgio Ottaviani, Università di Firenze

Tensors: a geometric view Open lecture November 24, 2014 Simons Institute, Berkeley. Giorgio Ottaviani, Università di Firenze Open lecture November 24, 2014 Simons Institute, Berkeley Plan of the talk Introduction to Tensors A few relevant applications Tensors as multidimensional matrices A (complex) a b c tensor is an element

More information

Math 1553, Introduction to Linear Algebra

Math 1553, Introduction to Linear Algebra Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level

More information

Definition 2.3. We define addition and multiplication of matrices as follows.

Definition 2.3. We define addition and multiplication of matrices as follows. 14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row

More information

x y B =. v u Note that the determinant of B is xu + yv = 1. Thus B is invertible, with inverse u y v x On the other hand, d BA = va + ub 2

x y B =. v u Note that the determinant of B is xu + yv = 1. Thus B is invertible, with inverse u y v x On the other hand, d BA = va + ub 2 5. Finitely Generated Modules over a PID We want to give a complete classification of finitely generated modules over a PID. ecall that a finitely generated module is a quotient of n, a free module. Let

More information

Probabilistic construction of t-designs over finite fields

Probabilistic construction of t-designs over finite fields Probabilistic construction of t-designs over finite fields Shachar Lovett (UCSD) Based on joint works with Arman Fazeli (UCSD), Greg Kuperberg (UC Davis), Ron Peled (Tel Aviv) and Alex Vardy (UCSD) Gent

More information

c i r i i=1 r 1 = [1, 2] r 2 = [0, 1] r 3 = [3, 4].

c i r i i=1 r 1 = [1, 2] r 2 = [0, 1] r 3 = [3, 4]. Lecture Notes: Rank of a Matrix Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk 1 Linear Independence Definition 1. Let r 1, r 2,..., r m

More information

MAT 1302B Mathematical Methods II

MAT 1302B Mathematical Methods II MAT 1302B Mathematical Methods II Alistair Savage Mathematics and Statistics University of Ottawa Winter 2015 Lecture 19 Alistair Savage (uottawa) MAT 1302B Mathematical Methods II Winter 2015 Lecture

More information

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F has characteristic zero. The following are facts (in

More information

Finitely Generated Modules over a PID, I

Finitely Generated Modules over a PID, I Finitely Generated Modules over a PID, I A will throughout be a fixed PID. We will develop the structure theory for finitely generated A-modules. Lemma 1 Any submodule M F of a free A-module is itself

More information

Lecture 4 : Quest for Structure in Counting Problems

Lecture 4 : Quest for Structure in Counting Problems CS6840: Advanced Complexity Theory Jan 10, 2012 Lecture 4 : Quest for Structure in Counting Problems Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K. Theme: Between P and PSPACE. Lecture Plan:Counting problems

More information

REGULI AND APPLICATIONS, ZARANKIEWICZ PROBLEM

REGULI AND APPLICATIONS, ZARANKIEWICZ PROBLEM REGULI AN APPLICATIONS, ZARANKIEWICZ PROBLEM One of our long-term goals is to pursue some questions of incidence geometry in R 3. We recall one question to direct our focus during this lecture. Question

More information

1 Last time: inverses

1 Last time: inverses MATH Linear algebra (Fall 8) Lecture 8 Last time: inverses The following all mean the same thing for a function f : X Y : f is invertible f is one-to-one and onto 3 For each b Y there is exactly one a

More information

List Decoding of Noisy Reed-Muller-like Codes

List Decoding of Noisy Reed-Muller-like Codes List Decoding of Noisy Reed-Muller-like Codes Martin J. Strauss University of Michigan Joint work with A. Robert Calderbank (Princeton) Anna C. Gilbert (Michigan) Joel Lepak (Michigan) Euclidean List Decoding

More information

Background on Chevalley Groups Constructed from a Root System

Background on Chevalley Groups Constructed from a Root System Background on Chevalley Groups Constructed from a Root System Paul Tokorcheck Department of Mathematics University of California, Santa Cruz 10 October 2011 Abstract In 1955, Claude Chevalley described

More information

Chapter 2: Matrix Algebra

Chapter 2: Matrix Algebra Chapter 2: Matrix Algebra (Last Updated: October 12, 2016) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). Write A = 1. Matrix operations [a 1 a n. Then entry

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 10/31/16

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 10/31/16 60.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 0/3/6 6. Introduction We talked a lot the last lecture about greedy algorithms. While both Prim

More information

FAMILIES OF IRREDUCIBLE REPRESENTATIONS OF S 2 S 3

FAMILIES OF IRREDUCIBLE REPRESENTATIONS OF S 2 S 3 FAMILIES OF IRREDUCIBLE REPRESENTATIONS OF S S 3 JAY TAYLOR We would like to consider the representation theory of the Weyl group of type B 3, which is isomorphic to the wreath product S S 3 = (S S S )

More information

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra MTH6140 Linear Algebra II Notes 2 21st October 2010 2 Matrices You have certainly seen matrices before; indeed, we met some in the first chapter of the notes Here we revise matrix algebra, consider row

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method

CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method CS168: The Modern Algorithmic Toolbox Lecture #8: PCA and the Power Iteration Method Tim Roughgarden & Gregory Valiant April 15, 015 This lecture began with an extended recap of Lecture 7. Recall that

More information

Lecture 4: Codes based on Concatenation

Lecture 4: Codes based on Concatenation Lecture 4: Codes based on Concatenation Error-Correcting Codes (Spring 206) Rutgers University Swastik Kopparty Scribe: Aditya Potukuchi and Meng-Tsung Tsai Overview In the last lecture, we studied codes

More information

Quantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002

Quantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002 Quantum NP - Cont. Classical and Quantum Computation A.Yu Kitaev, A. Shen, M. N. Vyalyi 2002 1 QMA - the quantum analog to MA (and NP). Definition 1 QMA. The complexity class QMA is the class of all languages

More information

Higher-order Fourier analysis of F n p and the complexity of systems of linear forms

Higher-order Fourier analysis of F n p and the complexity of systems of linear forms Higher-order Fourier analysis of F n p and the complexity of systems of linear forms Hamed Hatami School of Computer Science, McGill University, Montréal, Canada hatami@cs.mcgill.ca Shachar Lovett School

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Containment restrictions

Containment restrictions Containment restrictions Tibor Szabó Extremal Combinatorics, FU Berlin, WiSe 207 8 In this chapter we switch from studying constraints on the set operation intersection, to constraints on the set relation

More information

RESEARCH ARTICLE. An extension of the polytope of doubly stochastic matrices

RESEARCH ARTICLE. An extension of the polytope of doubly stochastic matrices Linear and Multilinear Algebra Vol. 00, No. 00, Month 200x, 1 15 RESEARCH ARTICLE An extension of the polytope of doubly stochastic matrices Richard A. Brualdi a and Geir Dahl b a Department of Mathematics,

More information

Trades in complex Hadamard matrices

Trades in complex Hadamard matrices Trades in complex Hadamard matrices Padraig Ó Catháin Ian M. Wanless School of Mathematical Sciences, Monash University, VIC 3800, Australia. February 9, 2015 Abstract A trade in a complex Hadamard matrix

More information

Math 18.6, Spring 213 Problem Set #6 April 5, 213 Problem 1 ( 5.2, 4). Identify all the nonzero terms in the big formula for the determinants of the following matrices: 1 1 1 2 A = 1 1 1 1 1 1, B = 3 4

More information

1 Algorithms for Permutation Groups

1 Algorithms for Permutation Groups AM 106/206: Applied Algebra Madhu Sudan Lecture Notes 9 October 3, 2016 References: Based on text by Akos Seress on Permutation Group Algorithms. Algorithm due to Sims. 1 Algorithms for Permutation Groups

More information

Math 730 Homework 6. Austin Mohr. October 14, 2009

Math 730 Homework 6. Austin Mohr. October 14, 2009 Math 730 Homework 6 Austin Mohr October 14, 2009 1 Problem 3A2 Proposition 1.1. If A X, then the family τ of all subsets of X which contain A, together with the empty set φ, is a topology on X. Proof.

More information

MATH JORDAN FORM

MATH JORDAN FORM MATH 53 JORDAN FORM Let A,, A k be square matrices of size n,, n k, respectively with entries in a field F We define the matrix A A k of size n = n + + n k as the block matrix A 0 0 0 0 A 0 0 0 0 A k It

More information

Math 343 Midterm I Fall 2006 sections 002 and 003 Instructor: Scott Glasgow

Math 343 Midterm I Fall 2006 sections 002 and 003 Instructor: Scott Glasgow Math 343 Midterm I Fall 006 sections 00 003 Instructor: Scott Glasgow 1 Assuming A B are invertible matrices of the same size, prove that ( ) 1 1 AB B A 1 = (11) B A 1 1 is the inverse of AB if only if

More information

Fundamental theorem of modules over a PID and applications

Fundamental theorem of modules over a PID and applications Fundamental theorem of modules over a PID and applications Travis Schedler, WOMP 2007 September 11, 2007 01 The fundamental theorem of modules over PIDs A PID (Principal Ideal Domain) is an integral domain

More information

Computational Lower Bounds for Statistical Estimation Problems

Computational Lower Bounds for Statistical Estimation Problems Computational Lower Bounds for Statistical Estimation Problems Ilias Diakonikolas (USC) (joint with Daniel Kane (UCSD) and Alistair Stewart (USC)) Workshop on Local Algorithms, MIT, June 2018 THIS TALK

More information

Algorithms for Tensor Decomposition via Numerical Homotopy

Algorithms for Tensor Decomposition via Numerical Homotopy Algorithms for Tensor Decomposition via Numerical Homotopy Session on Algebraic Geometry of Tensor Decompositions August 3, 2013 This talk involves joint work and discussions with: Hirotachi Abo Dan Bates

More information

Linear Systems. Math A Bianca Santoro. September 23, 2016

Linear Systems. Math A Bianca Santoro. September 23, 2016 Linear Systems Math A4600 - Bianca Santoro September 3, 06 Goal: Understand how to solve Ax = b. Toy Model: Let s study the following system There are two nice ways of thinking about this system: x + y

More information

Lecture 9: Low Rank Approximation

Lecture 9: Low Rank Approximation CSE 521: Design and Analysis of Algorithms I Fall 2018 Lecture 9: Low Rank Approximation Lecturer: Shayan Oveis Gharan February 8th Scribe: Jun Qi Disclaimer: These notes have not been subjected to the

More information

Citation Osaka Journal of Mathematics. 43(2)

Citation Osaka Journal of Mathematics. 43(2) TitleIrreducible representations of the Author(s) Kosuda, Masashi Citation Osaka Journal of Mathematics. 43(2) Issue 2006-06 Date Text Version publisher URL http://hdl.handle.net/094/0396 DOI Rights Osaka

More information

Linear Algebra, Summer 2011, pt. 2

Linear Algebra, Summer 2011, pt. 2 Linear Algebra, Summer 2, pt. 2 June 8, 2 Contents Inverses. 2 Vector Spaces. 3 2. Examples of vector spaces..................... 3 2.2 The column space......................... 6 2.3 The null space...........................

More information

REVIEW FOR EXAM II. The exam covers sections , the part of 3.7 on Markov chains, and

REVIEW FOR EXAM II. The exam covers sections , the part of 3.7 on Markov chains, and REVIEW FOR EXAM II The exam covers sections 3.4 3.6, the part of 3.7 on Markov chains, and 4.1 4.3. 1. The LU factorization: An n n matrix A has an LU factorization if A = LU, where L is lower triangular

More information

CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery

CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery Tim Roughgarden & Gregory Valiant May 3, 2017 Last lecture discussed singular value decomposition (SVD), and we

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

MTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics

MTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics MTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics Ulrich Meierfrankenfeld Department of Mathematics Michigan State University East Lansing MI 48824 meier@math.msu.edu

More information

Representations of Sp(6,R) and SU(3) carried by homogeneous polynomials

Representations of Sp(6,R) and SU(3) carried by homogeneous polynomials Representations of Sp(6,R) and SU(3) carried by homogeneous polynomials Govindan Rangarajan a) Department of Mathematics and Centre for Theoretical Studies, Indian Institute of Science, Bangalore 560 012,

More information

A Simple Proof of Fiedler's Conjecture Concerning Orthogonal Matrices

A Simple Proof of Fiedler's Conjecture Concerning Orthogonal Matrices University of Wyoming Wyoming Scholars Repository Mathematics Faculty Publications Mathematics 9-1-1997 A Simple Proof of Fiedler's Conjecture Concerning Orthogonal Matrices Bryan L. Shader University

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition (Com S 477/577 Notes Yan-Bin Jia Sep, 7 Introduction Now comes a highlight of linear algebra. Any real m n matrix can be factored as A = UΣV T where U is an m m orthogonal

More information

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document

More information

sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU)

sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU) sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU) 0 overview Our Contributions: 1 overview Our Contributions: A near optimal low-rank

More information

12x + 18y = 30? ax + by = m

12x + 18y = 30? ax + by = m Math 2201, Further Linear Algebra: a practical summary. February, 2009 There are just a few themes that were covered in the course. I. Algebra of integers and polynomials. II. Structure theory of one endomorphism.

More information

Quivers of Period 2. Mariya Sardarli Max Wimberley Heyi Zhu. November 26, 2014

Quivers of Period 2. Mariya Sardarli Max Wimberley Heyi Zhu. November 26, 2014 Quivers of Period 2 Mariya Sardarli Max Wimberley Heyi Zhu ovember 26, 2014 Abstract A quiver with vertices labeled from 1,..., n is said to have period 2 if the quiver obtained by mutating at 1 and then

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

More on Bracket Algebra

More on Bracket Algebra 7 More on Bracket Algebra Algebra is generous; she often gives more than is asked of her. D Alembert The last chapter demonstrated, that determinants (and in particular multihomogeneous bracket polynomials)

More information

The Cyclic Decomposition of a Nilpotent Operator

The Cyclic Decomposition of a Nilpotent Operator The Cyclic Decomposition of a Nilpotent Operator 1 Introduction. J.H. Shapiro Suppose T is a linear transformation on a vector space V. Recall Exercise #3 of Chapter 8 of our text, which we restate here

More information

VII Selected Topics. 28 Matrix Operations

VII Selected Topics. 28 Matrix Operations VII Selected Topics Matrix Operations Linear Programming Number Theoretic Algorithms Polynomials and the FFT Approximation Algorithms 28 Matrix Operations We focus on how to multiply matrices and solve

More information

Chapter 2. Error Correcting Codes. 2.1 Basic Notions

Chapter 2. Error Correcting Codes. 2.1 Basic Notions Chapter 2 Error Correcting Codes The identification number schemes we discussed in the previous chapter give us the ability to determine if an error has been made in recording or transmitting information.

More information

1. Introduction to commutative rings and fields

1. Introduction to commutative rings and fields 1. Introduction to commutative rings and fields Very informally speaking, a commutative ring is a set in which we can add, subtract and multiply elements so that the usual laws hold. A field is a commutative

More information

Singular Value Decomposition

Singular Value Decomposition Chapter 5 Singular Value Decomposition We now reach an important Chapter in this course concerned with the Singular Value Decomposition of a matrix A. SVD, as it is commonly referred to, is one of the

More information

Math 205, Summer I, Week 3a (continued): Chapter 4, Sections 5 and 6. Week 3b. Chapter 4, [Sections 7], 8 and 9

Math 205, Summer I, Week 3a (continued): Chapter 4, Sections 5 and 6. Week 3b. Chapter 4, [Sections 7], 8 and 9 Math 205, Summer I, 2016 Week 3a (continued): Chapter 4, Sections 5 and 6. Week 3b Chapter 4, [Sections 7], 8 and 9 4.5 Linear Dependence, Linear Independence 4.6 Bases and Dimension 4.7 Change of Basis,

More information

1 Review Session. 1.1 Lecture 2

1 Review Session. 1.1 Lecture 2 1 Review Session Note: The following lists give an overview of the material that was covered in the lectures and sections. Your TF will go through these lists. If anything is unclear or you have questions

More information

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0

c 1 v 1 + c 2 v 2 = 0 c 1 λ 1 v 1 + c 2 λ 1 v 2 = 0 LECTURE LECTURE 2 0. Distinct eigenvalues I haven t gotten around to stating the following important theorem: Theorem: A matrix with n distinct eigenvalues is diagonalizable. Proof (Sketch) Suppose n =

More information

Tangent spaces, normals and extrema

Tangent spaces, normals and extrema Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent

More information

Lecture 1: Systems of linear equations and their solutions

Lecture 1: Systems of linear equations and their solutions Lecture 1: Systems of linear equations and their solutions Course overview Topics to be covered this semester: Systems of linear equations and Gaussian elimination: Solving linear equations and applications

More information

Stochastic Realization of Binary Exchangeable Processes

Stochastic Realization of Binary Exchangeable Processes Stochastic Realization of Binary Exchangeable Processes Lorenzo Finesso and Cecilia Prosdocimi Abstract A discrete time stochastic process is called exchangeable if its n-dimensional distributions are,

More information

First we introduce the sets that are going to serve as the generalizations of the scalars.

First we introduce the sets that are going to serve as the generalizations of the scalars. Contents 1 Fields...................................... 2 2 Vector spaces.................................. 4 3 Matrices..................................... 7 4 Linear systems and matrices..........................

More information

Lecture 18: The Rank of a Matrix and Consistency of Linear Systems

Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Winfried Just Department of Mathematics, Ohio University February 28, 218 Review: The linear span Definition Let { v 1, v 2,..., v n }

More information

BALANCING GAUSSIAN VECTORS. 1. Introduction

BALANCING GAUSSIAN VECTORS. 1. Introduction BALANCING GAUSSIAN VECTORS KEVIN P. COSTELLO Abstract. Let x 1,... x n be independent normally distributed vectors on R d. We determine the distribution function of the minimum norm of the 2 n vectors

More information

Topics in linear algebra

Topics in linear algebra Chapter 6 Topics in linear algebra 6.1 Change of basis I want to remind you of one of the basic ideas in linear algebra: change of basis. Let F be a field, V and W be finite dimensional vector spaces over

More information

A NOTE ON THE JORDAN CANONICAL FORM

A NOTE ON THE JORDAN CANONICAL FORM A NOTE ON THE JORDAN CANONICAL FORM H. Azad Department of Mathematics and Statistics King Fahd University of Petroleum & Minerals Dhahran, Saudi Arabia hassanaz@kfupm.edu.sa Abstract A proof of the Jordan

More information

Min-Rank Conjecture for Log-Depth Circuits

Min-Rank Conjecture for Log-Depth Circuits Min-Rank Conjecture for Log-Depth Circuits Stasys Jukna a,,1, Georg Schnitger b,1 a Institute of Mathematics and Computer Science, Akademijos 4, LT-80663 Vilnius, Lithuania b University of Frankfurt, Institut

More information

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous: MATH 51H Section 4 October 16, 2015 1 Continuity Recall what it means for a function between metric spaces to be continuous: Definition. Let (X, d X ), (Y, d Y ) be metric spaces. A function f : X Y is

More information

By allowing randomization in the verification process, we obtain a class known as MA.

By allowing randomization in the verification process, we obtain a class known as MA. Lecture 2 Tel Aviv University, Spring 2006 Quantum Computation Witness-preserving Amplification of QMA Lecturer: Oded Regev Scribe: N. Aharon In the previous class, we have defined the class QMA, which

More information

An algebraic perspective on integer sparse recovery

An algebraic perspective on integer sparse recovery An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:

More information

Conditions for Robust Principal Component Analysis

Conditions for Robust Principal Component Analysis Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize

More information

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011 Compressed Sensing Huichao Xue CS3750 Fall 2011 Table of Contents Introduction From News Reports Abstract Definition How it works A review of L 1 norm The Algorithm Backgrounds for underdetermined linear

More information

MODEL ANSWERS TO THE FIRST HOMEWORK

MODEL ANSWERS TO THE FIRST HOMEWORK MODEL ANSWERS TO THE FIRST HOMEWORK 1. Chapter 4, 1: 2. Suppose that F is a field and that a and b are in F. Suppose that a b = 0, and that b 0. Let c be the inverse of b. Multiplying the equation above

More information

Algebraic Methods in Combinatorics

Algebraic Methods in Combinatorics Algebraic Methods in Combinatorics Po-Shen Loh 27 June 2008 1 Warm-up 1. (A result of Bourbaki on finite geometries, from Răzvan) Let X be a finite set, and let F be a family of distinct proper subsets

More information

Determinants of Partition Matrices

Determinants of Partition Matrices journal of number theory 56, 283297 (1996) article no. 0018 Determinants of Partition Matrices Georg Martin Reinhart Wellesley College Communicated by A. Hildebrand Received February 14, 1994; revised

More information

IDEALS DEFINING UNIONS OF MATRIX SCHUBERT VARIETIES

IDEALS DEFINING UNIONS OF MATRIX SCHUBERT VARIETIES IDEALS DEFINING UNIONS OF MATRIX SCHUBERT VARIETIES A. S. BERTIGER Abstract. This note computes a Gröbner basis for the ideal defining a union of matrix Schubert varieties. Moreover, the theorem presented

More information

12. Perturbed Matrices

12. Perturbed Matrices MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,

More information

Lecture 4: Proof of Shannon s theorem and an explicit code

Lecture 4: Proof of Shannon s theorem and an explicit code CSE 533: Error-Correcting Codes (Autumn 006 Lecture 4: Proof of Shannon s theorem and an explicit code October 11, 006 Lecturer: Venkatesan Guruswami Scribe: Atri Rudra 1 Overview Last lecture we stated

More information

Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J

Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J Frank Curtis, John Drew, Chi-Kwong Li, and Daniel Pragel September 25, 2003 Abstract We study central groupoids, central

More information

MODEL ANSWERS TO THE FIRST QUIZ. 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function

MODEL ANSWERS TO THE FIRST QUIZ. 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function MODEL ANSWERS TO THE FIRST QUIZ 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function A: I J F, where I is the set of integers between 1 and m and J is

More information

Representations of moderate growth Paul Garrett 1. Constructing norms on groups

Representations of moderate growth Paul Garrett 1. Constructing norms on groups (December 31, 2004) Representations of moderate growth Paul Garrett Representations of reductive real Lie groups on Banach spaces, and on the smooth vectors in Banach space representations,

More information

ACI-matrices all of whose completions have the same rank

ACI-matrices all of whose completions have the same rank ACI-matrices all of whose completions have the same rank Zejun Huang, Xingzhi Zhan Department of Mathematics East China Normal University Shanghai 200241, China Abstract We characterize the ACI-matrices

More information

0.2 Vector spaces. J.A.Beachy 1

0.2 Vector spaces. J.A.Beachy 1 J.A.Beachy 1 0.2 Vector spaces I m going to begin this section at a rather basic level, giving the definitions of a field and of a vector space in much that same detail as you would have met them in a

More information

series. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques..

series. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques.. 1 Use computational techniques and algebraic skills essential for success in an academic, personal, or workplace setting. (Computational and Algebraic Skills) MAT 203 MAT 204 MAT 205 MAT 206 Calculus I

More information

Lecture 1 and 2: Random Spanning Trees

Lecture 1 and 2: Random Spanning Trees Recent Advances in Approximation Algorithms Spring 2015 Lecture 1 and 2: Random Spanning Trees Lecturer: Shayan Oveis Gharan March 31st Disclaimer: These notes have not been subjected to the usual scrutiny

More information

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra 1.1. Introduction SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that

More information

Support weight enumerators and coset weight distributions of isodual codes

Support weight enumerators and coset weight distributions of isodual codes Support weight enumerators and coset weight distributions of isodual codes Olgica Milenkovic Department of Electrical and Computer Engineering University of Colorado, Boulder March 31, 2003 Abstract In

More information

Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang. May 11, 2015

Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang. May 11, 2015 Tensor Factorization via Matrix Factorization Volodymyr Kuleshov ú Arun Tejasvi Chaganty ú Percy Liang Stanford University May 11, 2015 Kuleshov, Chaganty, Liang (Stanford University) Tensor Factorization

More information