Lattices and Hermite normal form

Similar documents
Hermite normal form: Computation and applications

CSE 206A: Lattice Algorithms and Applications Winter The dual lattice. Instructor: Daniele Micciancio

Lecture 2: Lattices and Bases

1: Introduction to Lattices

An Algorithmist s Toolkit Lecture 18

Math 121 Homework 5: Notes on Selected Problems

The Dual Lattice, Integer Linear Systems and Hermite Normal Form

CSE 206A: Lattice Algorithms and Applications Spring Basic Algorithms. Instructor: Daniele Micciancio

Primitive sets in a lattice

CSE 206A: Lattice Algorithms and Applications Spring Basis Reduction. Instructor: Daniele Micciancio

MATH 2050 Assignment 8 Fall [10] 1. Find the determinant by reducing to triangular form for the following matrices.

Third Midterm Exam Name: Practice Problems November 11, Find a basis for the subspace spanned by the following vectors.

1 Shortest Vector Problem

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

(Inv) Computing Invariant Factors Math 683L (Summer 2003)

Solution to Homework 1

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

Calculating determinants for larger matrices

2 b 3 b 4. c c 2 c 3 c 4

x y B =. v u Note that the determinant of B is xu + yv = 1. Thus B is invertible, with inverse u y v x On the other hand, d BA = va + ub 2

Factorization of singular integer matrices

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Diagonalisierung. Eigenwerte, Eigenvektoren, Mathematische Methoden der Physik I. Vorlesungsnotizen zu

The Determinant: a Means to Calculate Volume

MAT Linear Algebra Collection of sample exams

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Gaussian Elimination and Back Substitution

MATH 149, FALL 2008 DISCRETE GEOMETRY LECTURE NOTES

Bricklaying and the Hermite Normal Form

Lattice Basis Reduction and the LLL Algorithm

Math 344 Lecture # Linear Systems

A = , A 32 = n ( 1) i +j a i j det(a i j). (1) j=1

MH1200 Final 2014/2015

CSE 206A: Lattice Algorithms and Applications Spring Minkowski s theorem. Instructor: Daniele Micciancio

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

MODEL ANSWERS TO THE FIRST QUIZ. 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function

Determinants Chapter 3 of Lay

Determinants. Recall that the 2 2 matrix a b c d. is invertible if

Linear Algebra. Chapter Linear Equations

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

MAT 2037 LINEAR ALGEBRA I web:

Review for Exam Find all a for which the following linear system has no solutions, one solution, and infinitely many solutions.

Definition 2.3. We define addition and multiplication of matrices as follows.

Review problems for MA 54, Fall 2004.

Diagonalisierung. Eigenwerte, Eigenvektoren, Mathematische Methoden der Physik I. Vorlesungsnotizen zu

Eigenvalues and Eigenvectors

A proof of the Jordan normal form theorem

Linear Algebra. and

CSL361 Problem set 4: Basic linear algebra

MATRICES. knowledge on matrices Knowledge on matrix operations. Matrix as a tool of solving linear equations with two or three unknowns.

NOTES on LINEAR ALGEBRA 1

CS412: Lecture #17. Mridul Aanjaneya. March 19, 2015

5.6. PSEUDOINVERSES 101. A H w.

2. Every linear system with the same number of equations as unknowns has a unique solution.

Topological Data Analysis - Spring 2018

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

Linear Systems and Matrices

Math 240 Calculus III

Math Linear Algebra Final Exam Review Sheet

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

47-831: Advanced Integer Programming Lecturer: Amitabh Basu Lecture 2 Date: 03/18/2010

Matrix Factorization and Analysis

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

ELEMENTARY LINEAR ALGEBRA

A matrix over a field F is a rectangular array of elements from F. The symbol

Evaluating Determinants by Row Reduction

Determinants. Beifang Chen

II. Determinant Functions

TRISTRAM BOGART AND REKHA R. THOMAS

Lattices and the Shortest Vector Problem

MIT Final Exam Solutions, Spring 2017

Components and change of basis

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

1 Last time: least-squares problems

THE DYNAMICS OF SUCCESSIVE DIFFERENCES OVER Z AND R

Extra Problems for Math 2050 Linear Algebra I

Introduction to Determinants

MATH 2030: ASSIGNMENT 4 SOLUTIONS

DETERMINANTS. , x 2 = a 11b 2 a 21 b 1

MATH 300, Second Exam REVIEW SOLUTIONS. NOTE: You may use a calculator for this exam- You only need something that will perform basic arithmetic.

How to Factor N 1 and N 2 When p 1 = p 2 mod 2 t

(b) If a multiple of one row of A is added to another row to produce B then det(b) =det(a).

MATH 2030: MATRICES. Example 0.2. Q:Define A 1 =, A. 3 4 A: We wish to find c 1, c 2, and c 3 such that. c 1 + c c

MA106 Linear Algebra lecture notes

Matrix Operations: Determinant

MATH 167: APPLIED LINEAR ALGEBRA Chapter 2

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra

CANONICAL FORMS FOR LINEAR TRANSFORMATIONS AND MATRICES. D. Katz

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions

(f + g)(s) = f(s) + g(s) for f, g V, s S (cf)(s) = cf(s) for c F, f V, s S

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

chapter 12 MORE MATRIX ALGEBRA 12.1 Systems of Linear Equations GOALS

Department of Aerospace Engineering AE602 Mathematics for Aerospace Engineers Assignment No. 4

[06.1] Given a 3-by-3 matrix M with integer entries, find A, B integer 3-by-3 matrices with determinant ±1 such that AMB is diagonal.

NAME MATH 304 Examination 2 Page 1

Lecture Summaries for Linear Algebra M51A

A Multiplicative Operation on Matrices with Entries in an Arbitrary Abelian Group

= W z1 + W z2 and W z1 z 2

What is A + B? What is A B? What is AB? What is BA? What is A 2? and B = QUESTION 2. What is the reduced row echelon matrix of A =

Transcription:

Integer Points in Polyhedra Lattices and Hermite normal form Gennady Shmonin February 17, 2009 1 Lattices Let B = { } b 1,b 2,...,b k be a set of linearly independent vectors in n-dimensional Euclidean space R n. The set of the form { k } Λ(B) := λ i b i : λ 1,λ 2,...,λ k Z i=1 is called a lattice with basis B (or generated by B) and k is the dimension of Λ(B). If k = n, we say that the lattice is full-dimensional. We shall mostly concentrate on full-dimensional lattices; otherwise we may apply an appropriate linear transformation and identify span(b) with R k. Yet, the basis can compactly be represented as a square n k-matrix (we also denote it by B) with vectors b 1,b 2,...,b n as columns. Then we can write Λ(B) = { Bx : x Z k}. (1) We shall often interchange between different notation without mentioning this explicitly, but it should not cause any confusion; thus, B is either a set of vectors b 1,b 2,...,b k, or a matrix with columns b 1,b 2,...,b k, depending on the context. Trivially, the vectors b 1,b 2,...,b k are linearly independent if and only if the matrix B has full column rank, i.e., rank(b) = k, and they span R n if and only if B has full row rank. It is natural to ask about sets of the form (1) when vectors in B are not necessarily linearly independent. Is it still a lattice? In other words, are there linearly independent vectors generating the same set Λ(B)? In general, the answer is no : just consider the one-dimensional vectors b 1 = [1 and b 2 = [ 2 and the corresponding set Λ(b 1,b 2 ) = { λ 1 + λ 2 2 : λ1,λ 2 Z }. However, if we restrict ourselves onto rational vectors only, then the answer is yes, as we shall see a little later. We shall also see that lattices are exactly the discrete subgroups of R n (a group Λ is called discrete if there is a neighbourhood of the origin containing no elements from Λ \ {0} clearly, this property does not hold for our example { λ 1 + λ 2 2 : λ1,λ 2 Z } ). The set of integral vectors Z n is a full-dimensional lattice in R n, generated by the unit vectors e 1,e 2,...,e n. But the same lattice can also be generated by the vectors e 1 + e 2,e 2,e 3,...,e n (since the sum λ 1 e 1 +λ 2 e 2, where λ 1,λ 2 Z, can be rewritten as λ 1 (e 1 +e 2 )+(λ 2 λ 1 )e 2, and the coefficients remain integral; for the converse, we rewrite λ 1 (e 1 +e 2 )+λ 2 e 2 as λ 1 e 1 +(λ 1 +λ 2 )e 2 ). Thus, the basis of a lattice is not unique. However, all the bases of a given lattice are equivalent modulo unimodular transformations. Recall that an integral square matrix U is called unimodular if det(u ) = 1. The following simple facts will be needed. Lemma 1. Let U be a unimodular matrix. Then (a) the inverse U 1 is also unimodular; (b) x is an integral vector if and only if Ux is an integral vector. 1

Proof. Since U is an integral matrix, all cofactors of U are integers. It follows that all entries of U 1 are also integers, as det(u ) = 1. Finally, UU 1 = I implies det(u )det(u 1 ) = 1, whence det(u 1 ) = 1. This proves part (a). It is obvious that if x is an integral vector, then Ux is also an integral vector. The converse follows now from part (a), since x = U 1 (Ux). Now, we can establish the connexion between different bases of the same lattice. For the sake of both completeness and consistency, in the following lemma we also consider general-form matrices of full row rank, where do not claim that the set Λ(B) is a lattice Λ(B) is just the group generated by the columns of B, i.e., the set defined as in (1), where the vectors are not necessarily linearly independent. Lemma 2. Let B and B be matrices. If B = B U for some unimodular matrix U, then Λ(B) = Λ(B ). Moreover, if B and B have full column rank and Λ(B) = Λ(B ), then B = B U for some unimodular matrix U. Proof. Let U be a unimodular matrix and suppose that B = B U. Then for every vector y = Bx with x R n, we have y = B (Ux), and x is integral if and only if Ux is integral. This implies Λ(B) = Λ(B ). If B and B have full column rank and Λ(B) = Λ(B ), then B Λ(B ), and therefore, B = B V for some integral matrix V. Similarly, B Λ(B), which implies that B = BW for some integral matrix W. It follows that B = BW V, and since B has full column rank, W V = I. Particularly, V = W 1 and det(v )det(w ) = 1. But V and W are integral matrices, and the last equality is only possible when det(w ) = det(v ) = 1. 2 Hermite normal form The following elementary column operations are particular unimodular transformations of a matrix B = [ b 1,b 2,...,b n : (1) swap two columns in B: b i b j (i j ); (2) multiply a column by 1: b i ( b i ); (3) add an integer multiple of a column to another column: b i b i + αb j (i j, α Z). We explicitly specify appropriate unimodular matrices for each of the elementary column operations. Thus, swapping columns b i and b j in matrix B is equivalent to multiplying B by the unimodular matrix U = [ u kl, which is obtained from the identity matrix by swapping its i -th and j -th columns. Precisely, u kk = 1 for all k i, j, u i j = u j i = 1, and u kl = 0 otherwise. Multiplying a column b i of matrix B by 1 is equivalent to multiplying B with the unimodular matrix U = [ u kl, which is obtained from the identity matrix by multiplying the i -th column by 1; hence, u kk = 1 for all k i, u ii = 1, and u kl = 0 otherwise. Lastly, adding α times column b j to column b i in matrix B is equivalent to multiplying B with the unimodular matrix U, which is obtained from the identity matrix by α times j -th column to the i -th column, i.e., u kk = 1 for all k, u j i = α, and u i j = 0 otherwise. We say that a matrix B of full row rank is in Hermite normal form if it has the form B = [ H 0, where H = [ h i j is a square matrix such that 2

(1) h i j = 0 for i < j (i.e., H is lower-triangular); (2) 0 h i j < h ii for i > j (i.e., H is non-negative and each row has a unique maximum entry, which is on the main diagonal) Particularly, matrix H is non-singular. Now, we show that any matrix can be brought into Hermite normal form by applying an appropriate sequence of elementary column operations. Since the elementary column operations are actually unimodular transformations of a matrix, the group generated by the columns of the matrix is invariant under these operations; in other words, if we had transformed the matrix into a matrix in Hermite normal form, we also proved that this group can be generated by linearly independent vectors, and therefore, is a lattice. Theorem 3 (Existence of Hermite normal form). Each rational matrix of full row rank can be brought into Hermite normal form by a sequence of elementary column operations. Proof. Let B be a rational matrix of full row rank. Without loss of generality, we may assume that B is integral; otherwise, B = 1 δ B, where δ is the least common multiple of all denominators in B, and we proceed with matrix B. We describe an algorithm converting B into a matrix in Hermite normal form. This algorithm constructs a sequence of matrices B 1,B 2,..., where [ Hk 0 B k =, C k where H k is a k k-matrix in Hermite normal form, and the matrix B k+1 is obtained from the matrix B k as follows. Let d 1,d 2,...,d n k be the entries in the first row of D k. By permuting some columns and multiplying some columns by 1, we may assure that they all are non-negative. Moreover, there is at least one non-zero entry, since B has full row rank. If d i > d j are two non-zero entries in the first row of D k, we add di d j D k times the j -th column to the i -th column of D k. All the entries in the first row of D k remain non-negative but their total sum strictly decreases. (In fact, we execute the Euclidean algorithm to compute the greatest common divisor of d i and d j, and therefore, terminate if both d i and d j are integers.) Therefore, by repeating this procedure we end up with exactly one non-zero entry, say d, in the first row of D k, which after swapping the columns is located in the first column. It remains to ensure that all entries in the first row of C k are non-negative and smaller than d. To do so, we add c i d times the (k + 1)-th column to the i -th column in B k, for each i = 1,2,...,k 1 (clearly, this does not affect the entries of H k ). Due to unimodularity of elementary column operations, we can derive the following corollary. Corollary 3a. Let B be a matrix of full row rank. Then there is a unimodular matrix U such that the matrix BU is in Hermite normal form. Corollary 3a implies that every rational lattice has a basis in Hermite normal form. Moreover, if B is a rational matrix of full row rank, then the group generated by B, Λ(B), is a lattice. In the next section we state these facts in a slightly more general form. In fact, the proof of Theorem 3 yields an algorithm to compute Hermite normal form of a matrix. But is this algorithm polynomial? Unfortunately, not yet. The problem is that the entries may grow exponentially during the execution of the algorithm (recall the Gaussian elmination method with cross-multiplication). Later, we shall consider this question once again and modify the algorithm, so that it runs in polynomial time. 3

3 Sublattices As we mentioned before, Theorem 3 shows, as a side-effect, that any group generated by rational vectors is a lattice, i.e., it can be generated by linearly independent vectors. In this section, we state this result in a slightly more general form. Let Λ = Λ(B) be a full-dimensional lattice in R n. It follows directly from Lemma 2 that the value det(b) > 0 is independent of the particular choice of a basis. Thus, we may define the determinant of lattice Λ as det(λ) := det(b). Let Λ = Λ(B ) be another full-dimensional lattice. If Λ Λ, we say that Λ is a sublattice of Λ. In this case, B = BV for some integral matrix V. The value D(Λ,Λ ) := det(v ) = det(b ) det(b) = det(λ ) det(λ) is then a positive integer and is called the index of Λ in Λ. In particular, if B is an integral matrix, then det(λ ), the determinant of lattice Λ, is the index of Λ in Z n. Lemma 4. Let Λ be a sublattice of a lattice Λ (both Λ and Λ are full-dimensional). Then DΛ Λ Λ, where D is the index of Λ in Λ and DΛ denotes the scaled lattice Λ(DB ) = { Dx : x Λ }. Proof. The second inclusion being trivial, we prove that DΛ Λ. Let B and B be bases of Λ and Λ, respectively. It is clear that DB is a basis of DΛ. Since Λ Λ, we have B = BV for some integral matrix V and D = det(v ) > 0. But then B = B V 1, DV 1 is an integral matrix, and therefore, DB Λ. This implies DΛ Λ. Theorem 5 (Bases in Hermite normal form). Let Λ be a sublattice of Λ. (a) For every basis B of Λ, there is a unique basis B of Λ such that B = B H, where H is a matrix in Hermite normal form. (b) For every basis B of Λ, there is a unique basis B of Λ such that B = B H, where H is a matrix in Hermite normal form. Proof. Let B be a basis of lattice Λ and choose any basis B of lattice Λ. Since Λ Λ, we have B = BV for some integral matrix V. By Corollary 3a, there is a unimodular matrix U such that H = V U is in Hermite normal form, whence B U = BV U = B H. By Lemma 2, B U is also a basis of Λ and Part (a) is proved. For Part (b), let B be a basis of lattice Λ and choose an arbitrary basis B of lattice Λ. By Lemma 4, we have DΛ Λ, where D is the index of B in B, and therefore B = 1 D B V for some integral matrix V. Again, there is a unimodular matrix U such that H = 1 V U is in Hermite D normal form. Hence, BU = 1 D B V U = B H is the required basis of B. A particular case of Theorem 5 is that of rational lattices, i.e., lattices generated by rational vectors: they always have a basis in Hermite normal form. 4

Theorem 6. Let Λ be a full-dimensional lattice with basis B R n n and let B = [ b 1,b 2,...,b k be a matrix composed of some vectors b 1,b 2,...,b k Λ that span Rn. Then the set group generated by B, Λ(B ), is a lattice. Proof. Again, B = BV for some integral matrix V Z n k. By Corollary 3a, there is a unimodular matrix U such that H = V U is in Hermite normal form. By Lemma 2, Λ(B ) = Λ(B U ) and B = BV U = B H. But H is in Hermite normal form, and therefore, has only n non-zero columns. The same is therefore true for B U, implying that Λ(B U ) is a lattice. For rational bases, this means that Λ(B) is a lattice whenever B is a rational matrix. 5