Bilinear and quadratic forms

Similar documents
1 Positive definiteness and semidefiniteness

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and

3.2 Real and complex symmetric bilinear forms

Linear Algebra. Chapter 8: Eigenvalues: Further Applications and Computations Section 8.2. Applications to Geometry Proofs of Theorems.

Matrix decompositions

8. Diagonalization.

5 Quiver Representations

13. Forms and polar spaces

Chapter 12 Review Vector. MATH 126 (Section 9.5) Vector and Scalar The University of Kansas 1 / 30

Chapter 2: Linear Independence and Bases

Math Matrix Algebra

Spectral radius, symmetric and positive matrices

β : V V k, (x, y) x yφ

Chapter 1: Systems of Linear Equations and Matrices

A course in low-dimensional geometry

Algebra Workshops 10 and 11

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Chapter 1 Vector Spaces

Solution to Homework 1

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

Linear Algebra Lecture Notes-II

1 Last time: least-squares problems

Quadratic forms. Defn

MATH 112 QUADRATIC AND BILINEAR FORMS NOVEMBER 24, Bilinear forms

Algebra II. Paulius Drungilas and Jonas Jankauskas

Projective Geometry Lecture Notes

MTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics

Quadratic Forms. Marco Schlichting Notes by Florian Bouyer. January 16, 2012

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

Solutions to Final Exam 2011 (Total: 100 pts)

MATH 110: LINEAR ALGEBRA HOMEWORK #2

Abstract Vector Spaces

Linear Algebra and Matrix Inversion

Determining Unitary Equivalence to a 3 3 Complex Symmetric Matrix from the Upper Triangular Form. Jay Daigle Advised by Stephan Garcia

A Do It Yourself Guide to Linear Algebra

MATRICES. a m,1 a m,n A =

First we introduce the sets that are going to serve as the generalizations of the scalars.

Linear Algebra Lecture Notes-I

6 Orthogonal groups. O 2m 1 q. q 2i 1 q 2i. 1 i 1. 1 q 2i 2. O 2m q. q m m 1. 1 q 2i 1 i 1. 1 q 2i. i 1. 2 q 1 q i 1 q i 1. m 1.

Linear Algebra. Paul Yiu. 6D: 2-planes in R 4. Department of Mathematics Florida Atlantic University. Fall 2011

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved

A linear algebra proof of the fundamental theorem of algebra

Honors Algebra II MATH251 Course Notes by Dr. Eyal Goren McGill University Winter 2007

A linear algebra proof of the fundamental theorem of algebra

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

EK102 Linear Algebra PRACTICE PROBLEMS for Final Exam Spring 2016

The following definition is fundamental.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

The geometry of projective space

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS

2. Every linear system with the same number of equations as unknowns has a unique solution.

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

MATH 369 Linear Algebra

SYLLABUS. 1 Linear maps and matrices

HASSE-MINKOWSKI THEOREM

NORMS ON SPACE OF MATRICES

7. Symmetric Matrices and Quadratic Forms

Introduction to linear programming

Basic Concepts in Linear Algebra

2: LINEAR TRANSFORMATIONS AND MATRICES

Operators with numerical range in a closed halfplane

Epilogue: Quivers. Gabriel s Theorem

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Lecture Summaries for Linear Algebra M51A

2 so Q[ 2] is closed under both additive and multiplicative inverses. a 2 2b 2 + b

Elementary linear algebra

Review of Basic Concepts in Linear Algebra

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Matrix operations Linear Algebra with Computer Science Application

From Wikipedia, the free encyclopedia

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

Elementary maths for GMT

Definition 1.2. Let p R n be a point and v R n be a non-zero vector. The line through p in direction v is the set

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

The Multivariate Gaussian Distribution

0.1 Rational Canonical Forms

Conics and their duals

Review of Linear Algebra

SYMBOL EXPLANATION EXAMPLE

Octonions? A non-associative geometric algebra. Benjamin Prather. October 19, Florida State University Department of Mathematics

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Math 321: Linear Algebra

Math 113 Winter 2013 Prof. Church Midterm Solutions

Eigenvalues and Eigenvectors

UMA Putnam Talk LINEAR ALGEBRA TRICKS FOR THE PUTNAM

Numerical Linear Algebra

Linear algebra 2. Yoav Zemel. March 1, 2012

MODEL ANSWERS TO THE FIRST QUIZ. 1. (18pts) (i) Give the definition of a m n matrix. A m n matrix with entries in a field F is a function

Solution. That ϕ W is a linear map W W follows from the definition of subspace. The map ϕ is ϕ(v + W ) = ϕ(v) + W, which is well-defined since

Math 110: Worksheet 3

Econ Slides from Lecture 8

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

9. Numerical linear algebra background

MODEL ANSWERS TO HWK #3

Classication of Quadratic Forms

MAT Linear Algebra Collection of sample exams

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

Transcription:

Bilinear and quadratic forms Zdeněk Dvořák April 8, 015 1 Bilinear forms Definition 1. Let V be a vector space over a field F. A function b : V V F is a bilinear form if b(u + v, w) = b(u, w) + b(v, w) b(u, v + w) = b(u, v) + b(u, w) b(αu, v) = αb(u, v) b(u, αv) = αb(u, v) for all u, v, w V and α F. The bilinear form b is symmetric if b(u, v) = b(v, u) for all u, v V. Remark: b(o, v) = b(v, o) = 0. Examples: b(x, y) = x, y in R n is bilinear and symmetric for any scalar product. b((x 1, y 1 ), (x, y )) = x 1 x + x 1 y + 3y 1 x + 4y 1 y is bilinear, but not symmetric. Definition. Let C = v 1,..., v n be a basis of V and let b be a bilinear form on V. The matrix of b with respect to C is b(v 1, v 1 ) b(v 1, v )... b(v 1, v n ) [b] C = b(v, v 1 ) b(v, v )... b(v, v n ).... b(v n, v 1 ) b(v n, v )... b(v n, v n ) Lemma 1. Let C = v 1,..., v n be a basis of V and let b be a bilinear form on V. For any x, y V, we have b(x, y) = [x] C [b] C [y] T C. 1

Proof. Let [x] C = (α 1,..., α n ) and [y] C = (β 1,..., β n ). We have b(x, y) = b(α 1 v 1 +... + α n v n, β 1 v 1 +... + β n v n ) n n = α i β j b(v i, v j ) = i=1 n j=1 i=1 j=1 n ( α i β j [b]c )i,j = [x] C [b] C [y] T C. Remark: [b] C is the only matrix with this property. A bilinear form b is symmetric if and only if [b] C is a symmetric matrix. Corollary. Let V be a vector space over a field F. Let C = v 1,..., v n be a basis of V. For every n n matrix M over F, there exists a unique bilinear form b : V V F such that b(v i, v j ) = M i,j for 1 i, j n. Proof. Define b(x, y) = [x] C M[y] T C and observe that b is bilinear. No other bilinear form with this property exists, since any bilinear form satisfying the assumptions has matrix M, which by Lemma 1 uniquely determines the values of the bilinear form. Example 1. The( bilinear ) form b((x 1, y 1 ), (x, y )) = x 1 x + x 1 y + 3y 1 x + 1 4y 1 y has matrix with respect to the standard basis; 3 4 b((x 1, y 1 ), (x, y )) = (x 1, y 1 ) ( 1 3 4 ) ( x Lemma 3. Let B = v 1,..., v n and C be two bases of V and let b be a bilinear form on V. Let S = [id] B,C. Then Proof. We have [b] B = S T [b] C S. ( S T [b] C S ) i,j = et i S T [b] C Se j y ). = [v i ] B S T [b] C S[v j ] T B = [v i ] C [b] C [v j ] T C = b(v i, v j ) = ([b] B ) i,j.

Quadratic forms Definition 3. A function f : V F is a quadratic form if there exists a bilinear form b : V V F such that f(x) = b(x, x) for every x V. Example. f((x, y)) = x +5xy+4y is a quadratic form, since f((x, y)) = b((x, y), (x, y)) for the bilinear form b((x 1, y 1 ), (x, y )) = x 1 x + x 1 y + 3y 1 x + 4y 1 y. Also, f((x, y)) = b 1 ((x, y), (x, y)) for the symmetric bilinear form b((x 1, y 1 ), (x, y )) = x 1 x + 5 (x 1y + y 1 x ) + 4y 1 y. Lemma 4. Let V be a vector space over a field F whose characteristic is not. For every quadratic form f, there exists a unique symmetric bilinear form b such that f(x) = b(x, x) for every x V. Proof. Since f is quadratic, there exists a bilinear form b 0 such that f(x) = b 0 (x, x) for every x V. Let b(x, y) = 1 (b 0(x, y) + b 0 (y, x)). Then b is a symmetric bilinear form and b(x, x) = b 0 (x, x) for every x V. Hence, b is a symmetric bilinear form such that f(x) = b(x, x) for every x V. To show that b is unique, it suffices to note that b(x, y) = b(x + y, x + y) b(x, x) b(y, y) = f(x + y) f(x) f(y) whenever b is a symmetric bilinear form b satisfying f(x) = b(x, x) for every x V. Hence, if f is a quadratic form on V and C is a basis of V, then f(x) = [x] C A[x] T C for a unique symmetric matrix A. We write [f] C = A. Also, by Lemma 3, if B is another basis of V and S = [id] B,C, then [f] B = S T [f] C S. Lemma 5. Let f : V F be a quadratic form. Let B 1 = v 1,..., v n and B = w 1,..., w n be two bases of V such that both [f] B1 and [f] B are diagonal matrices. Then [f] B1 and [f] B have the same number of positive entries. Proof. For i {1, }, let a i be the number of positive entries of [f] Bi, and suppose for a contradiction that a 1 > a. Let I = {i : f(v i ) > 0}, J = {i : f(w i ) 0}. Let U 1 = span({v i : i I}) and U = span({w i : i J}). Note 3

that dim(u 1 ) + dim(u ) = a 1 + (dim(v) a ) > dim(v), and thus there exists a non-zero vector v U 1 U. However, f(v) = [v] B1 [f] B1 [v] T B 1 = i I f(v) = [v] B [f] B [v] T B = i J ([v] B1 ) i f(v i) > 0 ([v] B ) i f(w i) 0 This is a contradiction. For integers a, b, and c, let D(a, b, c) be the diagonal matrix with +1 for i = 1,..., a, D i,i = 1 for i = a + 1,..., a + b,. 0 for i = a + b + 1,..., a + b + c. Theorem 6 (Sylvester s law of inertia). If f is a quadratic form on a vector space V over real numbers of finite dimension, then there exist unique integers a, b, and c and a basis B of V such that [f] B = D(a, b, c). Proof. Let A be a matrix of f with respect to any basis C of V. We need to find a regular matrix S (which will serve as the transition matrix from basis B to C) such that S T AS = D(a, b, c) for some a, b, and c. Perform Gaussian elimination, applying the same operations to the rows and columns of A, until we obtain a diagonal matrix with only +1, 1, or 0 on the diagonal. For the uniqueness, suppose that [f] B1 = D(a 1, b 1, c 1 ) and [f] B = D(a, b, c ) for some bases B 1 and B. Let S = [id] B1,B ; hence, D(a 1, b 1, c 1 ) = S T D(a, b, c )S. Since S is regular, we have rank(d(a 1, b 1, c 1 )) = rank(d(a, b, c )), and thus c 1 = c. Also, by Lemma 5, we have a 1 = a, and thus b 1 = b. Definition 4. We say that a quadratic form f on a vector space V of finite dimension has signature (a, b, c) if there exists a basis B such that [f] B = D(a, b, c). 1 1 3 1 Example 3. Let A = 1 7 1 5 3 1 1 1. By performing simultaneous 1 5 1 7 4

row and column operations, we have 1 1 3 1 1 0 3 1 A 0 8 4 4 3 1 1 1 0 8 4 4 3 4 1 1 1 5 1 7 1 4 1 7 1 0 3 1 1 0 0 1 0 8 4 4 0 4 8 4 0 8 4 4 0 4 8 4 1 4 1 7 1 4 4 7 1 0 0 1 1 0 0 0 0 8 4 4 0 4 8 4 0 8 4 4 0 4 8 4 0 4 4 8 0 4 4 8 1 0 0 0 1 0 0 0 0 4 0 4 8 4 0 0 8 4 0 4 4 8 0 4 8 1 0 0 0 1 0 0 0 0 0 0 6 6 0 0 0 0 6 6 0 4 8 0 6 8 1 0 0 0 1 0 0 0 0 0 0 0 6 6 0 0 0 0 0 6 6 0 0 6 6 0 0 6 6 1 0 0 0 1 0 0 0 0 0 0 0 0 6 6 0 0 0 0 0 6 0 0 0 0 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 6 0 0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 0 1 0 0 1 0 0 0 0 1 0 0 0 0 1 0. 0 0 0 0 0 0 0 0 Hence, the signature of A is (, 1, 1). 5

Corollary 7. Let A be a real symmetric n n matrix and let f(x) = x T Ax for x R n. If f has signature (a, b, c), then the sum of algebraic multiplicities of the positive eigenvalues of A is a, and the sum of algebraic multiplicities of the negative eigenvalues of A is b. Proof. Recall that since A is real and symmetric, we have A = Q 1 DQ = Q T DQ for a diagonal matrix D and an orthogonal matrix Q, where A and D have the same eigenvalues, equal to the diagonal entries of D. Equivalently, we have [f] B1 = D, where B 1 is the basis formed by the columns of Q T. Since the signature of A is (a, b, c), there exists a basis B such that [f] B = D(a, b, c). The claim follows by Lemma 5. 3 Quadrics and conics Definition 5. For any n n symmetric real matrix A, a real row vector b and a real number γ, the set {x R n : x T Ax + bx + γ = 0} is called a quadric. If n =, it is called a conic. Example 4. The set {(x, y, z) R 3 : x y z + 1 = 0}: 6

Consider a quadric C = {x R n : x T Ax + bx + γ = 0}. We have A = Q T DQ for an orthogonal matrix Q and a diagonal matrix D. Let b = bq T and C = {y R n : y T Dy + b y + γ = 0}. Observe that x C if and only if Qx C, and thus thus the sets C and C only differ by the isometry described by Q. Let (p, n, z) be the signature of A; without loss of generality, the first p entries of D are positive, the next n entries are negative and the last z are zeros. Furthermore, since C is also equal to {y R n : y T ( D)y b y γ = 0}, we can assume that p n. For any vector d, let b d = d T D + b, γ d = d T Dd + b d + γ and C d = {v R n : v T Dv + b d v + γ d }. Note that v C d if and only if v + d C, and thus C d is obtained from C by shifting it by the vector d (another isometry). Choose the first p + n coordinates of d so that the first p + n coordinates of the vector d T D are equal to the first p+n coordinates of b (the remaining coordinates of d T D are always 0). Furthermore, if z > 0 and at least one of the last z entries of b is not 0, we can choose the last z coordinates of d so that γ d is 0. Thus, we get the following. Lemma 8. Every quadric is up to isometry equal to a quadric {x R n : x T Ax + bx + γ = 0} satisfying the following conditions: A is diagonal with the first p entries positive, the next n entries negative and the last z equal to 0 for some p n, the first p + n entries of b are equal to 0, and either b = o or γ = 0. Example 5. Classification of conics: p = : α 1 x 1 + α x = γ for α 1, α > 0. empty if γ < 0, the point (0, 0) if γ = 0, an ellipse with axes γ/α 1 and γ/α if γ > 0. 7

3 y.5 1.5 1 0.5 x 3.5 1.5 1 0.5 0.5 1 1.5.5 3 0.5 1 1.5.5 3 p = n = 1: α 1 x 1 α x = γ for α 1, α > 0. Two intersecting lines x 1 = α /α 1 x if γ = 0. Hyperbola if γ 0. 3 y.5 1.5 1 0.5 x 3.5 1.5 1 0.5 0.5 1 1.5.5 3 0.5 1 1.5.5 3 p = z = 1, b = o: αx 1 = γ for α > 0. Empty, a line, or two parallel lines depending on γ. p = z = 1, b o, γ = 0: αx 1 = βx for α, β 0. A parabola. 8

3 y.5 1.5 1 0.5 x 3.5 1.5 1 0.5 0.5 1 1.5.5 3 0.5 z =, b = o: γ = 0. Empty or R depending on γ. z =, b o, γ = 0: β 1 x 1 + β x = 0. A line. Similarly, we can classify quadrics in higher dimensions, based on their signatures. 9