Lecture 7: Vectors and Matrices II Introduction to Matrices (See Sections, 3.3, 3.6, 3.7 and 3.9 in Boas)

Similar documents
,, rectilinear,, spherical,, cylindrical. (6.1)

Lecture 10: A (Brief) Introduction to Group Theory (See Chapter 3.13 in Boas, 3rd Edition)

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Phys 201. Matrices and Determinants

Mathematics. EC / EE / IN / ME / CE. for

MATRICES The numbers or letters in any given matrix are called its entries or elements

Introduction to Group Theory

Review of Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

[ Here 21 is the dot product of (3, 1, 2, 5) with (2, 3, 1, 2), and 31 is the dot product of

Matrices and Determinants

Chapter 7. Linear Algebra: Matrices, Vectors,

Foundations of Matrix Analysis

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Introduction to Matrices

Introduction to Determinants

Math 360 Linear Algebra Fall Class Notes. a a a a a a. a a a

Matrices Gaussian elimination Determinants. Graphics 2009/2010, period 1. Lecture 4: matrices

. =. a i1 x 1 + a i2 x 2 + a in x n = b i. a 11 a 12 a 1n a 21 a 22 a 1n. i1 a i2 a in

7.3. Determinants. Introduction. Prerequisites. Learning Outcomes

n n matrices The system of m linear equations in n variables x 1, x 2,..., x n can be written as a matrix equation by Ax = b, or in full

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

a11 a A = : a 21 a 22

Introduction to Matrices

Matrices. Chapter Definitions and Notations

Fundamentals of Engineering Analysis (650163)

Finite Mathematics Chapter 2. where a, b, c, d, h, and k are real numbers and neither a and b nor c and d are both zero.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Chemistry 5.76 Revised February, 1982 NOTES ON MATRIX METHODS

Quantum Computing Lecture 2. Review of Linear Algebra

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of

Matrices. Background mathematics review

QM and Angular Momentum

10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections )

Elementary maths for GMT

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

ENGI 9420 Lecture Notes 2 - Matrix Algebra Page Matrix operations can render the solution of a linear system much more efficient.

M. Matrices and Linear Algebra

A Review of Matrix Analysis

Matrix Arithmetic. j=1

Matrices BUSINESS MATHEMATICS

Materials engineering Collage \\ Ceramic & construction materials department Numerical Analysis \\Third stage by \\ Dalya Hekmat

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

MAC Learning Objectives. Learning Objectives (Cont.) Module 10 System of Equations and Inequalities II

Matrix Representation

Matrix Algebra Determinant, Inverse matrix. Matrices. A. Fabretti. Mathematics 2 A.Y. 2015/2016. A. Fabretti Matrices

Physics 557 Lecture 5

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02)

The word Matrices is the plural of the word Matrix. A matrix is a rectangular arrangement (or array) of numbers called elements.

Mathematics SKE: STRAND J. UNIT J4 Matrices: Introduction

Review from Bootcamp: Linear Algebra

7.4. The Inverse of a Matrix. Introduction. Prerequisites. Learning Outcomes

Lecture 6: Geometry of OLS Estimation of Linear Regession

A matrix over a field F is a rectangular array of elements from F. The symbol

1 Matrices and Systems of Linear Equations

Chapter 2 Notes, Linear Algebra 5e Lay

Linear Algebra March 16, 2019

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Review of Vectors and Matrices

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

MATH 106 LINEAR ALGEBRA LECTURE NOTES

Lesson U2.1 Study Guide

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Numerical Analysis Lecture Notes

A = 3 B = A 1 1 matrix is the same as a number or scalar, 3 = [3].

A VERY BRIEF LINEAR ALGEBRA REVIEW for MAP 5485 Introduction to Mathematical Biophysics Fall 2010

Exercise Set Suppose that A, B, C, D, and E are matrices with the following sizes: A B C D E

Section 9.2: Matrices.. a m1 a m2 a mn

A primer on matrices

MATRICES. a m,1 a m,n A =

Section 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns.

We use the overhead arrow to denote a column vector, i.e., a number with a direction. For example, in three-space, we write

PHYS 705: Classical Mechanics. Rigid Body Motion Introduction + Math Review

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

Multiplying matrices by diagonal matrices is faster than usual matrix multiplication.

Matrices A brief introduction

Elementary Row Operations on Matrices

Linear Algebra: Matrix Eigenvalue Problems

STRAND J: TRANSFORMATIONS, VECTORS and MATRICES

Symmetric and anti symmetric matrices

Think about systems of linear equations, their solutions, and how they might be represented with matrices.

MATH Mathematics for Agriculture II

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Introduction to Matrices and Linear Systems Ch. 3

1 Matrices and Systems of Linear Equations

1 Determinants. 1.1 Determinant

1 Matrices and matrix algebra

Linear Algebra. Min Yan

Topic 1: Matrix diagonalization

Appendix C Vector and matrix algebra

Prepared by: M. S. KumarSwamy, TGT(Maths) Page

CS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:

Lecture 10: Determinants and Cramer s Rule

MATH 2050 Assignment 8 Fall [10] 1. Find the determinant by reducing to triangular form for the following matrices.

Matrices A brief introduction

Transcription:

Lecture 7: Vectors and Matrices II Introduction to Matrices (See Sections 3.3 3.6 3.7 and 3.9 in Boas) Here we will continue our discussion of vectors and their transformations. In Lecture 6 we gained some facility with vectors (-tuples) and we now want to discuss generalized operations on vectors a very common activity in physics. In particular we will focus on linear functions of vectors which are themselves sometimes vectors. Consider first a general (scalar valued) function of a vector f r. Recall that a function is a linear function if and only if the following properties hold: f r r2 f r f r2 (7.) f ar af r where a is a scalar (i.e. an ordinary constant). Examples of scalar valued functions 2 2 2 of a vector are f r r xˆ x and f2 r r x y z. Comparing to Eq. (7.) we see that f is a linear function and f 2 is not! A similar definition of linearity holds for vector valued functions of vectors as in the following F r r2 F r F r2 (7.2) F ar af r. A trivial example of a linear vector valued function is constant. F r ar multiplication by a ASIDE: he definition of linearity in Eqs. (7.) and (7.2) is the standard definition for a linear operator (e.g. d dt ). In general we have A B A B aa a A. his line of reasoning (linear vector valued functions of vectors) leads us directly to the idea of matrices. As in Lecture 6 we want to connect the idea of vector valued linear functions of general vectors to explicit (concrete) expressions for this operation in terms of components defined by an explicit set of basis vectors v k. k Physics 227 Lecture 7 Autumn 2008

Expressing both F and r in terms of components we have (introducing some definitions; note we are assuming that Fr and r are in the same vector space and therefore can be expressed in terms of the same basis vectors) F r F r v R R v F r R r k k k k k k k k r v k k k F r F r v r F v k k k k k k (7.3) where the last step follows due to the linearity of F. Since Fv k is a vector we must be able to express it in terms of its components in the given basis set. Further these components are important and we should give them a name F v F v v M v. k j k j jk j j j (7.4) he x array of numbers F v M j k jk (note the order of the indices) is the matrix mentioned earlier. Substituting back in Eq. (7.3) (and changing the order of the sums) we have F r r F v r M v M r v R v R k k k jk j jk k j j j k k j j k j j k M r. jk k (7.5) he last step follows from the uniqueness of the components which is seen most directly if we choose our basis vectors to be orthogonal (but this is not necessary). We think of M as a square array of numbers where j defines the row and k the jk column of a specific entry. hus given a specific basis set a linear vector valued function of vectors is defined by the corresponding matrix M (just like a vector is jk Physics 227 Lecture 7 2 Autumn 2008

defined by its components in the chosen basis). If we think of the set of components of a vector as a x matrix then Eq. (7.5) illustrates matrix multiplication the elements in one row of the left-hand matrix times (one-at-a-time) the elements in a column of the right-hand matrix. Looking back to Lecture 6 (recall Eqs.(6.2) and (6.3)) we see that the scalar product of 2 vectors (in a Euclidean geometry) is just the usual matrix multiplication of an x matrix (the transpose of the vector) times an x matrix r r r r r r 2 2 k 2 k k (7.6) (or in the complex case r r r r r r r r 2 2 2 2 where the dagger means the complex conjugate transpose or Hermitian conjugate). We should also note that Eq. (7.5) has the usual structure of simultaneous linear equations for unknowns in the case we know M jk and the R j and we want to solve for the r k. We will return to this subject in the next Lecture. Clearly matrices are very useful in physics and we want to learn about their properties which will teach us about linear operators simultaneous linear equations and Group heory (see Section 3.3). In the following discussion we will use capital Roman letters to represent linear operators and their abstract matrices. Just as in our discussion of vectors we will be able to define abstract relations independent of any basis vector set and concrete relations that involve the explicit components of vectors and operators (matrices) with respect to specific basis vectors. R Mr abstract R M r M r k l l l concrete where in the very last expression we have employed the (standard) convention of summing over repeated indices. An interesting question is what kind of operations can we apply to the matrices themselves? Clearly we can add (or subtract) matrices (i.e. add linear operators) (7.7) Physics 227 Lecture 7 3 Autumn 2008

A B r Ar Br X X A B r k l l A B A B. (7.8) Adding matrices just means adding the corresponding elements. Addition of matrices is associative and commutative A B C A B C A B B A. Also we can multiply matrices by constants (scalars) which is commutative associative and distributive A r Ar X X k l A r l A A A A B A B A A A A 2 2 2 2. (7.9) (7.0) Finally we can define the multiplication of matrices as we did earlier as long as the matrices are conformable i.e. as long as the numbers of rows and columns match. We will generally be considering square matrices x of the same dimension so this point will not be an issue (i.e. our linear operation on a vector produces a vector in the same vector space). We have Physics 227 Lecture 7 4 Autumn 2008

AB r A Br X X k lm m lm km A B r AB A B A B l lm lm. (7.) Matrix multiplication is associative distributive but not generally commutative (and this last point is really important in physics) ABC A BC AB C A B C AB AC AB BA except in special cases. (7.2) he difference implied by this last expression AB BA is so important that we give it a name the commutator and a symbol (square brackets) AB BA A B C (7.3) where the commutator has the form of another matrix C. (ote that the commutator only makes sense when A and B are square matrices. Otherwise even if AB and BA both exist they will be of different dimensions.) As we will see the special case of commuting matrices AB BA (AB 0) also plays an important role in physics. (ransformations associated with commuting matrices are called Abelian transformations while non-commuting transformations (matrices) are labeled non- Abelian.) o complete our introduction to matrices we need 2 special matrices (in analogy to our discussion of vectors). he first is the null matrix (all elements zero) 0 0 0r 0 all vectors r 0 0 0 (7.4) Physics 227 Lecture 7 5 Autumn 2008

and the second is the unit matrix that we met first in Lecture 6 0 r r all vectors r 0 (7.5). ote also that the equality of 2 matrices A B means that Ar Br for all r. In component notation two matrices are equal if and only if every element is identical A B. While still focusing on square matrices a special question (essential in the application of matrices to simultaneous equations) is whether given a matrix A there exists a second matrix B such that AB BA B A (7.6) i.e. we say that B is the inverse matrix to A. Further if exists we say that A is a non-singular matrix (see below). In operator language A represents the inverse of the operation represented by A. For example if A rotates vectors clockwise through some angle about some axis then A rotates them counterclockwise through the same size angle and about the same axis. ASIDE: Even if the matrix A is not square but rectangular 2 2 we can still consider the question of finding an inverse but in this case we must distinguish the left-inverse A A from the right-inverse AA since L AL AR (they necessarily have different dimensions if they exist i.e. they satisfy different conformality constraints and the 2 unit matrices in these equations have different dimensions). his question will not be an issue in this course and we will generally assume that only square matrices can be non-singular. o be able to work with matrices it is useful to define 2 scalar quantities related to each matrix A i.e. these quantities characterize the matrix and are invariant under the usual transformations of the matrix (i.e. scalar means invariant). hey play a role A R Physics 227 Lecture 7 6 Autumn 2008

similar to that played by the length of a vector r. he first quantity is the trace of the matrix written as A r A (7.7) i.e. the trace is the sum of the diagonal elements. he second scalar quantity is the determinant which we already assumed in our discussion of the vector product of 2 vectors in Lecture 6. For a 2x2 matrix we have k kk a b a b A det A ad bc. c d c d (7.8) o analyze larger matrices we first define a couple more quantities. For the moment just accept the following definitions. heir intrinsic value will become clear shortly. Consider an x matrix A with elements A. For specific values of the indices k and l construct a new dimensional matrix by eliminating the k th row and the l th column. he determinant of this smaller matrix is called the minor M of the original element A. Since the minor yields a number for each value of k and l these numbers can be thought of as defining another x matrix. ext we multiply the minor by a power of minus defined by and construct the so-called cofactor C of A which is itself an element of a x matrix take away the kth row C M det A. take away the lth column (7.9) For example as applied to a specific 3x3 matrix we have a b c 3 b c A d e f C2 bj ch. h j g h j (7.20) Physics 227 Lecture 7 7 Autumn 2008

ow we can use the cofactors to connect 2x2 determinants (which we know from Eq. (7.8)) to the desired x determinants. he general result is that we can evaluate the determinant of an x matrix by first picking a specific (but arbitrary) row or column here we choose row k and then evaluating the sum of each element in that row times its cofactor A A A C k det any fixed (7.2) l where l is summed over and k is not. Since the cofactor itself involves a determinant of a smaller matrix we can iterate this expression in terms of determinants of successively smaller matrices until we reach the 2x2 case (this construction is called the Laplace development for determinants). Square matrices with nonzero determinants are nonsingular matrices and as we will see shortly have inverses. If follows from this definition of the determinant (although we will not go through the arithmetic here) that the determinant of a product of matrices is the product of the individual determinants det[ AB] det[ A] det[ B]. (7.22) Consider some simple examples applying the rules above : A A A A 2 3 2: A A A A A 22 2 2 22 2 2 A A A A A A 3: A A A A etc. A A 22 23 2 23 2 22 2 3 A32 A33 A3 A33 A3 A32 here are several important properties of determinants following from Eq. (7.2) that help to simplify the process of evaluating determinants (recall our motto). ) If (all of the elements in) a single row or a single column are multiplied by a constant c the determinant of the new matrix is c times the determinant of the Physics 227 Lecture 7 8 Autumn 2008

original matrix. 2) he sign of a determinant (but not its absolute value) is changed if we interchange 2 rows or 2 columns. 3) he determinant is unchanged if we interchange all the columns with all the rows A A i.e. the determinant of the transpose matrix is the same as the determinant of the original matrix A A A A. 4) he determinant is unchanged if we add a constant times any row to another row or the same for columns. 5) he determinant of a matrix vanishes if all of the elements in any row or column vanish. 6) he determinant of a matrix vanishes if any two rows or any two columns are identical (as follows from points 4) and 5)). 7) he determinant of a matrix vanishes if any two rows or any two columns differ by a constant multiplier i.e. are proportional (as follows from points 4) and 5)). We can use these properties to simplify a matrix i.e. get as many zero elements as possible before taking the determinant (again recall our motto!). As an example consider the following matrix and evaluate its determinant by using the elements in the first row 2 3 4 5 6 45 48 2 36 42 3 32 35 7 8 9 32 9 0. (7.23) Instead consider proceeding by first subtracting 4 times the first row from the second (R2-4*R) then take 7 times the first row from the third row (R3-7*R) and finally subtract 2 times the second row from the third row (R3-2*R2) Physics 227 Lecture 7 9 Autumn 2008

2 3 2 3 2 3 2 3 4 5 6 0 3 6 0 3 6 0 3 6 0. R23* R R37* R R32* R2 7 8 9 7 8 9 0 6 2 0 0 0 (7.24) We obtain the same answer but in some sense the arithmetic is simpler. ote that this analysis also works the other way in the sense that for a vanishing determinant we can always manipulate the form of the determinant by using the rules above into an expression with all zeros along one row or one column. o return to the question of finding the inverse of a matrix consider a slightly edited version of Eq. (7.2) i.e. consider what happens if we multiply a row (or column) of elements by the cofactors of a different row (or column). For example consider the quantity k m A Cml det[ A] (7.25) l which we recognize as yielding the determinant of a matrix A that differs from the matrix A in that both the k th and the m th row of A are equal to the k th row of A. hus these two rows in A are equal to each other and by rule 6) above we can conclude that det A 0. he same is true for summing on the elements of a column. So the right hand side of Eq. (7.25) can be nonzero only if (i.e. we only get a nonzero result for the original expression for the determinant if) k = m A C A C det[ A]. (7.26) ml lm km l l Recalling that the transpose of a matrix is obtained by exchanging rows and columns we are led to consider the product A A (7.27) AC for which we can write using Eq. (7.26) that Physics 227 Lecture 7 0 Autumn 2008

AC A C A Cml km det A km lm l l C AC det A A. det A (7.28) Comparing this last equation with Eq. (7.6) we see that we have found a general expression for the inverse of the matrix A. We have A A C det A C det A C det A. (7.29) It should now be clear why nonsingular matrices with inverses require nonzero determinants. he analogue statement to Eq. (7.22) about inverses of products is (note the order of the matrices) AB B A AB AB B A AB B B B B. (7.30) As an explicit example consider a 0 b 2 2 A 0 0 det A aa 0 bb a b b c a a 0 b a bc b 2 2 2 2 C bc a b ac C 0 a b 0 b 0 a b ac a Physics 227 Lecture 7 Autumn 2008

a bc b C 2 2 A 0 a b 0. 2 2 det A a b b ac a (7.3) he reader is encouraged to verify that with the above expressions AA We close this introduction to/review of matrices with a summary of vocabulary/notation and general properties. Operations on matrices: A A. t ranspose A A A A : A A AB B A : AB AB A B B A Complex Conjugate A A : A A Hermitian Conjugate [or Adjoint] A A : A A. ˆ Old Adjoint : ˆ A C A C C lm mk km ml ypes of matrices: Symmetric Matrix A A : A A Antisymmnetric Matrix A A : A A Real Matrix A A : A A Pure Imaginary Matrix A A : A A Orthogonal Matrix : A A A A : AA Hermitian Matrix A A Unitary Matrix : A A A A : AA. : A A Physics 227 Lecture 7 2 Autumn 2008