Hebbian Learning II. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. July 20, 2017
|
|
- Pamela Higgins
- 5 years ago
- Views:
Transcription
1 Hebbian Learning II Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester July 20, 2017
2 Goals Teach about one-half of an undergraduate course on Linear Algebra Understand when supervised Hebbian learning works perfectly (and when it does not) Patten completion: Supervised Hebbian learning finds a weight matrix such that the input vectors are the eigenvectors of this weight matrix
3 Vector age height weight Joe = Mary = Vectors have both a length (magnitude) and a direction
4 Graphical representation for the vector Mary :
5 Multiplication of a Vector by a Scalar 2 [ 2 1 ] = [ 4 2 ] Scalar multiplication corresponds to lengthening or shortening a vector (while leaving it pointing in the same direction)
6
7 Addition of Vectors = 3 3 4
8
9 Linear Combination of Vectors u = c 1 v 1 + c 2 v 2 The set of all linear combinations of the { v i } is called the set spanned by the { v i }
10
11 The three vectors 1 0 0, 0 1 0, span all of three-dimensional space since any vector u = written as a linear combination: 1 u = a 0 + b c a b c can be In general, n vectors suffice to span n-dimensional space
12 Linear Independence If none of the vectors in a set can be written as a linear combination of the others, then the set is called linearly independent n-dimensional space is the set of vectors spanned by a set of n linearly independent vectors. The n vectors are referred to as a basis for the space
13 [ ] [ ] 1 2, are colinear and, thus, are linearly dependent. They 1 2 span only a 1-dimensional space [ ] [ ] 1 2, are linearly independent (and, thus, span a dimensional space)
14 [ 1 1 ] [ 2, 1 ] [ 1, 3 ] are linearly dependent ( v 3 = 7 v 1 4 v 2 ) 1 2 0, 3 2 0, are linearly dependent (no vector with a non-zero third component can be generated from this set)
15 For a given n-dimensional space, there are an infinite number of basis for that space Every vector has a different representation (i.e., set of coordinates) for each basis Which basis is best?
16 Inner (Dot) Product of Two Vectors v = 3 1 2, w = v w = i v i w i = (3 1) + ( 1 2) + (2 1) = 3
17 Length of a Vector The length of a vector (denoted ) is the square root of the inner product of the vector with itself Let v = v = ( v v) 1/2 = ( 1)
18 Follows from the Pythagorean Theorem:
19 Angle Between Two Vectors cos θ = = v w v w i v iw i ( i v2 i )1/2 ( i w2 i )1/2 Roughly a measure of similarity between two vectors: If v and w are random variables (so v and w are vectors of values for these variables) with zero mean, then this formula is their correlation If the inner product is zero, then cos θ = 0 (meaning that the two vectors are orthogonal)
20 Projection of One Vector Onto Another Vector
21 Let x be the projection of v onto w (a number, not a vector): x = v cos θ = v w w If w = 1, then x = v w
22 Example: Linear Neural Network with One Output Unit x 1 w 1 x2 w 2 y x 3 w 3 The output (y = w 1 x 1 + w 2 x 2 + w 3 x 3 = w x) gives an indication of how close the input x is to the weight vector w: If y > 0, then x is similar to w If y = 0, then x is orthogonal to w If y < 0, then x is dissimilar to w
23 Matrix An array of numbers. For example: [ ] W = 1 0 1
24 Multiplication of a Matrix and a Vector u = W v: Matrix W maps from one space of vectors ( v) to a new space of vectors ( u) In general, vectors v and u may have different dimensionalities
25 Multiplication of a Matrix and a Vector W = [ ] v = u = W v = = = [ ] [ (3 1) + (4 0) + (5 2) (1 1) + (0 0) + (1 2) [ ] 13 3 ]
26 Multiplication of a Matrix and a Vector The following are equivalent: Form inner product of each row of matrix with vector u = W v is a linear combination of the columns of W. The coefficients are the components of v
27 Neural Network with Multiple Input and Multiple Output Units x 1 w 11 y 1 x2 y 2 x 3 w 23 [ y1 y 2 ] = [ w11 w 12 w 13 w 21 w 22 w 23 ] x 1 x 2 x 3 y = W x
28 Linearity A function is said to be linear if: f(cx) = cf(x) f(x 1 + x 2 ) = f(x 1 ) + f(x 2 )
29 Implication: If we know how a system responds to the basis of a space, then we can easily compute how it responds to all vectors in that space Let { v i } be a basis for a space Let v be an arbitrary vector in this space Then: W v = W (c 1 v 1 + c 2 v c n v n ) = c 1 W v 1 + c 2 W v c n W v n
30 Eigenvectors and Eigenvalues Limit our attention to square matrices (i.e., v and u have the same dimensionality) In general, multiplication by a matrix changes both a vector s direction and length However, there are some vectors that will change only in length, not direction For these vectors, multiplication by the matrix is no different than multiplication by a scalar where λ is a scalar W v = λ v Such vectors are called eigenvectors, and the scalar λ is called an eigenvalue
31 [ ] [ 1 2 ] [ 1 = 2 2 ] Each vector that is colinear with an eigenvector is itself an eigenvector: [ ] [ 2 4 ] [ 2 = 2 4 ] We will reserve the term eigenvector only for vectors of length 1
32 An n n matrix can have up to (but no more than) n distinct eigenvalues If it has n distinct eigenvalues, then the n associated eigenvectors are linearly independent Thus, these eigenvectors form a basis
33 Let { v i } be linearly independent eigenvectors of matrix W, and let v be an arbitrary vector. Then: u = W v = W (c 1 v c n v n ) = c 1 W v c n W v n = c 1 λ 1 v 1 + c n λ n v n There are no matrices in this last equation. Just a simple linear combination of eigenvectors.
34 Eigenvectors and eigenvalues reveal the directions in which matrix multiplication stretches and shrinks a space (i.e., it reveals which input vectors a system gives small and large responses to). v W u Power method for finding the largest eigenvector of a matrix
35 Transpose Turn a column vector into a row vector. For example, we can re-write an inner product as follows: w v = w T v = [w 1 w 2 w 3 ] v 1 v 2 v 3 = (w 1 v 1 ) + (w 2 v 2 ) + (w 3 v 3 )
36 Outer Product w v T = = w 1 w 2 w 3 [v 1 v 2 v 3 ] w 1 v 1 w 1 v 2 w 1 v 3 w 2 v 1 w 2 v 2 w 2 v 3 w 3 v 1 w 3 v 2 w 3 v 3 If w and v are random variables (the components of w and v are values of these variables) with zero means, and if w = v, then this is a covariance matrix
37 Using Linear Algebra to Study Supervised Hebbian Learning
38 Neural Network With One Output Unit: One Input-Output Pattern One input-output pattern: x y Assume x = 1 If we choose w = x, then y = w T x = x T x = 1 But we want y to equal y. So let w = y x w T x = (y x) T x = y ( x T x) = y
39 Problem of finding w corresponds to finding a vector whose projection onto x is y x y * There are an infinite number of solutions On the previous slide, we made the simple choice of the vector that points in the same directions as x
40 Neural Network With Multiple Output Units: One Input-Output Pattern One input-output pattern: x y Assume x = 1 Let W = y x T y = W x = ( y x T ) x = y ( x T x) = y
41 Example With Multiple Input-Output Patterns x 1 = x 2 = x 3 = y 1 = 3 y 2 = 2 y 3 = 4
42 Based on 1st pattern, w 1 = Based on 2nd pattern, w 2 = Based on 3rd pattern, w 3 = Next, set weight matrix W : W = w 1 + w 2 + w =
43 Verify: W x 1 = y 1 W x 2 = y 2 W x 3 = y 3 Q: Why does this work? A: If the input vectors are orthogonal, then the Hebb rule works perfectly (!!!)
44 Hebb Rule Works Perfectly When Inputs Are Orthogonal Assume input vectors are unit length and mutually orthogonal: { 1 if i = j x T i x j = 0 else Set W i = y i x T Set W = W W n
45 For all i: y = W x i = (W W n ) x i = ( y 1 x T 1 + y n x T n) x i = y 1 x T 1 x i + + y n x T n x i = y i = y i
46 Caveat If the input vectors are not orthogonal, the Hebb rule is not guaranteed to work perfectly If the input vectors are linearly independent, the LMS rule works perfectly
47 Example: Hebb Fails Input Output Hebb rule: overall weight changes for w 1, w 2, and w 4 are 0 (i.e., Hebb rule does not work) There are successful weights: w 1 = 1, w 2 = 1, w 3 = 2, and w 4 = 1 (but Hebb rule won t find these values)
48 Hebb Learning and Pattern Completion
49 Recurrent Network
50 Associate input vectors with scalar copies of themselves: Assume λ i are distinct y i = λ i x i Assume input vectors are unit length and mutually orthogonal: { 1 if i = j x T i x j = 0 else
51 Set W i = y i x T i = λ i x i x T i Set W = W W n W x i = (W W n ) x i = (λ 1 x 1 x T 1 + λ n x n x T n) x i = λ 1 x 1 x T 1 x i + + λ n x n x T n x i = λ i x i = λ i x i Hebb rule creates a weight matrix such that the input vectors are the eigenvectors of this matrix (!!!)
Review of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationELE/MCE 503 Linear Algebra Facts Fall 2018
ELE/MCE 503 Linear Algebra Facts Fall 2018 Fact N.1 A set of vectors is linearly independent if and only if none of the vectors in the set can be written as a linear combination of the others. Fact N.2
More information2018 Fall 2210Q Section 013 Midterm Exam I Solution
8 Fall Q Section 3 Midterm Exam I Solution True or False questions ( points = points) () An example of a linear combination of vectors v, v is the vector v. True. We can write v as v + v. () If two matrices
More informationThis appendix provides a very basic introduction to linear algebra concepts.
APPENDIX Basic Linear Algebra Concepts This appendix provides a very basic introduction to linear algebra concepts. Some of these concepts are intentionally presented here in a somewhat simplified (not
More informationVectors and Matrices Statistics with Vectors and Matrices
Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationVAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:
VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationThere are two things that are particularly nice about the first basis
Orthogonality and the Gram-Schmidt Process In Chapter 4, we spent a great deal of time studying the problem of finding a basis for a vector space We know that a basis for a vector space can potentially
More informationMATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.
MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors. Orthogonal sets Let V be a vector space with an inner product. Definition. Nonzero vectors v 1,v
More informationWe use the overhead arrow to denote a column vector, i.e., a number with a direction. For example, in three-space, we write
1 MATH FACTS 11 Vectors 111 Definition We use the overhead arrow to denote a column vector, ie, a number with a direction For example, in three-space, we write The elements of a vector have a graphical
More informationMath 3191 Applied Linear Algebra
Math 191 Applied Linear Algebra Lecture 1: Inner Products, Length, Orthogonality Stephen Billups University of Colorado at Denver Math 191Applied Linear Algebra p.1/ Motivation Not all linear systems have
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationVectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =
Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.
More information1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?
. Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in
More informationMATH 115A: SAMPLE FINAL SOLUTIONS
MATH A: SAMPLE FINAL SOLUTIONS JOE HUGHES. Let V be the set of all functions f : R R such that f( x) = f(x) for all x R. Show that V is a vector space over R under the usual addition and scalar multiplication
More informationwhich are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar.
It follows that S is linearly dependent since the equation is satisfied by which are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar. is
More informationFamily Feud Review. Linear Algebra. October 22, 2013
Review Linear Algebra October 22, 2013 Question 1 Let A and B be matrices. If AB is a 4 7 matrix, then determine the dimensions of A and B if A has 19 columns. Answer 1 Answer A is a 4 19 matrix, while
More information22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes
Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationLecture II: Linear Algebra Revisited
Lecture II: Linear Algebra Revisited Overview Vector spaces, Hilbert & Banach Spaces, etrics & Norms atrices, Eigenvalues, Orthogonal Transformations, Singular Values Operators, Operator Norms, Function
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationLinear Equations in Linear Algebra
Linear Equations in Linear Algebra.7 LINEAR INDEPENDENCE LINEAR INDEPENDENCE Definition: An indexed set of vectors {v,, v p } in n is said to be linearly independent if the vector equation x x x 2 2 p
More informationLinear Algebra V = T = ( 4 3 ).
Linear Algebra Vectors A column vector is a list of numbers stored vertically The dimension of a column vector is the number of values in the vector W is a -dimensional column vector and V is a 5-dimensional
More informationLinear Combination. v = a 1 v 1 + a 2 v a k v k
Linear Combination Definition 1 Given a set of vectors {v 1, v 2,..., v k } in a vector space V, any vector of the form v = a 1 v 1 + a 2 v 2 +... + a k v k for some scalars a 1, a 2,..., a k, is called
More informationThe set of all solutions to the homogeneous equation Ax = 0 is a subspace of R n if A is m n.
0 Subspaces (Now, we are ready to start the course....) Definitions: A linear combination of the vectors v, v,..., v m is any vector of the form c v + c v +... + c m v m, where c,..., c m R. A subset V
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationLinear Algebra (Review) Volker Tresp 2017
Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationMatrix Algebra: Summary
May, 27 Appendix E Matrix Algebra: Summary ontents E. Vectors and Matrtices.......................... 2 E.. Notation.................................. 2 E..2 Special Types of Vectors.........................
More information5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.
Linear Algebra - Test File - Spring Test # For problems - consider the following system of equations. x + y - z = x + y + 4z = x + y + 6z =.) Solve the system without using your calculator..) Find the
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationEigenvalue and Eigenvector Homework
Eigenvalue and Eigenvector Homework Olena Bormashenko November 4, 2 For each of the matrices A below, do the following:. Find the characteristic polynomial of A, and use it to find all the eigenvalues
More informationj=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.
Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u
More informationIntro Vectors 2D implicit curves 2D parametric curves. Graphics 2012/2013, 4th quarter. Lecture 2: vectors, curves, and surfaces
Lecture 2, curves, and surfaces Organizational remarks Tutorials: TA sessions for tutorial 1 start today Tutorial 2 will go online after lecture 3 Practicals: Make sure to find a team partner very soon
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationEcon Slides from Lecture 7
Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for
More informationLinear Algebra - Part II
Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationLinear Algebra Review
Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite
More informationUsually, when we first formulate a problem in mathematics, we use the most familiar
Change of basis Usually, when we first formulate a problem in mathematics, we use the most familiar coordinates. In R, this means using the Cartesian coordinates x, y, and z. In vector terms, this is equivalent
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationChapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of
Chapter 2 Linear Algebra In this chapter, we study the formal structure that provides the background for quantum mechanics. The basic ideas of the mathematical machinery, linear algebra, are rather simple
More information(v, w) = arccos( < v, w >
MA322 Sathaye Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v
More information2. Review of Linear Algebra
2. Review of Linear Algebra ECE 83, Spring 217 In this course we will represent signals as vectors and operators (e.g., filters, transforms, etc) as matrices. This lecture reviews basic concepts from linear
More informationThe 'linear algebra way' of talking about "angle" and "similarity" between two vectors is called "inner product". We'll define this next.
Orthogonality and QR The 'linear algebra way' of talking about "angle" and "similarity" between two vectors is called "inner product". We'll define this next. So, what is an inner product? An inner product
More informationDefinitions for Quizzes
Definitions for Quizzes Italicized text (or something close to it) will be given to you. Plain text is (an example of) what you should write as a definition. [Bracketed text will not be given, nor does
More informationNOTES ON LINEAR ALGEBRA CLASS HANDOUT
NOTES ON LINEAR ALGEBRA CLASS HANDOUT ANTHONY S. MAIDA CONTENTS 1. Introduction 2 2. Basis Vectors 2 3. Linear Transformations 2 3.1. Example: Rotation Transformation 3 4. Matrix Multiplication and Function
More informationVectors Coordinate frames 2D implicit curves 2D parametric curves. Graphics 2008/2009, period 1. Lecture 2: vectors, curves, and surfaces
Graphics 2008/2009, period 1 Lecture 2 Vectors, curves, and surfaces Computer graphics example: Pixar (source: http://www.pixar.com) Computer graphics example: Pixar (source: http://www.pixar.com) Computer
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationDot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.
Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................
More informationMAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:
MAC Module Eigenvalues and Eigenvectors Learning Objectives Upon completing this module, you should be able to: Solve the eigenvalue problem by finding the eigenvalues and the corresponding eigenvectors
More informationExample Linear Algebra Competency Test
Example Linear Algebra Competency Test The 4 questions below are a combination of True or False, multiple choice, fill in the blank, and computations involving matrices and vectors. In the latter case,
More informationMAC Module 12 Eigenvalues and Eigenvectors
MAC 23 Module 2 Eigenvalues and Eigenvectors Learning Objectives Upon completing this module, you should be able to:. Solve the eigenvalue problem by finding the eigenvalues and the corresponding eigenvectors
More informationImage Registration Lecture 2: Vectors and Matrices
Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this
More information6.1. Inner Product, Length and Orthogonality
These are brief notes for the lecture on Friday November 13, and Monday November 1, 2009: they are not complete, but they are a guide to what I want to say on those days. They are guaranteed to be incorrect..1.
More informationMTH 2032 SemesterII
MTH 202 SemesterII 2010-11 Linear Algebra Worked Examples Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education December 28, 2011 ii Contents Table of Contents
More informationConsequences of Orthogonality
Consequences of Orthogonality Philippe B. Laval KSU Today Philippe B. Laval (KSU) Consequences of Orthogonality Today 1 / 23 Introduction The three kind of examples we did above involved Dirichlet, Neumann
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationMath 1553, Introduction to Linear Algebra
Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level
More informationLinear Algebra and Dirac Notation, Pt. 1
Linear Algebra and Dirac Notation, Pt. 1 PHYS 500 - Southern Illinois University February 1, 2017 PHYS 500 - Southern Illinois University Linear Algebra and Dirac Notation, Pt. 1 February 1, 2017 1 / 13
More informationRepeated Eigenvalues and Symmetric Matrices
Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one
More informationFFTs in Graphics and Vision. The Laplace Operator
FFTs in Graphics and Vision The Laplace Operator 1 Outline Math Stuff Symmetric/Hermitian Matrices Lagrange Multipliers Diagonalizing Symmetric Matrices The Laplacian Operator 2 Linear Operators Definition:
More informationIn biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.
In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required. In humans, association is known to be a prominent feature of memory.
More informationLinear Equations in Linear Algebra
1 Linear Equations in Linear Algebra 1.7 LINEAR INDEPENDENCE LINEAR INDEPENDENCE Definition: An indexed set of vectors {v 1,, v p } in n is said to be linearly independent if the vector equation x x x
More informationIntroduction to Matrix Algebra
Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p
More informationDefinition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition
6 Vector Spaces with Inned Product Basis and Dimension Section Objective(s): Vector Spaces and Subspaces Linear (In)dependence Basis and Dimension Inner Product 6 Vector Spaces and Subspaces Definition
More informationElements of linear algebra
Elements of linear algebra Elements of linear algebra A vector space S is a set (numbers, vectors, functions) which has addition and scalar multiplication defined, so that the linear combination c 1 v
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationLinear Algebra Practice Problems
Math 7, Professor Ramras Linear Algebra Practice Problems () Consider the following system of linear equations in the variables x, y, and z, in which the constants a and b are real numbers. x y + z = a
More informationChapter Two Elements of Linear Algebra
Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationSolution: By inspection, the standard matrix of T is: A = Where, Ae 1 = 3. , and Ae 3 = 4. , Ae 2 =
This is a typical assignment, but you may not be familiar with the material. You should also be aware that many schools only give two exams, but also collect homework which is usually worth a small part
More informationEigenvalues and Eigenvectors
5 Eigenvalues and Eigenvectors 5.2 THE CHARACTERISTIC EQUATION DETERMINANATS n n Let A be an matrix, let U be any echelon form obtained from A by row replacements and row interchanges (without scaling),
More informationDiagonalizing Matrices
Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B
More informationLinear Independence. Linear Algebra MATH Linear Algebra LI or LD Chapter 1, Section 7 1 / 1
Linear Independence Linear Algebra MATH 76 Linear Algebra LI or LD Chapter, Section 7 / Linear Combinations and Span Suppose s, s,..., s p are scalars and v, v,..., v p are vectors (all in the same space
More informationMATRIX ALGEBRA. or x = (x 1,..., x n ) R n. y 1 y 2. x 2. x m. y m. y = cos θ 1 = x 1 L x. sin θ 1 = x 2. cos θ 2 = y 1 L y.
as Basics Vectors MATRIX ALGEBRA An array of n real numbers x, x,, x n is called a vector and it is written x = x x n or x = x,, x n R n prime operation=transposing a column to a row Basic vector operations
More informationLinear Algebra. Alvin Lin. August December 2017
Linear Algebra Alvin Lin August 207 - December 207 Linear Algebra The study of linear algebra is about two basic things. We study vector spaces and structure preserving maps between vector spaces. A vector
More informationLinear Algebra Practice Problems
Linear Algebra Practice Problems Page of 7 Linear Algebra Practice Problems These problems cover Chapters 4, 5, 6, and 7 of Elementary Linear Algebra, 6th ed, by Ron Larson and David Falvo (ISBN-3 = 978--68-78376-2,
More informationIntro Vectors 2D implicit curves 2D parametric curves. Graphics 2011/2012, 4th quarter. Lecture 2: vectors, curves, and surfaces
Lecture 2, curves, and surfaces Organizational remarks Tutorials: Tutorial 1 will be online later today TA sessions for questions start next week Practicals: Exams: Make sure to find a team partner very
More informationMATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve:
MATH 2331 Linear Algebra Section 1.1 Systems of Linear Equations Finding the solution to a set of two equations in two variables: Example 1: Solve: x x = 3 1 2 2x + 4x = 12 1 2 Geometric meaning: Do these
More informationSymmetric and anti symmetric matrices
Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal
More informationMath 2114 Common Final Exam May 13, 2015 Form A
Math 4 Common Final Exam May 3, 5 Form A Instructions: Using a # pencil only, write your name and your instructor s name in the blanks provided. Write your student ID number and your CRN in the blanks
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationCS123 INTRODUCTION TO COMPUTER GRAPHICS. Linear Algebra 1/33
Linear Algebra 1/33 Vectors A vector is a magnitude and a direction Magnitude = v Direction Also known as norm, length Represented by unit vectors (vectors with a length of 1 that point along distinct
More information3 a 21 a a 2N. 3 a 21 a a 2M
APPENDIX: MATHEMATICS REVIEW G 12.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers 2 A = 6 4 a 11 a 12... a 1N a 21 a 22... a 2N. 7..... 5 a M1 a M2...
More informationHOSTOS COMMUNITY COLLEGE DEPARTMENT OF MATHEMATICS
HOSTOS COMMUNITY COLLEGE DEPARTMENT OF MATHEMATICS MAT 217 Linear Algebra CREDIT HOURS: 4.0 EQUATED HOURS: 4.0 CLASS HOURS: 4.0 PREREQUISITE: PRE/COREQUISITE: MAT 210 Calculus I MAT 220 Calculus II RECOMMENDED
More informationCS123 INTRODUCTION TO COMPUTER GRAPHICS. Linear Algebra /34
Linear Algebra /34 Vectors A vector is a magnitude and a direction Magnitude = v Direction Also known as norm, length Represented by unit vectors (vectors with a length of 1 that point along distinct axes)
More informationMath 2331 Linear Algebra
6.1 Inner Product, Length & Orthogonality Math 2331 Linear Algebra 6.1 Inner Product, Length & Orthogonality Shang-Huan Chiu Department of Mathematics, University of Houston schiu@math.uh.edu math.uh.edu/
More informationChapter 1 Vector Spaces
Chapter 1 Vector Spaces Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 110 Linear Algebra Vector Spaces Definition A vector space V over a field
More informationThe Transpose of a Vector
8 CHAPTER Vectors The Transpose of a Vector We now consider the transpose of a vector in R n, which is a row vector. For a vector u 1 u. u n the transpose is denoted by u T = [ u 1 u u n ] EXAMPLE -5 Find
More informationProperties of Linear Transformations from R n to R m
Properties of Linear Transformations from R n to R m MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Topic Overview Relationship between the properties of a matrix transformation
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More informationReview: Linear and Vector Algebra
Review: Linear and Vector Algebra Points in Euclidean Space Location in space Tuple of n coordinates x, y, z, etc Cannot be added or multiplied together Vectors: Arrows in Space Vectors are point changes
More informationIFT 6760A - Lecture 1 Linear Algebra Refresher
IFT 6760A - Lecture 1 Linear Algebra Refresher Scribe(s): Tianyu Li Instructor: Guillaume Rabusseau 1 Summary In the previous lecture we have introduced some applications of linear algebra in machine learning,
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More informationELEMENTARY LINEAR ALGEBRA
ELEMENTARY LINEAR ALGEBRA DAVID MEREDITH These notes are about the application and computation of linear algebra objects in F n, where F is a field. I think everything here works over any field including
More informationLarge Scale Data Analysis Using Deep Learning
Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset
More information