Data representation and classification

Similar documents
Dot product. The dot product is an inner product on a coordinate vector space (Definition 1, Theorem

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Fourier and Wavelet Signal Processing

INNER PRODUCT SPACE. Definition 1

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

Math Linear Algebra

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

MATH 167: APPLIED LINEAR ALGEBRA Chapter 3

This pre-publication material is for review purposes only. Any typographical or technical errors will be corrected prior to publication.

Linear Algebra V = T = ( 4 3 ).

Applied Linear Algebra in Geoscience Using MATLAB

LAKELAND COMMUNITY COLLEGE COURSE OUTLINE FORM

Cambridge University Press The Mathematics of Signal Processing Steven B. Damelin and Willard Miller Excerpt More information

EGR2013 Tutorial 8. Linear Algebra. Powers of a Matrix and Matrix Polynomial

Vectors in Function Spaces

Vector Spaces. Commutativity of +: u + v = v + u, u, v, V ; Associativity of +: u + (v + w) = (u + v) + w, u, v, w V ;

Review of Linear Algebra

Econ Slides from Lecture 7

2 Metric Spaces Definitions Exotic Examples... 3

Lecture Notes 1: Vector spaces

Quantum Computing Lecture 2. Review of Linear Algebra

Lecture 3: Vectors. In Song Kim. September 1, 2011

MATH 304 Linear Algebra Lecture 19: Least squares problems (continued). Norms and inner products.

The following definition is fundamental.

Linear Algebra for Machine Learning. Sargur N. Srihari

Typical Problem: Compute.

Definitions and Properties of R N

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Linear Algebra. Min Yan

MTH 2032 SemesterII

Data representation and approximation

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Categories and Quantum Informatics: Hilbert spaces

Euclidean Space. This is a brief review of some basic concepts that I hope will already be familiar to you.

Math Linear Algebra II. 1. Inner Products and Norms

Review of Some Concepts from Linear Algebra: Part 2

Linear Analysis Lecture 5

The geometry of least squares

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Math 3191 Applied Linear Algebra

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors

Linear Algebra Review. Vectors

1. General Vector Spaces

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

0.2 Vector spaces. J.A.Beachy 1

Inner Product and Orthogonality

(v, w) = arccos( < v, w >

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

Linear Algebra Massoud Malek

Fairfield Public Schools

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

Lecture 2: Vector-Vector Operations

Lecture 23: 6.1 Inner Products

3 a 21 a a 2N. 3 a 21 a a 2M

MATH 320, WEEK 7: Matrices, Matrix Operations

Principal Components Theory Notes

MTH 2310, FALL Introduction

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space.

MATH 0960 ELEMENTARY ALGEBRA FOR COLLEGE STUDENTS (8 TH EDITION) BY ANGEL & RUNDE Course Outline

Lecture 02: 2D Geometry. August 24, 2017

MTAEA Vectors in Euclidean Spaces

12x + 18y = 30? ax + by = m

(v, w) = arccos( < v, w >

Inner Product Spaces

Precalculus Graphical, Numerical, Algebraic Media Update 7th Edition 2010, (Demana, et al)

Chapter 3 Transformations

2. Signal Space Concepts

Vector Spaces. Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms.

Linear Algebra Practice Problems

Inner Product Spaces

Vectors Coordinate frames 2D implicit curves 2D parametric curves. Graphics 2008/2009, period 1. Lecture 2: vectors, curves, and surfaces

Review: Linear and Vector Algebra

Numerical Linear Algebra

On Some Properties of Generalized Fock Space F 2 (d v α ) by Frame Theory on the C n

SYLLABUS UNDER AUTONOMY MATHEMATICS

SYLLABUS. 1 Linear maps and matrices

ON GENERALIZED n-inner PRODUCT SPACES

Math 302 Outcome Statements Winter 2013

Section 3.1: Definition and Examples (Vector Spaces)

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

ORTHOGONALITY AND LEAST-SQUARES [CHAP. 6]

MATH 22A: LINEAR ALGEBRA Chapter 1

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Lecture 2. Econ August 11

Linear Algebra I for Science (NYC)

Lecture 7. Econ August 18

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

Tues Feb Vector spaces and subspaces. Announcements: Warm-up Exercise:

Introduction to Hilbert Space Frames

Algebra I. Course Outline

2.1 Definition. Let n be a positive integer. An n-dimensional vector is an ordered list of n real numbers.

Stat 159/259: Linear Algebra Notes

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Tune-Up Lecture Notes Linear Algebra I

Transcription:

Representation and classification of data January 25, 2016

Outline Lecture 1: Data Representation 1 Lecture 1: Data Representation

Data representation The phrase data representation usually refers to the methods used to represent information stored in a computer. Computers can store many different types of information, including, but not limited to: numbers, text, graphics (many varieties), sound, etc etc. Each such type of information must be formed using appropriate codes, because in the end, all of the information stored in a computer must be represented by finite sequences of 0s and 1s. This is how computer scientists and engineers think about data representation. Moreover, from this perspective, all processing of the stored information is performed on those binary sequences, and then appropriately interpreted through the use of aforementioned codes and other software techniques. In mathematics, the concept of data representation is much broader: it is any mathematical form that data can be described by. This includes analytic, algebraic, statistical, or geometric representations. In this class we will narrow this scope to focus on analytic/geometric forms of data representation by means of vector spaces.

Linear Representation by Vector Spaces A vector space is a mathematical structure formed by two types of objects: a collection of elements called vectors, and two operations on pairs of vectors: addition and scalar multiplication. As such two vectors may be added together, and any vector can multiplied by numbers, called scalars. Thus, for a vector space X, for any two of its elements x, y X, and any numbers α, β F (a field of numbers), we can form the linear combinations as new elements of X: α x + β y X.

Axioms of Vector Spaces The operations of addition and scalar multiplication in a vector space must satisfy a number of conditions called axioms: Associativity of addition Commutativity of addition Identity element of addition Inverse elements of addition Compatibility of scalar multiplication with field multiplication Identity element of scalar multiplication Distributivity of scalar multiplication with respect to vector addition Distributivity of scalar multiplication with respect to field addition Giuseppe Peano, Geometrical Calculus, 1888

Diversity of representations: from bases to frames A basis is a set of linearly independent vectors which can represent every vector in a given vector space through their linear combinations. An orthogonal basis for a vector space with an inner product, is a basis with vectors which are mutually orthogonal (perpendicular). If the vectors of an orthogonal basis are or length (norm) 1, the resulting basis is an orthonormal basis (ONB). An ONB may cease to be an ONB after even a small perturbation, or when any of its elements is removed. We seek thus representation systems with more stability. Frames were introduced by Dunford and Schaeffer in 1952. R. J. Dunford and A. C. Schaeffer, A class of nonharmonic Fourier series, Trans. Amer. Math. Soc., 1952, Vol. 72, pp. 341 366.

Bases Lecture 1: Data Representation The role of a basis is to allow us to represent elements of the vector space in terms of sequences of scalars (numbers), which are called vector coordinates. This is an important step, because thanks to this representation, abstract or complicated objects obtain a uniform mathematical format. The reason for this may not necessarily be clear when we think of the most typical example of a vector space: d-dimensional Euclidean vector space. This is because the Euclidean space is not just a good example of a vector space, it is also a prototypical example, and last but not least - a finite dimensional vector space. Infinite dimensional vector spaces provide us with more intriguing examples of objects, and the role of a basis which allows us to replace these complicated objects by sequences of numbers becomes much more clear. Vector spaces of polynomials; Function spaces (Lipschitz, integrable, finite energy functions, etc.)

Example: Polynomials of 1 variable with real coefficients The set of polynomials with coefficients in R is a vector space over R, denoted typically by P(R). Vector addition and scalar multiplication are defined in the obvious manner. If the degree of the polynomials is unrestricted then the dimension of P(R) is infinite. If instead one restricts the polynomials to those with degree less than or equal to an integer n, then we obtain a vector space of dimension n + 1, commonly denoted by P n (R). One possible basis for P(R) is the monomial basis: 1, x, x 2, x 3,.... The coordinates of any polynomial with respect to this basis are its coefficients. Say, for f = 3 + x 2 + 17.1x 7, the coefficients are 3, 0, 1, 0, 0, 0, 0, 17.1, 0,.... Now, the true advantage of using basis representations for polynomials requires us to recognize polynomials as functions, rather than just abstract algebraic objects. For this, however, we need to introduce some new concepts allowing us to measure distances in vector spaces.

Normed vector spaces A norm in a vector space X is a non-negative real-valued function x x, which satisfies the axioms of sublinearlity (also known as Minkowski inequality or triangle inequality), and non-degeneracy. A normed vector space is a pair (X, ) where X is a vector space and a norm on X. Every finite (N) dimensional vector space X can be equipped with a norm. Indeed, let {b 1,..., b N } be a basis for X. Then, for any x X, we can write uniquely: N x = x n b n. n=1 With this notation we can define various norms, e.g.,: x p = ( x 1 p +... + x N p ) 1 p, p 1, x = max{ x 1,..., x N }.

Euclidean norm When we fix p = 2 in the definition of the p-norm: x 2 = x 1 2 +... + x N 2, which is the standard Euclidean norm for the N-dimensional vector space. This is the same norm that is used in the least squares optimization problems.

Euclidean space R N The Euclidean space R N is a normed vector space consisting of N-tuples of real numbers (which we call vectors), with the operations of standard (coordinate-wise) vector addition and (real) scalar multiplication, and equipped with the Euclidean norm x 2. One of the major advantages of the Euclidean vector space over other normed spaces is the fact that we can measure distances between vectors using the concept of an inner product: x, y = x 1 y 1 + x 2 y 2 +... + x N y N. This inner product is naturally related to Euclidean norm via the following formula: x 2 2 = x, x.

Example: Fourier Basis for R N Given N > 0, define the following N N matrix: F(m, n) = 1 N e 2πimn/N, m, n = 0,... N 1. The columns (or rows) of this matrix form an orthonormal basis for the space of N-dimensional complex vectors R N. This basis is called the Fourier basis. And the matrix F is known as the Discrete Fourier Transform. Clearly F is a unitary matrix, and as such it is, in particular, invertible.