The Structure of Digital Imaging: The Haar Wavelet Transform. Wavelet Transformation and Image Compression

Size: px
Start display at page:

Download "The Structure of Digital Imaging: The Haar Wavelet Transform. Wavelet Transformation and Image Compression"

Transcription

1 The Structure of Digital Imaging: The Haar Wavelet Transformation and Image Compression December 7, 2009

2

3 Abstract The Haar wavelet transform allows images to be compressed and sent over a network of computers so that it is visable to anyone on the network. The mathematical apparatus for the transform is that of linear algebra and other elements of matrix theory, the main part of this being manipulations applied to the initial matrix representing a certian resolution of an image. It shall be shown how these techniques are used in conjuction with each other to create a given digital image of a certain resolution. Prerequisites: A rudimentary knowledge of matrices and arithmetical operations. Also, an indtroductory course in linear algebra would be helpful for the third section.

4 Overview How is an image sent over a computer network?

5 Overview How is an image sent over a computer network? It s possible through a technique called the Haar wavelet transformation.

6 Overview How is an image sent over a computer network? It s possible through a technique called the Haar wavelet transformation. There are two main mathematical processes averaging and differencing.

7 Overview How is an image sent over a computer network? It s possible through a technique called the Haar wavelet transformation. There are two main mathematical processes averaging and differencing. They are simple, but to use them in a large matrix requires some linear algebra.

8 Overview How is an image sent over a computer network? It s possible through a technique called the Haar wavelet transformation. There are two main mathematical processes averaging and differencing. They are simple, but to use them in a large matrix requires some linear algebra. Most of the operations can also be programmed into Matlab.

9 An Image as a Matrix Every Image represents a matrix

10 An Image as a Matrix Every Image represents a matrix The numbers within the matrix represent different shades of black and white or color.

11 An Image as a Matrix Every Image represents a matrix The numbers within the matrix represent different shades of black and white or color. The values of the numbers range from 0 to some positive whole number.

12 An Image as a Matrix Every Image represents a matrix The numbers within the matrix represent different shades of black and white or color. The values of the numbers range from 0 to some positive whole number. For this presentation we will only consider gray-scale images.

13 An Image as a Matrix So consider the matrix: A =

14 An Image as a Matrix So consider the matrix: A = This could represent some portion of an image

15 An Image as a Matrix The image Figure: Bird is a 256 by 256 matrix.

16 Averaging and Differencing So how do we transport the image?

17 Averaging and Differencing So how do we transport the image? Using averaging and differencing we can actually compress an image, making it easier to transport.

18 Averaging and Differencing So how do we transport the image? Using averaging and differencing we can actually compress an image, making it easier to transport. How is this done?

19 Averaging and Differencing To average first we isolate a row, called a data string, such as S 1 = ( ) This is row one of Matrix A.

20 Averaging and Differencing To average first we isolate a row, called a data string, such as S 1 = ( ) This is row one of Matrix A. We now can find the basic average of these terms, given by y = x 1 + x 2. 2 where x 1 and x 2 are two elements in S 1 which are next to each other.

21 Averaging and Differencing The difference is found by taking the average and subtracting them from x 1 like d = x 1 y

22 Averaging and Differencing The difference is found by taking the average and subtracting them from x 1 like d = x 1 y These differenced numbers are the detail coefficients.

23 Averaging and Differencing Both averaging and differencing must take place 3 times with a string with length 8.

24 Averaging and Differencing Both averaging and differencing must take place 3 times with a string with length 8. This is because 2 3 = 8.

25 Averaging and Differencing Both averaging and differencing must take place 3 times with a string with length 8. This is because 2 3 = 8. Here is a table which summarizes the results on S 1 : Table: Averaging and Differencing of S 1 Compression No. average/detail coefficient

26 Basic Compression Matrices Consider images as matrices. There must be an efficient way to apply the previous idea of averaging and differencing to a matrix the size of 256 by 256.

27 Basic Compression Matrices Consider images as matrices. There must be an efficient way to apply the previous idea of averaging and differencing to a matrix the size of 256 by 256. This would bring in a compression matrix: 1/ / / / / / / / / / / / / / / /2

28 This simplifies what could be a rather long process otherwise.

29 This simplifies what could be a rather long process otherwise. Here is our compression matrix multiplied with our initial matrix three times to create matrix T: T =

30 Compression The process of applying the compression matrix to the initial matrix multiple times is wavelet transform.

31 Compression The process of applying the compression matrix to the initial matrix multiple times is wavelet transform. This process is meant to transform data in the matrix to zero or near zero.

32 Compression The process of applying the compression matrix to the initial matrix multiple times is wavelet transform. This process is meant to transform data in the matrix to zero or near zero. A matrix is considered sparse when it is highly composed of zeros.

33 Compression The process of applying the compression matrix to the initial matrix multiple times is wavelet transform. This process is meant to transform data in the matrix to zero or near zero. A matrix is considered sparse when it is highly composed of zeros. Compression involves choosing a threshold value e, let s set e = 20.

34 Compression The process of applying the compression matrix to the initial matrix multiple times is wavelet transform. This process is meant to transform data in the matrix to zero or near zero. A matrix is considered sparse when it is highly composed of zeros. Compression involves choosing a threshold value e, let s set e = 20. Then all numbers that fall within the absolute value of e will be made into a zero.

35 Compression The process of applying the compression matrix to the initial matrix multiple times is wavelet transform. This process is meant to transform data in the matrix to zero or near zero. A matrix is considered sparse when it is highly composed of zeros. Compression involves choosing a threshold value e, let s set e = 20. Then all numbers that fall within the absolute value of e will be made into a zero. This helps to maitain the images beginning integrity while saving space or transmission time.

36 This is our matrix T after the threshold value e = 20 has been applied. We ll call it matrix D: D =

37 This is our matrix T after the threshold value e = 20 has been applied. We ll call it matrix D: D = This has created 27 zero components in our matrix creating a 2:1 compression ratio.

38 Progressive Image Transmission Progressive Image transmission utilizes the previous techniques.

39 Progressive Image Transmission Progressive Image transmission utilizes the previous techniques. The compression process allows the first image we attain to be comparable to our matrix T.

40 Progressive Image Transmission Progressive Image transmission utilizes the previous techniques. The compression process allows the first image we attain to be comparable to our matrix T. Matrix T is brought up starting with the overall average,larger detail coefficients and finally the smallest detail coefficients.

41 Progressive Image Transmission Progressive Image transmission utilizes the previous techniques. The compression process allows the first image we attain to be comparable to our matrix T. Matrix T is brought up starting with the overall average,larger detail coefficients and finally the smallest detail coefficients. T matrix initial image is crude but as more wavelet coefficients are used the image slowly becomes an exact copy of the initial image.

42 Figure: The progressive transmission of figure 1 (3rd compression to 1st compression)

43 Programming Compressions with Matlab Creating compressions of large matrices, such as 256 by 256 is difficult and time consuming by hand

44 Programming Compressions with Matlab Creating compressions of large matrices, such as 256 by 256 is difficult and time consuming by hand Therefore we program the compression matrices with Matlab

45 Programming Compressions with Matlab Creating compressions of large matrices, such as 256 by 256 is difficult and time consuming by hand Therefore we program the compression matrices with Matlab This is by defining the matrix as a function and writting code for a matrix with variable dimentions

46 Programming Compressions with Matlab Frist we define our function and variable. Using the name indmat

47 Programming Compressions with Matlab Frist we define our function and variable. Using the name indmat we code this like function a = indmat(n) b=[1;1]/2 c=[1;-1]/2

48 Programming Compressions with Matlab Frist we define our function and variable. Using the name indmat we code this like function a = indmat(n) b=[1;1]/2 c=[1;-1]/2 This creates two vectors, b, and c which consists of < 1 2, 1 2 > and < 1 2, 1 2 > respectively.

49 Programming Compressions with Matlab next we specify the first half of the matrix which will average the image.

50 Programming Compressions with Matlab next we specify the first half of the matrix which will average the image. this is while min(size(b))< n/2 b=[b, zeros(max(size(b)),min(size(b)));... zeros(max(size(b)),min(size(b))), b]; end

51 Programming Compressions with Matlab next we specify the first half of the matrix which will average the image. this is while min(size(b))< n/2 b=[b, zeros(max(size(b)),min(size(b)));... zeros(max(size(b)),min(size(b))), b]; end Then, to specify the differencing half, we use the same code, replacing c s for the b s.

52 Programming Compressions with Matlab next we specify the first half of the matrix which will average the image. this is while min(size(b))< n/2 b=[b, zeros(max(size(b)),min(size(b)));... zeros(max(size(b)),min(size(b))), b]; end Then, to specify the differencing half, we use the same code, replacing c s for the b s. We finally create a larger matrix out of the two of these with a=[b,c]

53 Programming Compressions with Matlab Now that we have the compression matrix we apply indmat(256) to our original image. This creates the 1st compression

54 Programming Compressions with Matlab Now that we have the compression matrix we apply indmat(256) to our original image. This creates the 1st compression To compress again we apply the matrix 2ndcompression=[indmat(128),zeroes(128);zeroes(128),eye(128)]

55 Programming Compressions with Matlab Now that we have the compression matrix we apply indmat(256) to our original image. This creates the 1st compression To compress again we apply the matrix 2ndcompression=[indmat(128),zeroes(128);zeroes(128),eye(128)] For the next compression the dimention is reduced by 2 again, but there must be more blocks. 3rdcompression=[indmat(64),zeroes(64),zeroes(64),zeroes(64); zeroes(64),eye(64),zeroes(64),zeroes(64);zeroes(64), zeroes(64),eye(64),zeroes(64);zeroes(64),zeroes(64),zeroes(64),eye(64)

56 Programming Compressions with Matlab Now that we have the compression matrix we apply indmat(256) to our original image. This creates the 1st compression To compress again we apply the matrix 2ndcompression=[indmat(128),zeroes(128);zeroes(128),eye(128)] For the next compression the dimention is reduced by 2 again, but there must be more blocks. 3rdcompression=[indmat(64),zeroes(64),zeroes(64),zeroes(64); zeroes(64),eye(64),zeroes(64),zeroes(64);zeroes(64), zeroes(64),eye(64),zeroes(64);zeroes(64),zeroes(64),zeroes(64),eye(64) Thus, eventually the image becomes compressed to a significant enough degree to be sent

57 The Linear Algebra of Image Compression We can generalize the averaging and differencing constituting the compression

58 The Linear Algebra of Image Compression We can generalize the averaging and differencing constituting the compression For any string the equation c 1 = sa 1 represents the first compression, where A is the general compression matrix.

59 The Linear Algebra of Image Compression We can generalize the averaging and differencing constituting the compression For any string the equation c 1 = sa 1 represents the first compression, where A is the general compression matrix. for the second compression, the equation is c 2 = c 1 A 2 the A matrix is the block matrix ( ) A 0 A 2 = 0 I

60 The Linear Algebra of Image Compression We can generalize the averaging and differencing constituting the compression For any string the equation c 1 = sa 1 represents the first compression, where A is the general compression matrix. for the second compression, the equation is c 2 = c 1 A 2 the A matrix is the block matrix ( ) A 0 A 2 = 0 I for the 3rd compression the equation is c 3 = c 2 A 3 with the A matrix being A A 3 = 0 I I I

61 The Linear Algebra of Image Compression in general the equation for the nth compression will be c n = c n 1 A n

62 The Linear Algebra of Image Compression in general the equation for the nth compression will be c n = c n 1 A n The matrix A n is A I c n = c n I I 2 n 1 1

63 The Linear Algebra of Image Compression We can multiply all of the A matrices together to do the entire series of compressions in one step

64 The Linear Algebra of Image Compression We can multiply all of the A matrices together to do the entire series of compressions in one step We call this W and say that W = A 1 A 2 A 3... A n where n is the total amount of compressions.

65 The Linear Algebra of Image Compression We can multiply all of the A matrices together to do the entire series of compressions in one step We call this W and say that W = A 1 A 2 A 3... A n where n is the total amount of compressions. So the equation c = sw, where c is the complete compression of the image, will do all of the avereging and differencing in one step.

66 The Linear Algebra of Image Compression The beauty of this process is that it is invertable; once compressed, the original image can be reconstructed from the compression.

67 The Linear Algebra of Image Compression The beauty of this process is that it is invertable; once compressed, the original image can be reconstructed from the compression. Not suprisingly, the equation s = cw 1 does this.

68 The Linear Algebra of Image Compression The beauty of this process is that it is invertable; once compressed, the original image can be reconstructed from the compression. Not suprisingly, the equation s = cw 1 does this. This is called the Inverse Haar Wavelet Transform.

69 The Linear Algebra of Image Compression The beauty of this process is that it is invertable; once compressed, the original image can be reconstructed from the compression. Not suprisingly, the equation s = cw 1 does this. This is called the Inverse Haar Wavelet Transform. For the general case, the equation W 1 = A 1 n A 1 n 1 A 1 n 2... A 1 1 gives the matrix W

70 The Linear Algebra of Image Compression For a String of length 2 n, W must have a dimension of n by n, and therefore so must the A matrices. Also note that n matrices are neededto completely compress the image.

71 The Linear Algebra of Image Compression For a String of length 2 n, W must have a dimension of n by n, and therefore so must the A matrices. Also note that n matrices are neededto completely compress the image. We can derive a product series which aids in the calculation of W and W 1 for strings of various lengths.

72 The Linear Algebra of Image Compression This is and W = W 1 = A n 0 I I I 2 i 1 1 i=1 A I I I 2 i 1 1 i=n 1

73 The Linear Algebra of Image Compression This is and W = W 1 = A n 0 I I I 2 i 1 1 i=1 A I I I 2 i 1 1 i=n 1 The general A matrix represented here is in blocks, the dimension of which are D/2 i 1 where D is the original dimension of the image.

74 The Linear Algebra of Image Compression Up to this point we have only been averaging the rows of the entire matrix (which we have called strings). Now we will talk about how to average the columns.

75 The Linear Algebra of Image Compression Up to this point we have only been averaging the rows of the entire matrix (which we have called strings). Now we will talk about how to average the columns. What we have computed so far is called the row-reduced form of a matrix. The complete compression will be a row-and-column-reduced form.

76 The Linear Algebra of Image Compression Up to this point we have only been averaging the rows of the entire matrix (which we have called strings). Now we will talk about how to average the columns. What we have computed so far is called the row-reduced form of a matrix. The complete compression will be a row-and-column-reduced form. the simplest way to do this is by transposing our equations and T = ((PW ) T W ) T = W T PW P = ((T ) T W 1 ) T = (W 1 ) T TW 1 Where T is the compressed image, and P is the original.

77 The Linear Algebra of Image Compression Up to this point we have only been averaging the rows of the entire matrix (which we have called strings). Now we will talk about how to average the columns. What we have computed so far is called the row-reduced form of a matrix. The complete compression will be a row-and-column-reduced form. the simplest way to do this is by transposing our equations and T = ((PW ) T W ) T = W T PW P = ((T ) T W 1 ) T = (W 1 ) T TW 1 Where T is the compressed image, and P is the original. If we normalized the columns of W we could simplify this even more, because the columns would then be orthonormal, making W an orthogonal matrix, thus W 1 = W T.

78 The Linear Algebra of Image Compression Then the equation for P would be P = ((T ) T W 1 o ) T = (Wo 1 ) T TWo 1 = (Wo T ) T TWo T = W o TWo T

79 The Linear Algebra of Image Compression Then the equation for P would be P = ((T ) T W 1 o ) T = (Wo 1 ) T TWo 1 = (Wo T ) T TWo T = W o TWo T This is now much easier to calculate than before, which would optimize the use of a computer s capacity if it was to be included in an algorithm.

80 Questions Any questions?

81 Bibliography Colm Mulcahy. Image Compression Using the Haar Wavelet Transform

Image Compression Using the Haar Wavelet Transform

Image Compression Using the Haar Wavelet Transform College of the Redwoods http://online.redwoods.cc.ca.us/instruct/darnold/laproj/fall2002/ames/ 1/33 Image Compression Using the Haar Wavelet Transform Greg Ames College of the Redwoods Math 45 Linear Algebra

More information

Image Compression. 1. Introduction. Greg Ames Dec 07, 2002

Image Compression. 1. Introduction. Greg Ames Dec 07, 2002 Image Compression Greg Ames Dec 07, 2002 Abstract Digital images require large amounts of memory to store and, when retrieved from the internet, can take a considerable amount of time to download. The

More information

The Haar Wavelet Transform: Compression and. Reconstruction

The Haar Wavelet Transform: Compression and. Reconstruction The Haar Wavelet Transform: Compression and Damien Adams and Halsey Patterson December 14, 2006 Abstract The Haar Wavelet Transformation is a simple form of compression involved in averaging and differencing

More information

The Haar Wavelet Transform: Compression and Reconstruction

The Haar Wavelet Transform: Compression and Reconstruction and Reconstruction December 13, 2006 Have you ever looked at an image on your computer? Of course you have. Images today aren t just stored on rolls of film. Most images today are stored or compressed

More information

Matrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1

Matrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1 Matrix notation A nm : n m : size of the matrix m : no of columns, n: no of rows Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1 n = m square matrix Symmetric matrix Upper triangular matrix: matrix

More information

Learning goals: students learn to use the SVD to find good approximations to matrices and to compute the pseudoinverse.

Learning goals: students learn to use the SVD to find good approximations to matrices and to compute the pseudoinverse. Application of the SVD: Compression and Pseudoinverse Learning goals: students learn to use the SVD to find good approximations to matrices and to compute the pseudoinverse. Low rank approximation One

More information

Matrix operations Linear Algebra with Computer Science Application

Matrix operations Linear Algebra with Computer Science Application Linear Algebra with Computer Science Application February 14, 2018 1 Matrix operations 11 Matrix operations If A is an m n matrix that is, a matrix with m rows and n columns then the scalar entry in the

More information

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1 1 Rows first, columns second. Remember that. R then C. 1 A matrix is a set of real or complex numbers arranged in a rectangular array. They can be any size and shape (provided they are rectangular). A

More information

Matrix Algebra. Learning Objectives. Size of Matrix

Matrix Algebra. Learning Objectives. Size of Matrix Matrix Algebra 1 Learning Objectives 1. Find the sum and difference of two matrices 2. Find scalar multiples of a matrix 3. Find the product of two matrices 4. Find the inverse of a matrix 5. Solve a system

More information

Multi-Resolution Analysis for the Haar Wavelet: A Minimalist Approach

Multi-Resolution Analysis for the Haar Wavelet: A Minimalist Approach Multi-Resolution Analysis for the Haar Wavelet: A Minimalist Approach Helmut Knaust Department of Mathematical Sciences The University of Texas at El Paso El Paso TX 79968-0514 hknaust@utep.edu Joint Mathematics

More information

7.3. Determinants. Introduction. Prerequisites. Learning Outcomes

7.3. Determinants. Introduction. Prerequisites. Learning Outcomes Determinants 7.3 Introduction Among other uses, determinants allow us to determine whether a system of linear equations has a unique solution or not. The evaluation of a determinant is a key skill in engineering

More information

SNAP Centre Workshop. Solving Systems of Equations

SNAP Centre Workshop. Solving Systems of Equations SNAP Centre Workshop Solving Systems of Equations 35 Introduction When presented with an equation containing one variable, finding a solution is usually done using basic algebraic manipulation. Example

More information

TOPIC 2 Computer application for manipulating matrix using MATLAB

TOPIC 2 Computer application for manipulating matrix using MATLAB YOGYAKARTA STATE UNIVERSITY MATHEMATICS AND NATURAL SCIENCES FACULTY MATHEMATICS EDUCATION STUDY PROGRAM TOPIC 2 Computer application for manipulating matrix using MATLAB Definition of Matrices in MATLAB

More information

An overview of key ideas

An overview of key ideas An overview of key ideas This is an overview of linear algebra given at the start of a course on the mathematics of engineering. Linear algebra progresses from vectors to matrices to subspaces. Vectors

More information

Clock Arithmetic. 1. If it is 9 o clock and you get out of school in 4 hours, when do you get out of school?

Clock Arithmetic. 1. If it is 9 o clock and you get out of school in 4 hours, when do you get out of school? Clock Arithmetic We are going to learn all about clock addition and the relationship to remainders when you divide numbers. 1 Standard Clock Addition 1. If it is 9 o clock and you get out of school in

More information

Deep Learning Book Notes Chapter 2: Linear Algebra

Deep Learning Book Notes Chapter 2: Linear Algebra Deep Learning Book Notes Chapter 2: Linear Algebra Compiled By: Abhinaba Bala, Dakshit Agrawal, Mohit Jain Section 2.1: Scalars, Vectors, Matrices and Tensors Scalar Single Number Lowercase names in italic

More information

Fractions. Review R.7. Dr. Doug Ensley. January 7, Dr. Doug Ensley Review R.7

Fractions. Review R.7. Dr. Doug Ensley. January 7, Dr. Doug Ensley Review R.7 Review R.7 Dr. Doug Ensley January 7, 2015 Equivalence of fractions As long as c 0, a b = a c b c Equivalence of fractions As long as c 0, a b = a c b c Examples True or False? 10 18 = 2 5 2 9 = 5 9 10

More information

10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections )

10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections ) c Dr. Igor Zelenko, Fall 2017 1 10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections 7.2-7.4) 1. When each of the functions F 1, F 2,..., F n in right-hand side

More information

Answer Key for Exam #2

Answer Key for Exam #2 . Use elimination on an augmented matrix: Answer Key for Exam # 4 4 8 4 4 4 The fourth column has no pivot, so x 4 is a free variable. The corresponding system is x + x 4 =, x =, x x 4 = which we solve

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

Wavelet Methods

Wavelet Methods -- - - - ~ - ::. - --::i 180 4. Wavelet Methods clear j % main program filename='lena128'; dim=128; fid=fopen(filename,'r')j if fid==-l disp('file not found') else img=fread(fid,[dim,dim])'; fclose(fid);

More information

Lesson U2.1 Study Guide

Lesson U2.1 Study Guide Lesson U2.1 Study Guide Sunday, June 3, 2018 2:05 PM Matrix operations, The Inverse of a Matrix and Matrix Factorization Reading: Lay, Sections 2.1, 2.2, 2.3 and 2.5 (about 24 pages). MyMathLab: Lesson

More information

Basic Linear Algebra. Florida State University. Acknowledgements: Daniele Panozzo. CAP Computer Graphics - Fall 18 Xifeng Gao

Basic Linear Algebra. Florida State University. Acknowledgements: Daniele Panozzo. CAP Computer Graphics - Fall 18 Xifeng Gao Basic Linear Algebra Acknowledgements: Daniele Panozzo Overview We will briefly overview the basic linear algebra concepts that we will need in the class You will not be able to follow the next lectures

More information

Linear Algebra and Matrices

Linear Algebra and Matrices Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.

More information

Lecture 3: Matrix and Matrix Operations

Lecture 3: Matrix and Matrix Operations Lecture 3: Matrix and Matrix Operations Representation, row vector, column vector, element of a matrix. Examples of matrix representations Tables and spreadsheets Scalar-Matrix operation: Scaling a matrix

More information

Properties of Real Numbers

Properties of Real Numbers Properties of Real Numbers Essential Understanding. Relationships that are always true for real numbers are called properties, which are rules used to rewrite and compare expressions. Two algebraic expressions

More information

Expressions that always have the same value. The Identity Property of Addition states that For any value a; a + 0 = a so = 3

Expressions that always have the same value. The Identity Property of Addition states that For any value a; a + 0 = a so = 3 Name Key Words/Topic 2.1 Identity and Zero Properties Topic 2 Guided Notes Equivalent Expressions Identity Property of Addition Identity Property of Multiplication Zero Property of Multiplication The sum

More information

Module 7:Data Representation Lecture 35: Wavelets. The Lecture Contains: Wavelets. Discrete Wavelet Transform (DWT) Haar wavelets: Example

Module 7:Data Representation Lecture 35: Wavelets. The Lecture Contains: Wavelets. Discrete Wavelet Transform (DWT) Haar wavelets: Example The Lecture Contains: Wavelets Discrete Wavelet Transform (DWT) Haar wavelets: Example Haar wavelets: Theory Matrix form Haar wavelet matrices Dimensionality reduction using Haar wavelets file:///c /Documents%20and%20Settings/iitkrana1/My%20Documents/Google%20Talk%20Received%20Files/ist_data/lecture35/35_1.htm[6/14/2012

More information

Roberto s Notes on Linear Algebra Chapter 9: Orthogonality Section 2. Orthogonal matrices

Roberto s Notes on Linear Algebra Chapter 9: Orthogonality Section 2. Orthogonal matrices Roberto s Notes on Linear Algebra Chapter 9: Orthogonality Section 2 Orthogonal matrices What you need to know already: What orthogonal and orthonormal bases for subspaces are. What you can learn here:

More information

Tutorial 2 - Learning about the Discrete Fourier Transform

Tutorial 2 - Learning about the Discrete Fourier Transform Tutorial - Learning about the Discrete Fourier Transform This tutorial will be about the Discrete Fourier Transform basis, or the DFT basis in short. What is a basis? If we google define basis, we get:

More information

Singular Value Decomposition and Digital Image Compression

Singular Value Decomposition and Digital Image Compression Singular Value Decomposition and Digital Image Compression Chris Bingham December 1, 016 Page 1 of Abstract The purpose of this document is to be a very basic introduction to the singular value decomposition

More information

Math 671: Tensor Train decomposition methods II

Math 671: Tensor Train decomposition methods II Math 671: Tensor Train decomposition methods II Eduardo Corona 1 1 University of Michigan at Ann Arbor December 13, 2016 Table of Contents 1 What we ve talked about so far: 2 The Tensor Train decomposition

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =

More information

Image and Multidimensional Signal Processing

Image and Multidimensional Signal Processing Image and Multidimensional Signal Processing Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ Image Compression 2 Image Compression Goal: Reduce amount

More information

Discrete Wavelet Transform

Discrete Wavelet Transform Discrete Wavelet Transform [11] Kartik Mehra July 2017 Math 190s Duke University "1 Introduction Wavelets break signals up and then analyse them separately with a resolution that is matched with scale.

More information

Digital Design for Multiplication

Digital Design for Multiplication Digital Design for Multiplication Norman Matloff October 15, 2003 c 2003, N.S. Matloff 1 Overview A cottage industry exists in developing fast digital logic to perform arithmetic computations. Fast addition,

More information

Data representation by field calculus and leading to the orthonormal linear transforms

Data representation by field calculus and leading to the orthonormal linear transforms International Journal of Applied Electromagnetics and Mechanics 19 (2004) 153 157 153 IOS Press Data representation by field calculus and leading to the orthonormal linear transforms Hisashi Endo a,, Yoshihiro

More information

Review of linear algebra

Review of linear algebra Review of linear algebra 1 Vectors and matrices We will just touch very briefly on certain aspects of linear algebra, most of which should be familiar. Recall that we deal with vectors, i.e. elements of

More information

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................

More information

Example Linear Algebra Competency Test

Example Linear Algebra Competency Test Example Linear Algebra Competency Test The 4 questions below are a combination of True or False, multiple choice, fill in the blank, and computations involving matrices and vectors. In the latter case,

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

DISCRETE HAAR WAVELET TRANSFORMS

DISCRETE HAAR WAVELET TRANSFORMS DISCRETE HAAR WAVELET TRANSFORMS Catherine Bénéteau University of South Florida Tampa, FL USA UNM - PNM Statewide Mathematics Contest, 2011 SATURDAY, FEBRUARY 5, 2011 (UNM) DISCRETE HAAR WAVELET TRANSFORMS

More information

Matrix Algebra 2.1 MATRIX OPERATIONS Pearson Education, Inc.

Matrix Algebra 2.1 MATRIX OPERATIONS Pearson Education, Inc. 2 Matrix Algebra 2.1 MATRIX OPERATIONS MATRIX OPERATIONS m n If A is an matrixthat is, a matrix with m rows and n columnsthen the scalar entry in the ith row and jth column of A is denoted by a ij and

More information

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition

More information

College Algebra. Course Text. Course Description. Course Objectives. StraighterLine MAT101: College Algebra

College Algebra. Course Text. Course Description. Course Objectives. StraighterLine MAT101: College Algebra College Algebra Course Text Barnett, Raymond A., Michael R. Ziegler, and Karl E. Byleen. College Algebra, 8th edition, McGraw-Hill, 2008, ISBN: 9780072867381 [This text is available as an etextbook at

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Linear Algebra II. 7 Inner product spaces. Notes 7 16th December Inner products and orthonormal bases

Linear Algebra II. 7 Inner product spaces. Notes 7 16th December Inner products and orthonormal bases MTH6140 Linear Algebra II Notes 7 16th December 2010 7 Inner product spaces Ordinary Euclidean space is a 3-dimensional vector space over R, but it is more than that: the extra geometric structure (lengths,

More information

Vector/Matrix operations. *Remember: All parts of HW 1 are due on 1/31 or 2/1

Vector/Matrix operations. *Remember: All parts of HW 1 are due on 1/31 or 2/1 Lecture 4: Topics: Linear Algebra II Vector/Matrix operations Homework: HW, Part *Remember: All parts of HW are due on / or / Solving Axb Row reduction method can be used Simple operations on equations

More information

The Laplace Expansion Theorem: Computing the Determinants and Inverses of Matrices

The Laplace Expansion Theorem: Computing the Determinants and Inverses of Matrices The Laplace Expansion Theorem: Computing the Determinants and Inverses of Matrices David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative

More information

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix

More information

a (b + c) = a b + a c

a (b + c) = a b + a c Chapter 1 Vector spaces In the Linear Algebra I module, we encountered two kinds of vector space, namely real and complex. The real numbers and the complex numbers are both examples of an algebraic structure

More information

Exercise Set Suppose that A, B, C, D, and E are matrices with the following sizes: A B C D E

Exercise Set Suppose that A, B, C, D, and E are matrices with the following sizes: A B C D E Determine the size of a given matrix. Identify the row vectors and column vectors of a given matrix. Perform the arithmetic operations of matrix addition, subtraction, scalar multiplication, and multiplication.

More information

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of Chapter 2 Linear Algebra In this chapter, we study the formal structure that provides the background for quantum mechanics. The basic ideas of the mathematical machinery, linear algebra, are rather simple

More information

Introduction to Data Mining

Introduction to Data Mining Introduction to Data Mining Lecture #21: Dimensionality Reduction Seoul National University 1 In This Lecture Understand the motivation and applications of dimensionality reduction Learn the definition

More information

Factorizing Algebraic Expressions

Factorizing Algebraic Expressions 1 of 60 Factorizing Algebraic Expressions 2 of 60 Factorizing expressions Factorizing an expression is the opposite of expanding it. Expanding or multiplying out a(b + c) ab + ac Factorizing Often: When

More information

Algebra 2 Matrices. Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Find.

Algebra 2 Matrices. Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Find. Algebra 2 Matrices Review Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Find. Evaluate the determinant of the matrix. 2. 3. A matrix contains 48 elements.

More information

Next topics: Solving systems of linear equations

Next topics: Solving systems of linear equations Next topics: Solving systems of linear equations 1 Gaussian elimination (today) 2 Gaussian elimination with partial pivoting (Week 9) 3 The method of LU-decomposition (Week 10) 4 Iterative techniques:

More information

1111: Linear Algebra I

1111: Linear Algebra I 1111: Linear Algebra I Dr. Vladimir Dotsenko (Vlad) Lecture 5 Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 5 1 / 12 Systems of linear equations Geometrically, we are quite used to the fact

More information

1.Chapter Objectives

1.Chapter Objectives LU Factorization INDEX 1.Chapter objectives 2.Overview of LU factorization 2.1GAUSS ELIMINATION AS LU FACTORIZATION 2.2LU Factorization with Pivoting 2.3 MATLAB Function: lu 3. CHOLESKY FACTORIZATION 3.1

More information

Derivation of the Kalman Filter

Derivation of the Kalman Filter Derivation of the Kalman Filter Kai Borre Danish GPS Center, Denmark Block Matrix Identities The key formulas give the inverse of a 2 by 2 block matrix, assuming T is invertible: T U 1 L M. (1) V W N P

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

Least Squares Optimization

Least Squares Optimization Least Squares Optimization The following is a brief review of least squares optimization and constrained optimization techniques. Broadly, these techniques can be used in data analysis and visualization

More information

Day 1: Introduction to Vectors + Vector Arithmetic

Day 1: Introduction to Vectors + Vector Arithmetic Day 1: Introduction to Vectors + Vector Arithmetic A is a quantity that has magnitude but no direction. You can have signed scalar quantities as well. A is a quantity that has both magnitude and direction.

More information

Kevin James. MTHSC 3110 Section 2.1 Matrix Operations

Kevin James. MTHSC 3110 Section 2.1 Matrix Operations MTHSC 3110 Section 2.1 Matrix Operations Notation Let A be an m n matrix, that is, m rows and n columns. We ll refer to the entries of A by their row and column indices. The entry in the i th row and j

More information

MITOCW ocw f99-lec31_300k

MITOCW ocw f99-lec31_300k MITOCW ocw-18.06-f99-lec31_300k OK. So, coming nearer the end of the course, this lecture will be a mixture of the linear algebra that comes with a change of basis. And a change of basis from one basis

More information

Linear Equations in Linear Algebra

Linear Equations in Linear Algebra 1 Linear Equations in Linear Algebra 1.1 SYSTEMS OF LINEAR EQUATIONS LINEAR EQUATION,, 1 n A linear equation in the variables equation that can be written in the form a a a b 1 1 2 2 n n a a is an where

More information

Image Compression by Using Haar Wavelet Transform and Singular Value Decomposition

Image Compression by Using Haar Wavelet Transform and Singular Value Decomposition Master Thesis Image Compression by Using Haar Wavelet Transform and Singular Value Decomposition Zunera Idrees 9--5 Eliza Hashemiaghjekandi 979-- Subject: Mathematics Level: Advance Course code: 4MAE Abstract

More information

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0. ) Find all solutions of the linear system. Express the answer in vector form. x + 2x + x + x 5 = 2 2x 2 + 2x + 2x + x 5 = 8 x + 2x + x + 9x 5 = 2 2 Solution: Reduce the augmented matrix [ 2 2 2 8 ] to

More information

Problem Set # 1 Solution, 18.06

Problem Set # 1 Solution, 18.06 Problem Set # 1 Solution, 1.06 For grading: Each problem worths 10 points, and there is points of extra credit in problem. The total maximum is 100. 1. (10pts) In Lecture 1, Prof. Strang drew the cone

More information

Linear Equations in Linear Algebra

Linear Equations in Linear Algebra 1 Linear Equations in Linear Algebra 1.7 LINEAR INDEPENDENCE LINEAR INDEPENDENCE Definition: An indexed set of vectors {v 1,, v p } in n is said to be linearly independent if the vector equation x x x

More information

Chapter 2 Notes, Linear Algebra 5e Lay

Chapter 2 Notes, Linear Algebra 5e Lay Contents.1 Operations with Matrices..................................1.1 Addition and Subtraction.............................1. Multiplication by a scalar............................ 3.1.3 Multiplication

More information

Math 360 Linear Algebra Fall Class Notes. a a a a a a. a a a

Math 360 Linear Algebra Fall Class Notes. a a a a a a. a a a Math 360 Linear Algebra Fall 2008 9-10-08 Class Notes Matrices As we have already seen, a matrix is a rectangular array of numbers. If a matrix A has m columns and n rows, we say that its dimensions are

More information

22A-2 SUMMER 2014 LECTURE 5

22A-2 SUMMER 2014 LECTURE 5 A- SUMMER 0 LECTURE 5 NATHANIEL GALLUP Agenda Elimination to the identity matrix Inverse matrices LU factorization Elimination to the identity matrix Previously, we have used elimination to get a system

More information

Mon Feb Matrix inverses, the elementary matrix approach overview of skipped section 2.5. Announcements: Warm-up Exercise:

Mon Feb Matrix inverses, the elementary matrix approach overview of skipped section 2.5. Announcements: Warm-up Exercise: Math 2270-004 Week 6 notes We will not necessarily finish the material from a given day's notes on that day We may also add or subtract some material as the week progresses, but these notes represent an

More information

Let p 2 ( t), (2 t k), we have the scaling relation,

Let p 2 ( t), (2 t k), we have the scaling relation, Multiresolution Analysis and Daubechies N Wavelet We have discussed decomposing a signal into its Haar wavelet components of varying frequencies. The Haar wavelet scheme relied on two functions: the Haar

More information

A primer on matrices

A primer on matrices A primer on matrices Stephen Boyd August 4, 2007 These notes describe the notation of matrices, the mechanics of matrix manipulation, and how to use matrices to formulate and solve sets of simultaneous

More information

Matrix Operations. Linear Combination Vector Algebra Angle Between Vectors Projections and Reflections Equality of matrices, Augmented Matrix

Matrix Operations. Linear Combination Vector Algebra Angle Between Vectors Projections and Reflections Equality of matrices, Augmented Matrix Linear Combination Vector Algebra Angle Between Vectors Projections and Reflections Equality of matrices, Augmented Matrix Matrix Operations Matrix Addition and Matrix Scalar Multiply Matrix Multiply Matrix

More information

Roberto s Notes on Linear Algebra Chapter 4: Matrix Algebra Section 4. Matrix products

Roberto s Notes on Linear Algebra Chapter 4: Matrix Algebra Section 4. Matrix products Roberto s Notes on Linear Algebra Chapter 4: Matrix Algebra Section 4 Matrix products What you need to know already: The dot product of vectors Basic matrix operations. Special types of matrices What you

More information

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4 Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix

More information

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Review of matrices. Let m, n IN. A rectangle of numbers written like A = Review of matrices Let m, n IN. A rectangle of numbers written like a 11 a 12... a 1n a 21 a 22... a 2n A =...... a m1 a m2... a mn where each a ij IR is called a matrix with m rows and n columns or an

More information

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007 Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007 You have 1 hour and 20 minutes. No notes, books, or other references. You are permitted to use Maple during this exam, but you must start with a blank

More information

Math Linear Algebra Final Exam Review Sheet

Math Linear Algebra Final Exam Review Sheet Math 15-1 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a component-wise operation. Two vectors v and w may be added together as long as they contain the same number n of

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic A FIRST COURSE IN LINEAR ALGEBRA An Open Text by Ken Kuttler Matrix Arithmetic Lecture Notes by Karen Seyffarth Adapted by LYRYX SERVICE COURSE SOLUTION Attribution-NonCommercial-ShareAlike (CC BY-NC-SA)

More information

Wavelet Transform And Principal Component Analysis Based Feature Extraction

Wavelet Transform And Principal Component Analysis Based Feature Extraction Wavelet Transform And Principal Component Analysis Based Feature Extraction Keyun Tong June 3, 2010 As the amount of information grows rapidly and widely, feature extraction become an indispensable technique

More information

7.4. The Inverse of a Matrix. Introduction. Prerequisites. Learning Outcomes

7.4. The Inverse of a Matrix. Introduction. Prerequisites. Learning Outcomes The Inverse of a Matrix 7.4 Introduction In number arithmetic every number a 0has a reciprocal b written as a or such that a ba = ab =. Similarly a square matrix A may have an inverse B = A where AB =

More information

Stage-structured Populations

Stage-structured Populations Department of Biology New Mexico State University Las Cruces, New Mexico 88003 brook@nmsu.edu Fall 2009 Age-Structured Populations All individuals are not equivalent to each other Rates of survivorship

More information

Module 4 MULTI- RESOLUTION ANALYSIS. Version 2 ECE IIT, Kharagpur

Module 4 MULTI- RESOLUTION ANALYSIS. Version 2 ECE IIT, Kharagpur Module MULTI- RESOLUTION ANALYSIS Version ECE IIT, Kharagpur Lesson Multi-resolution Analysis: Theory of Subband Coding Version ECE IIT, Kharagpur Instructional Objectives At the end of this lesson, the

More information

Counting in Different Number Systems

Counting in Different Number Systems Counting in Different Number Systems Base 1 (Decimal) is important because that is the base that we first learn in our culture. Base 2 (Binary) is important because that is the base used for computer codes

More information

Digital Image Processing

Digital Image Processing Digital Image Processing Image Transforms Unitary Transforms and the 2D Discrete Fourier Transform DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON What is this

More information

Elementary Linear Algebra

Elementary Linear Algebra Elementary Linear Algebra Linear algebra is the study of; linear sets of equations and their transformation properties. Linear algebra allows the analysis of; rotations in space, least squares fitting,

More information

Mathematics for Graphics and Vision

Mathematics for Graphics and Vision Mathematics for Graphics and Vision Steven Mills March 3, 06 Contents Introduction 5 Scalars 6. Visualising Scalars........................ 6. Operations on Scalars...................... 6.3 A Note on

More information

Section Summary. Definition of a Function.

Section Summary. Definition of a Function. Section 2.3 Section Summary Definition of a Function. Domain, Codomain Image, Preimage Injection, Surjection, Bijection Inverse Function Function Composition Graphing Functions Floor, Ceiling, Factorial

More information

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition Prof. Tesler Math 283 Fall 2016 Also see the separate version of this with Matlab and R commands. Prof. Tesler Diagonalizing

More information

8 The SVD Applied to Signal and Image Deblurring

8 The SVD Applied to Signal and Image Deblurring 8 The SVD Applied to Signal and Image Deblurring We will discuss the restoration of one-dimensional signals and two-dimensional gray-scale images that have been contaminated by blur and noise. After an

More information

Linear Algebra Tutorial for Math3315/CSE3365 Daniel R. Reynolds

Linear Algebra Tutorial for Math3315/CSE3365 Daniel R. Reynolds Linear Algebra Tutorial for Math3315/CSE3365 Daniel R. Reynolds These notes are meant to provide a brief introduction to the topics from Linear Algebra that will be useful in Math3315/CSE3365, Introduction

More information

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2

More information

ACE Transfer Credit Packet Your Guide to Earning ACE Credit for StraighterLine Courses

ACE Transfer Credit Packet Your Guide to Earning ACE Credit for StraighterLine Courses ACE Transfer Credit Packet Your Guide to Earning ACE Credit for StraighterLine Courses What You Need To Know Before You Start Finding out if your college or university will award credit for a course at

More information