Numerical Methods. Rafał Zdunek Underdetermined problems (2h.) Applications) (FOCUSS, M-FOCUSS,

Size: px
Start display at page:

Download "Numerical Methods. Rafał Zdunek Underdetermined problems (2h.) Applications) (FOCUSS, M-FOCUSS,"

Transcription

1 Numerical Methods Rafał Zdunek Underdetermined problems (h.) (FOCUSS, M-FOCUSS, M Applications)

2 Introduction Solutions to underdetermined linear systems, Morphological constraints, FOCUSS algorithm, M-FOCUSS algorithm.

3 Bibliography [] A. Borck, Numerical Methods for Least Squares Problems, SIAM, Philadelphia, 996, [] G. Golub, C. F. Van Loan, Matrix Computations, he John Hopkins University Press, (hird Edition), 996, [3] I. F. Gorodnitsky and B. D. Rao, Sparse signal reconstructions from limited data using FOCUSS: A re-weighted minimum norm algorithm, IEEE rans. Signal Process., vol. 45, no. 3, pp , Mar. 997, [4] B. D. Rao and K. Kreutz-Delgado, Deriving algorithms for computing sparse solutions to linear inverse problems, in Proc. 3st Asilomar Conf. Signals Syst. Comput., CA, Nov. 5, 997, vol., pp , [5] B. D. Rao, K. Engan, S. F. Cotter, J. Palmer, and K. Kreutz-Delgado, Subset selection in noise based on diversity measure minimization, IEEE rans. Signal Process., vol. 5, no. 3, pp , Mar. 003, [6] S. F. Cotter, B. Rao, K. Engan, and K. Kreutz-Delgado, Sparse solutions to linear inverse problems with multiple measurement vectors, IEEE rans. Signal Process., vol. 53, no. 7, pp , Jul. 005,

4 Solutions to linear systems A system of linear equations can be expressed in the following matrix form: M N where = [ a i ] I N = [ x ] I Ax = b, () A is a coefficient matrix, [ ] M M ( N +) x is a solution to be estimated. Let [ ] b = b I is a data vector, and i B = A b = I be the augmented matrix to the system (). he system of linear equations may behave in any one of three possible ways: A he system has no solution if ( A) rank( B) rank <. B he system has a single unique solution if rank ( ) = rank( B) = N A. C he system has infinitely many solutions if rank ( ) = rank( B) < N A.

5 Solutions to linear systems he case C occurs for rank-deficient problems or under-determined problems. In spite of infinitely many solutions, a good approximation to a true solution can be obtained if some a priori knowledge about the nature of the true solution is accessible. he additional constraints are usually concerned with a degree of sparsity or smoothness of the true solution.

6 RREF Basic variables Free variables

7 Example x + 3x + x = a x x + x3 = b 3x+ 7x x3 = c Let: 3 [ A b] (Basic variables) (Free variable) 3 a 0 5 a 3b Gauss-Jordan = b 0 b+ a 3 7 c c a+ b he system is consistent (it has solutions) if c a+ b= 0 c= a b Solution: x 3 = free variable, x = a+ b x, x = a 3b+ 5 x3. 3

8 Example Let: x+ x + x3+ 3x4 = 4 x+ 4x + x3+ 3x4 = 5 3x+ 6x + x3+ 4x4 = 7 [ A b] Gauss-Jordan = x = x + x, x = free variable, Solution: 4 Consistent x = x, x 4 = free variable. 3 4

9 Example hus: x x x4 x x 0 0 = = + x + x 4 x 3 x 4 0 x x General solution Particular solution xgeneral = xparticular + xhomogeneous Homogenous solution N(A) he homogeneous solution is a solution to the system Ax = 0. Problem: How to choose free variables? Remark: Replacing free variables with zero-values is not always a good solution!

10 Sparseness constraints Regularized least-squares problem: min x { ( ) ( )} p Ax b +γ E x L p diversity measure (Gorodnitsky, Rao, 997): E ( p) N ( ) = sgn( p) x x p. = p

11 FOCUSS algorithm Let Stationary point: ( p) ( ) J( x) = Ax b +γ E x ( ) J ( x ) = A Ax A b+ λw x x = 0 x * * * * ( ) { p/ = x } W x* diag, p λ = γ Hence: ( ) ( ) Iterative updates: ( λ ) M x W x A AW x A I b * = * * +. ( λ ) ( ) ( ) x W x A AW x A I b = + k+ k k M.

12 FOCUSS algorithm [ 0,] p, λ > 0 (0) Randomly initialize x, For k = 0,,, until convergence do { / } ( k ) p x W = diag, k ( ) λ x = W A AW A + I b, ( k+ ) k k M

13 Wiener filtered FOCUSS [ 0,] p, λ > 0 algorithm ( 0 ) Randomly initialize x, For k = 0,,, until convergence do { / } ( k ) p x W = diag, k ( ) λ x = W A AW A + I b, ( k+ ) k k M x ( ( k ) ) F x ( k + ) + R. Zdunek, Z. He, A. Cichocki, Proc. IEEE ISBI, 008

14 Wiener filtered FOCUSS algorithm Markov Random Field (MRF) μ σ = ( k + ) x n L n = N ( k + ) ( ) x n L n N First and second-order interactions N μ = h,, +, + h, L = 9 h, + h, h +, + h + - local mean around -th pixel - local variance

15 Wiener filtered FOCUSS algorithm Update rule Mean noise variance x ν σ ν ( k+ ) ) μ + σ N = N = σ ( ( k+ x μ )

16 Limited-view tomographic imaging ( A ) rank = LSS ( ) 3 ( A ; b ) = { x R N : Ax b = min! } ( A; b) x N ( A) 0 0 α 0 0 α = 0 α α A 5 ( ) α = (Rank-deficient) { [ ] } N( A( ) ) = span,,,, LSS = LS + where x LS = x r = P ( ) x exact R A

17 omographic imaging example Phantom image Minimal l norm least squares solution: (LS algorithms: AR, SIR)

18 omographic imaging (noise-free data) FOCUSS algorithm (p =, k = 5, ) λ =0 8 Wiener filtered FOCUSS algorithm (p =, k = 5, ) λ =0 8

19 omographic imaging (noisy data SNR = 30 db) FOCUSS algorithm (p =, k = 5, = 40 ) Wiener filtered FOCUSS algorithm (p =, k = 5, ) λ λ = 40

20 omographic imaging (Normalized RMSE, noise-free)

21 omographic imaging (Normalized RMSE, p =, noisy data)

22 M-FOCUSS M N Let AX + N = B where R, A X [ x x ] ( A) M N, (under-determined) rank = M, N,,, N, M = R B R, > N [ n n ],, M =. R (Cotter, Rao, Engan, Kreutz-Delgado, 005) (additive noise) If M < N, the nullspace of A is non-trivial, thus additional constraints are necessary to select the right solution. he M-FOCUSS assumes sparse solutions. heorem: he sparse solution to the consistent system AX = B is unique, ( ) rank B =, M, if with any M columns of A are linearly independent (unique representation property (URP) condition), and for each t: x t has at most ( M + )/ nonzero entries, where is a ceil function.

23 M-FOCUSS he M-FOCUSS algorithm iteratively solves the following equality constrained problem: where p J ( ) ( ) ( p) min J, X N N ( ) p x t = = t= X s.t. AX = B, p/ X = x =. - l diversity measure for sparsity p 0 p degree of sparsity x - the -th row of X (LS solution with minimal l -norm) (l 0 norm solution NP-hard problem)

24 M-FOCUSS For inconsistent data, i.e. B R(A), Cotter at al. developped the regularized M-FOCUSS algorithm that solves the ikhonov regularized least-squares problem in a single iterative step: ( k) ( k ) k X = arg min Ψ( X X ), where ( ) X p/ diag ( w / ( k ) W = ), with w = ( xt ) = ( x ) Regularized M-FOCUSS: For k =,,... A ( k+ ) ( k = AW ), ( ) ( ) t= ( ) λ M ( k+ ) ( k ) ( k ) ( k ) ( k X = W A A A ) + I B ( ) Ψ X X = B AX + λ W X, ( k ) λ > 0 regularization parameter. F. F

Chapter 1. Vectors, Matrices, and Linear Spaces

Chapter 1. Vectors, Matrices, and Linear Spaces 1.6 Homogeneous Systems, Subspaces and Bases 1 Chapter 1. Vectors, Matrices, and Linear Spaces 1.6. Homogeneous Systems, Subspaces and Bases Note. In this section we explore the structure of the solution

More information

A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT

A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Progress In Electromagnetics Research Letters, Vol. 16, 53 60, 2010 A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Y. P. Liu and Q. Wan School of Electronic Engineering University of Electronic

More information

EXAM. Exam #2. Math 2360 Summer II, 2000 Morning Class. Nov. 15, 2000 ANSWERS

EXAM. Exam #2. Math 2360 Summer II, 2000 Morning Class. Nov. 15, 2000 ANSWERS EXAM Exam # Math 6 Summer II Morning Class Nov 5 ANSWERS i Problem Consider the matrix 6 pts A = 6 4 9 5 7 6 5 5 5 4 The RREF of A is the matrix R = A Find a basis for the nullspace of A Solve the homogeneous

More information

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0. ) Find all solutions of the linear system. Express the answer in vector form. x + 2x + x + x 5 = 2 2x 2 + 2x + 2x + x 5 = 8 x + 2x + x + 9x 5 = 2 2 Solution: Reduce the augmented matrix [ 2 2 2 8 ] to

More information

Lecture 21: 5.6 Rank and Nullity

Lecture 21: 5.6 Rank and Nullity Lecture 21: 5.6 Rank and Nullity Wei-Ta Chu 2008/12/5 Rank and Nullity Definition The common dimension of the row and column space of a matrix A is called the rank ( 秩 ) of A and is denoted by rank(a);

More information

Row Reduction and Echelon Forms

Row Reduction and Echelon Forms Row Reduction and Echelon Forms 1 / 29 Key Concepts row echelon form, reduced row echelon form pivot position, pivot, pivot column basic variable, free variable general solution, parametric solution existence

More information

MATH 2360 REVIEW PROBLEMS

MATH 2360 REVIEW PROBLEMS MATH 2360 REVIEW PROBLEMS Problem 1: In (a) (d) below, either compute the matrix product or indicate why it does not exist: ( )( ) 1 2 2 1 (a) 0 1 1 2 ( ) 0 1 2 (b) 0 3 1 4 3 4 5 2 5 (c) 0 3 ) 1 4 ( 1

More information

Midterm 1 Review. Written by Victoria Kala SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015

Midterm 1 Review. Written by Victoria Kala SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015 Midterm 1 Review Written by Victoria Kala vtkala@math.ucsb.edu SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015 Summary This Midterm Review contains notes on sections 1.1 1.5 and 1.7 in your

More information

A Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems

A Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems A Continuation Approach to Estimate a Solution Path of Mixed L2-L Minimization Problems Junbo Duan, Charles Soussen, David Brie, Jérôme Idier Centre de Recherche en Automatique de Nancy Nancy-University,

More information

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 12, DECEMBER 2008 2009 Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation Yuanqing Li, Member, IEEE, Andrzej Cichocki,

More information

Solutions of Linear system, vector and matrix equation

Solutions of Linear system, vector and matrix equation Goals: Solutions of Linear system, vector and matrix equation Solutions of linear system. Vectors, vector equation. Matrix equation. Math 112, Week 2 Suggested Textbook Readings: Sections 1.3, 1.4, 1.5

More information

Simultaneous Sparsity

Simultaneous Sparsity Simultaneous Sparsity Joel A. Tropp Anna C. Gilbert Martin J. Strauss {jtropp annacg martinjs}@umich.edu Department of Mathematics The University of Michigan 1 Simple Sparse Approximation Work in the d-dimensional,

More information

SECTION 3.3. PROBLEM 22. The null space of a matrix A is: N(A) = {X : AX = 0}. Here are the calculations of AX for X = a,b,c,d, and e. =

SECTION 3.3. PROBLEM 22. The null space of a matrix A is: N(A) = {X : AX = 0}. Here are the calculations of AX for X = a,b,c,d, and e. = SECTION 3.3. PROBLEM. The null space of a matrix A is: N(A) {X : AX }. Here are the calculations of AX for X a,b,c,d, and e. Aa [ ][ ] 3 3 [ ][ ] Ac 3 3 [ ] 3 3 [ ] 4+4 6+6 Ae [ ], Ab [ ][ ] 3 3 3 [ ]

More information

Lecture 3: Linear Algebra Review, Part II

Lecture 3: Linear Algebra Review, Part II Lecture 3: Linear Algebra Review, Part II Brian Borchers January 4, Linear Independence Definition The vectors v, v,..., v n are linearly independent if the system of equations c v + c v +...+ c n v n

More information

Improved FOCUSS Method With Conjugate Gradient Iterations

Improved FOCUSS Method With Conjugate Gradient Iterations IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 1, JANUARY 2009 399 Improved FOCUSS Method With Conjugate Gradient Iterations Zhaoshui He, Andrzej Cichocki, Rafal Zdunek, and Shengli Xie Abstract

More information

Linear Algebra I Lecture 10

Linear Algebra I Lecture 10 Linear Algebra I Lecture 10 Xi Chen 1 1 University of Alberta January 30, 2019 Outline 1 Gauss-Jordan Algorithm ] Let A = [a ij m n be an m n matrix. To reduce A to a reduced row echelon form using elementary

More information

MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix.

MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix. MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix. Basis Definition. Let V be a vector space. A linearly independent spanning set for V is called a basis.

More information

DM559 Linear and Integer Programming. Lecture 2 Systems of Linear Equations. Marco Chiarandini

DM559 Linear and Integer Programming. Lecture 2 Systems of Linear Equations. Marco Chiarandini DM559 Linear and Integer Programming Lecture Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. Outline 1. 3 A Motivating Example You are organizing

More information

Linear equations in linear algebra

Linear equations in linear algebra Linear equations in linear algebra Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra Pearson Collections Samy T. Linear

More information

Math 314H EXAM I. 1. (28 points) The row reduced echelon form of the augmented matrix for the system. is the matrix

Math 314H EXAM I. 1. (28 points) The row reduced echelon form of the augmented matrix for the system. is the matrix Math 34H EXAM I Do all of the problems below. Point values for each of the problems are adjacent to the problem number. Calculators may be used to check your answer but not to arrive at your answer. That

More information

Math 2030 Assignment 5 Solutions

Math 2030 Assignment 5 Solutions Math 030 Assignment 5 Solutions Question 1: Which of the following sets of vectors are linearly independent? If the set is linear dependent, find a linear dependence relation for the vectors (a) {(1, 0,

More information

Chapter 1. Vectors, Matrices, and Linear Spaces

Chapter 1. Vectors, Matrices, and Linear Spaces 1.4 Solving Systems of Linear Equations 1 Chapter 1. Vectors, Matrices, and Linear Spaces 1.4. Solving Systems of Linear Equations Note. We give an algorithm for solving a system of linear equations (called

More information

The definition of a vector space (V, +, )

The definition of a vector space (V, +, ) The definition of a vector space (V, +, ) 1. For any u and v in V, u + v is also in V. 2. For any u and v in V, u + v = v + u. 3. For any u, v, w in V, u + ( v + w) = ( u + v) + w. 4. There is an element

More information

b for the linear system x 1 + x 2 + a 2 x 3 = a x 1 + x 3 = 3 x 1 + x 2 + 9x 3 = 3 ] 1 1 a 2 a

b for the linear system x 1 + x 2 + a 2 x 3 = a x 1 + x 3 = 3 x 1 + x 2 + 9x 3 = 3 ] 1 1 a 2 a Practice Exercises for Exam Exam will be on Monday, September 8, 7. The syllabus for Exam consists of Sections One.I, One.III, Two.I, and Two.II. You should know the main definitions, results and computational

More information

Solutions to Homework 5 - Math 3410

Solutions to Homework 5 - Math 3410 Solutions to Homework 5 - Math 34 (Page 57: # 489) Determine whether the following vectors in R 4 are linearly dependent or independent: (a) (, 2, 3, ), (3, 7,, 2), (, 3, 7, 4) Solution From x(, 2, 3,

More information

Lecture 2 Systems of Linear Equations and Matrices, Continued

Lecture 2 Systems of Linear Equations and Matrices, Continued Lecture 2 Systems of Linear Equations and Matrices, Continued Math 19620 Outline of Lecture Algorithm for putting a matrix in row reduced echelon form - i.e. Gauss-Jordan Elimination Number of Solutions

More information

Chapter 3. Directions: For questions 1-11 mark each statement True or False. Justify each answer.

Chapter 3. Directions: For questions 1-11 mark each statement True or False. Justify each answer. Chapter 3 Directions: For questions 1-11 mark each statement True or False. Justify each answer. 1. (True False) Asking whether the linear system corresponding to an augmented matrix [ a 1 a 2 a 3 b ]

More information

Type II variational methods in Bayesian estimation

Type II variational methods in Bayesian estimation Type II variational methods in Bayesian estimation J. A. Palmer, D. P. Wipf, and K. Kreutz-Delgado Department of Electrical and Computer Engineering University of California San Diego, La Jolla, CA 9093

More information

1. Determine by inspection which of the following sets of vectors is linearly independent. 3 3.

1. Determine by inspection which of the following sets of vectors is linearly independent. 3 3. 1. Determine by inspection which of the following sets of vectors is linearly independent. (a) (d) 1, 3 4, 1 { [ [,, 1 1] 3]} (b) 1, 4 5, (c) 3 6 (e) 1, 3, 4 4 3 1 4 Solution. The answer is (a): v 1 is

More information

MATH 2050 Assignment 6 Fall 2018 Due: Thursday, November 1. x + y + 2z = 2 x + y + z = c 4x + 2z = 2

MATH 2050 Assignment 6 Fall 2018 Due: Thursday, November 1. x + y + 2z = 2 x + y + z = c 4x + 2z = 2 MATH 5 Assignment 6 Fall 8 Due: Thursday, November [5]. For what value of c does have a solution? Is it unique? x + y + z = x + y + z = c 4x + z = Writing the system as an augmented matrix, we have c R

More information

Pseudoinverse & Moore-Penrose Conditions

Pseudoinverse & Moore-Penrose Conditions ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 1/1 Lecture 7 ECE 275A Pseudoinverse & Moore-Penrose Conditions ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego

More information

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 222 3. M Test # July, 23 Solutions. For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For

More information

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible. MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:

More information

Linear Equations in Linear Algebra

Linear Equations in Linear Algebra Linear Equations in Linear Algebra.7 LINEAR INDEPENDENCE LINEAR INDEPENDENCE Definition: An indexed set of vectors {v,, v p } in n is said to be linearly independent if the vector equation x x x 2 2 p

More information

EUSIPCO

EUSIPCO EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,

More information

Find the solution set of 2x 3y = 5. Answer: We solve for x = (5 + 3y)/2. Hence the solution space consists of all vectors of the form

Find the solution set of 2x 3y = 5. Answer: We solve for x = (5 + 3y)/2. Hence the solution space consists of all vectors of the form Math 2 Homework #7 March 4, 2 7.3.3. Find the solution set of 2x 3y = 5. Answer: We solve for x = (5 + 3y/2. Hence the solution space consists of all vectors of the form ( ( ( ( x (5 + 3y/2 5/2 3/2 x =

More information

Non-Negative Matrix Factorization with Quasi-Newton Optimization

Non-Negative Matrix Factorization with Quasi-Newton Optimization Non-Negative Matrix Factorization with Quasi-Newton Optimization Rafal ZDUNEK, Andrzej CICHOCKI Laboratory for Advanced Brain Signal Processing BSI, RIKEN, Wako-shi, JAPAN Abstract. Non-negative matrix

More information

Review Notes for Linear Algebra True or False Last Updated: February 22, 2010

Review Notes for Linear Algebra True or False Last Updated: February 22, 2010 Review Notes for Linear Algebra True or False Last Updated: February 22, 2010 Chapter 4 [ Vector Spaces 4.1 If {v 1,v 2,,v n } and {w 1,w 2,,w n } are linearly independent, then {v 1 +w 1,v 2 +w 2,,v n

More information

SIGNAL SEPARATION USING RE-WEIGHTED AND ADAPTIVE MORPHOLOGICAL COMPONENT ANALYSIS

SIGNAL SEPARATION USING RE-WEIGHTED AND ADAPTIVE MORPHOLOGICAL COMPONENT ANALYSIS TR-IIS-4-002 SIGNAL SEPARATION USING RE-WEIGHTED AND ADAPTIVE MORPHOLOGICAL COMPONENT ANALYSIS GUAN-JU PENG AND WEN-LIANG HWANG Feb. 24, 204 Technical Report No. TR-IIS-4-002 http://www.iis.sinica.edu.tw/page/library/techreport/tr204/tr4.html

More information

Week #4: Midterm 1 Review

Week #4: Midterm 1 Review Week #4: Midterm Review April 5, NAMES: TARDIS : http://math.ucsb.edu/ kgracekennedy/spring 4A.html Week : Introduction to Systems of Linear Equations Problem.. What row operations are allowed and why?...

More information

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,

More information

Section Gaussian Elimination

Section Gaussian Elimination Section. - Gaussian Elimination A matrix is said to be in row echelon form (REF) if it has the following properties:. The first nonzero entry in any row is a. We call this a leading one or pivot one..

More information

Multiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing

Multiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing Multiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing ibeltal F. Alem, Daniel H. Chae, and Rodney A. Kennedy The Australian National

More information

MTH 35, SPRING 2017 NIKOS APOSTOLAKIS

MTH 35, SPRING 2017 NIKOS APOSTOLAKIS MTH 35, SPRING 2017 NIKOS APOSTOLAKIS 1. Linear independence Example 1. Recall the set S = {a i : i = 1,...,5} R 4 of the last two lectures, where a 1 = (1,1,3,1) a 2 = (2,1,2, 1) a 3 = (7,3,5, 5) a 4

More information

RECENTLY, there has been a great deal of interest in

RECENTLY, there has been a great deal of interest in IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 47, NO. 1, JANUARY 1999 187 An Affine Scaling Methodology for Best Basis Selection Bhaskar D. Rao, Senior Member, IEEE, Kenneth Kreutz-Delgado, Senior Member,

More information

Matrices and systems of linear equations

Matrices and systems of linear equations Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.

More information

1. Solve each linear system using Gaussian elimination or Gauss-Jordan reduction. The augmented matrix of this linear system is

1. Solve each linear system using Gaussian elimination or Gauss-Jordan reduction. The augmented matrix of this linear system is Solutions to Homework Additional Problems. Solve each linear system using Gaussian elimination or Gauss-Jordan reduction. (a) x + y = 8 3x + 4y = 7 x + y = 3 The augmented matrix of this linear system

More information

Chapter 6. Linear Independence. Chapter 6

Chapter 6. Linear Independence. Chapter 6 Linear Independence Linear Dependence/Independence A set of vectors {v, v 2,..., v p } is linearly dependent if we can express the zero vector, 0, as a non-trivial linear combination of the vectors. α

More information

Introduction to Compressed Sensing

Introduction to Compressed Sensing Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral

More information

ROBUST BLIND CALIBRATION VIA TOTAL LEAST SQUARES

ROBUST BLIND CALIBRATION VIA TOTAL LEAST SQUARES ROBUST BLIND CALIBRATION VIA TOTAL LEAST SQUARES John Lipor Laura Balzano University of Michigan, Ann Arbor Department of Electrical and Computer Engineering {lipor,girasole}@umich.edu ABSTRACT This paper

More information

χ 1 χ 2 and ψ 1 ψ 2 is also in the plane: αχ 1 αχ 2 which has a zero first component and hence is in the plane. is also in the plane: ψ 1

χ 1 χ 2 and ψ 1 ψ 2 is also in the plane: αχ 1 αχ 2 which has a zero first component and hence is in the plane. is also in the plane: ψ 1 58 Chapter 5 Vector Spaces: Theory and Practice 57 s Which of the following subsets of R 3 are actually subspaces? (a The plane of vectors x =(χ,χ, T R 3 such that the first component χ = In other words,

More information

MATH 152 Exam 1-Solutions 135 pts. Write your answers on separate paper. You do not need to copy the questions. Show your work!!!

MATH 152 Exam 1-Solutions 135 pts. Write your answers on separate paper. You do not need to copy the questions. Show your work!!! MATH Exam -Solutions pts Write your answers on separate paper. You do not need to copy the questions. Show your work!!!. ( pts) Find the reduced row echelon form of the matrix Solution : 4 4 6 4 4 R R

More information

System of Linear Equations

System of Linear Equations Chapter 7 - S&B Gaussian and Gauss-Jordan Elimination We will study systems of linear equations by describing techniques for solving such systems. The preferred solution technique- Gaussian elimination-

More information

On the Role of the Properties of the Nonzero Entries on Sparse Signal Recovery

On the Role of the Properties of the Nonzero Entries on Sparse Signal Recovery On the Role of the Properties of the Nonzero Entries on Sparse Signal Recovery Yuzhe Jin and Bhaskar D. Rao Department of Electrical and Computer Engineering, University of California at San Diego, La

More information

Linear System Equations

Linear System Equations King Saud University September 24, 2018 Table of contents 1 2 3 4 Definition A linear system of equations with m equations and n unknowns is defined as follows: a 1,1 x 1 + a 1,2 x 2 + + a 1,n x n = b

More information

1 Regression with High Dimensional Data

1 Regression with High Dimensional Data 6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:

More information

Nonnegative Tensor Factorization with Smoothness Constraints

Nonnegative Tensor Factorization with Smoothness Constraints Nonnegative Tensor Factorization with Smoothness Constraints Rafal ZDUNEK 1 and Tomasz M. RUTKOWSKI 2 1 Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology,

More information

DNNs for Sparse Coding and Dictionary Learning

DNNs for Sparse Coding and Dictionary Learning DNNs for Sparse Coding and Dictionary Learning Subhadip Mukherjee, Debabrata Mahapatra, and Chandra Sekhar Seelamantula Department of Electrical Engineering, Indian Institute of Science, Bangalore 5612,

More information

l 0 -norm Minimization for Basis Selection

l 0 -norm Minimization for Basis Selection l 0 -norm Minimization for Basis Selection David Wipf and Bhaskar Rao Department of Electrical and Computer Engineering University of California, San Diego, CA 92092 dwipf@ucsd.edu, brao@ece.ucsd.edu Abstract

More information

Review for Chapter 1. Selected Topics

Review for Chapter 1. Selected Topics Review for Chapter 1 Selected Topics Linear Equations We have four equivalent ways of writing linear systems: 1 As a system of equations: 2x 1 + 3x 2 = 7 x 1 x 2 = 5 2 As an augmented matrix: ( 2 3 ) 7

More information

Math 369 Exam #2 Practice Problem Solutions

Math 369 Exam #2 Practice Problem Solutions Math 369 Exam #2 Practice Problem Solutions 2 5. Is { 2, 3, 8 } a basis for R 3? Answer: No, it is not. To show that it is not a basis, it suffices to show that this is not a linearly independent set.

More information

2. Every linear system with the same number of equations as unknowns has a unique solution.

2. Every linear system with the same number of equations as unknowns has a unique solution. 1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations

More information

MA 242 LINEAR ALGEBRA C1, Solutions to First Midterm Exam

MA 242 LINEAR ALGEBRA C1, Solutions to First Midterm Exam MA 242 LINEAR ALGEBRA C Solutions to First Midterm Exam Prof Nikola Popovic October 2 9:am - :am Problem ( points) Determine h and k such that the solution set of x + = k 4x + h = 8 (a) is empty (b) contains

More information

Robust multichannel sparse recovery

Robust multichannel sparse recovery Robust multichannel sparse recovery Esa Ollila Department of Signal Processing and Acoustics Aalto University, Finland SUPELEC, Feb 4th, 2015 1 Introduction 2 Nonparametric sparse recovery 3 Simulation

More information

Sparse Signal Recovery: Theory, Applications and Algorithms

Sparse Signal Recovery: Theory, Applications and Algorithms Sparse Signal Recovery: Theory, Applications and Algorithms Bhaskar Rao Department of Electrical and Computer Engineering University of California, San Diego Collaborators: I. Gorodonitsky, S. Cotter,

More information

SGN Advanced Signal Processing Project bonus: Sparse model estimation

SGN Advanced Signal Processing Project bonus: Sparse model estimation SGN 21006 Advanced Signal Processing Project bonus: Sparse model estimation Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 12 Sparse models Initial problem: solve

More information

Dictionary Learning for L1-Exact Sparse Coding

Dictionary Learning for L1-Exact Sparse Coding Dictionary Learning for L1-Exact Sparse Coding Mar D. Plumbley Department of Electronic Engineering, Queen Mary University of London, Mile End Road, London E1 4NS, United Kingdom. Email: mar.plumbley@elec.qmul.ac.u

More information

Compressed Sensing and Robust Recovery of Low Rank Matrices

Compressed Sensing and Robust Recovery of Low Rank Matrices Compressed Sensing and Robust Recovery of Low Rank Matrices M. Fazel, E. Candès, B. Recht, P. Parrilo Electrical Engineering, University of Washington Applied and Computational Mathematics Dept., Caltech

More information

squares based sparse system identification for the error in variables

squares based sparse system identification for the error in variables Lim and Pang SpringerPlus 2016)5:1460 DOI 10.1186/s40064-016-3120-6 RESEARCH Open Access l 1 regularized recursive total least squares based sparse system identification for the error in variables Jun

More information

Problem Sheet 1 with Solutions GRA 6035 Mathematics

Problem Sheet 1 with Solutions GRA 6035 Mathematics Problem Sheet 1 with Solutions GRA 6035 Mathematics BI Norwegian Business School 2 Problems 1. From linear system to augmented matrix Write down the coefficient matrix and the augmented matrix of the following

More information

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the

More information

COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION. Aditya K. Jagannatham and Bhaskar D.

COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION. Aditya K. Jagannatham and Bhaskar D. COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION Aditya K Jagannatham and Bhaskar D Rao University of California, SanDiego 9500 Gilman Drive, La Jolla, CA 92093-0407

More information

Math 54 HW 4 solutions

Math 54 HW 4 solutions Math 54 HW 4 solutions 2.2. Section 2.2 (a) False: Recall that performing a series of elementary row operations A is equivalent to multiplying A by a series of elementary matrices. Suppose that E,...,

More information

EXAM. Exam #3. Math 2360 Fall 2000 Morning Class. Nov. 29, 2000 ANSWERS

EXAM. Exam #3. Math 2360 Fall 2000 Morning Class. Nov. 29, 2000 ANSWERS EXAM Exam # Math 6 Fall Morning Class Nov. 9, ANSWERS i 5 pts. Problem. In each part you are given the augmented matrix of a system of linear equations, with the coefficent matrix in reduced row echelon

More information

MATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve:

MATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve: MATH 2331 Linear Algebra Section 1.1 Systems of Linear Equations Finding the solution to a set of two equations in two variables: Example 1: Solve: x x = 3 1 2 2x + 4x = 12 1 2 Geometric meaning: Do these

More information

Chapter 2 Notes, Linear Algebra 5e Lay

Chapter 2 Notes, Linear Algebra 5e Lay Contents.1 Operations with Matrices..................................1.1 Addition and Subtraction.............................1. Multiplication by a scalar............................ 3.1.3 Multiplication

More information

Lecture 22: More On Compressed Sensing

Lecture 22: More On Compressed Sensing Lecture 22: More On Compressed Sensing Scribed by Eric Lee, Chengrun Yang, and Sebastian Ament Nov. 2, 207 Recap and Introduction Basis pursuit was the method of recovering the sparsest solution to an

More information

IEOR 265 Lecture 3 Sparse Linear Regression

IEOR 265 Lecture 3 Sparse Linear Regression IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M

More information

Linear Independence Reading: Lay 1.7

Linear Independence Reading: Lay 1.7 Linear Independence Reading: Lay 17 September 11, 213 In this section, we discuss the concept of linear dependence and independence I am going to introduce the definitions and then work some examples and

More information

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017 2 Classification of hazardous

More information

Notes on Row Reduction

Notes on Row Reduction Notes on Row Reduction Francis J. Narcowich Department of Mathematics Texas A&M University September The Row-Reduction Algorithm The row-reduced form of a matrix contains a great deal of information, both

More information

Lecture 6: Spanning Set & Linear Independency

Lecture 6: Spanning Set & Linear Independency Lecture 6: Elif Tan Ankara University Elif Tan (Ankara University) Lecture 6 / 0 Definition (Linear Combination) Let v, v 2,..., v k be vectors in (V,, ) a vector space. A vector v V is called a linear

More information

10. Rank-nullity Definition Let A M m,n (F ). The row space of A is the span of the rows. The column space of A is the span of the columns.

10. Rank-nullity Definition Let A M m,n (F ). The row space of A is the span of the rows. The column space of A is the span of the columns. 10. Rank-nullity Definition 10.1. Let A M m,n (F ). The row space of A is the span of the rows. The column space of A is the span of the columns. The nullity ν(a) of A is the dimension of the kernel. The

More information

1 - Systems of Linear Equations

1 - Systems of Linear Equations 1 - Systems of Linear Equations 1.1 Introduction to Systems of Linear Equations Almost every problem in linear algebra will involve solving a system of equations. ü LINEAR EQUATIONS IN n VARIABLES We are

More information

Sections 1.5, 1.7. Ma 322 Fall Ma 322. Sept

Sections 1.5, 1.7. Ma 322 Fall Ma 322. Sept Sections 1.5, 1.7 Ma 322 Fall 213 Ma 322 Sept. 9-13 Summary ˆ Solutions of homogeneous equations AX =. ˆ Using the rank. ˆ Parametric solution of AX = B. ˆ Linear dependence and independence of vectors

More information

Adaptive Filters. un [ ] yn [ ] w. yn n wun k. - Adaptive filter (FIR): yn n n w nun k. (1) Identification. Unknown System + (2) Inverse modeling

Adaptive Filters. un [ ] yn [ ] w. yn n wun k. - Adaptive filter (FIR): yn n n w nun k. (1) Identification. Unknown System + (2) Inverse modeling Adaptive Filters - Statistical digital signal processing: in many problems of interest, the signals exhibit some inherent variability plus additive noise we use probabilistic laws to model the statistical

More information

Math x + 3y 5z = 14 3x 2y + 3z = 17 4x + 3y 2z = 1

Math x + 3y 5z = 14 3x 2y + 3z = 17 4x + 3y 2z = 1 Math 210 1. Solve the system: x + y + z = 1 2x + 3y + 4z = 5 (a z = 2, y = 1 and x = 0 (b z =any value, y = 3 2z and x = z 2 (c z =any value, y = 3 2z and x = z + 2 (d z =any value, y = 3 + 2z and x =

More information

Lecture 18: The Rank of a Matrix and Consistency of Linear Systems

Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Winfried Just Department of Mathematics, Ohio University February 28, 218 Review: The linear span Definition Let { v 1, v 2,..., v n }

More information

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra Lecture: Linear algebra. 1. Subspaces. 2. Orthogonal complement. 3. The four fundamental subspaces 4. Solutions of linear equation systems The fundamental theorem of linear algebra 5. Determining the fundamental

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

Fast Algorithms for Sparse Recovery with Perturbed Dictionary

Fast Algorithms for Sparse Recovery with Perturbed Dictionary Fast Algorithms for Sparse Recovery with Perturbed Dictionary Xuebing Han, Hao Zhang, Gang Li arxiv:.637v3 cs.it May Abstract In this paper, we account for approaches of sparse recovery from large underdetermined

More information

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix

More information

Number of solutions of a system

Number of solutions of a system Roberto s Notes on Linear Algebra Chapter 3: Linear systems and matrices Section 7 Number of solutions of a system What you need to know already: How to solve a linear system by using Gauss- Jordan elimination.

More information

Linear Combination. v = a 1 v 1 + a 2 v a k v k

Linear Combination. v = a 1 v 1 + a 2 v a k v k Linear Combination Definition 1 Given a set of vectors {v 1, v 2,..., v k } in a vector space V, any vector of the form v = a 1 v 1 + a 2 v 2 +... + a k v k for some scalars a 1, a 2,..., a k, is called

More information

Linear Independence x

Linear Independence x Linear Independence A consistent system of linear equations with matrix equation Ax = b, where A is an m n matrix, has a solution set whose graph in R n is a linear object, that is, has one of only n +

More information

Normed & Inner Product Vector Spaces

Normed & Inner Product Vector Spaces Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed

More information

Row Space, Column Space, and Nullspace

Row Space, Column Space, and Nullspace Row Space, Column Space, and Nullspace MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Introduction Every matrix has associated with it three vector spaces: row space

More information

Rank and Nullity. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics

Rank and Nullity. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics Rank and Nullity MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Objectives We have defined and studied the important vector spaces associated with matrices (row space,

More information

Sections 1.5, 1.7. Ma 322 Spring Ma 322. Jan 24-28

Sections 1.5, 1.7. Ma 322 Spring Ma 322. Jan 24-28 Sections 1.5, 1.7 Ma 322 Spring 217 Ma 322 Jan 24-28 Summary ˆ Text: Solution Sets of Linear Systems (1.5),Linear Independence (1.7) ˆ Solutions of homogeneous equations AX =. ˆ Using the rank. ˆ Parametric

More information

MAT Linear Algebra Collection of sample exams

MAT Linear Algebra Collection of sample exams MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system

More information