Numerical Methods. Rafał Zdunek Underdetermined problems (2h.) Applications) (FOCUSS, M-FOCUSS,
|
|
- Geraldine Johnston
- 5 years ago
- Views:
Transcription
1 Numerical Methods Rafał Zdunek Underdetermined problems (h.) (FOCUSS, M-FOCUSS, M Applications)
2 Introduction Solutions to underdetermined linear systems, Morphological constraints, FOCUSS algorithm, M-FOCUSS algorithm.
3 Bibliography [] A. Borck, Numerical Methods for Least Squares Problems, SIAM, Philadelphia, 996, [] G. Golub, C. F. Van Loan, Matrix Computations, he John Hopkins University Press, (hird Edition), 996, [3] I. F. Gorodnitsky and B. D. Rao, Sparse signal reconstructions from limited data using FOCUSS: A re-weighted minimum norm algorithm, IEEE rans. Signal Process., vol. 45, no. 3, pp , Mar. 997, [4] B. D. Rao and K. Kreutz-Delgado, Deriving algorithms for computing sparse solutions to linear inverse problems, in Proc. 3st Asilomar Conf. Signals Syst. Comput., CA, Nov. 5, 997, vol., pp , [5] B. D. Rao, K. Engan, S. F. Cotter, J. Palmer, and K. Kreutz-Delgado, Subset selection in noise based on diversity measure minimization, IEEE rans. Signal Process., vol. 5, no. 3, pp , Mar. 003, [6] S. F. Cotter, B. Rao, K. Engan, and K. Kreutz-Delgado, Sparse solutions to linear inverse problems with multiple measurement vectors, IEEE rans. Signal Process., vol. 53, no. 7, pp , Jul. 005,
4 Solutions to linear systems A system of linear equations can be expressed in the following matrix form: M N where = [ a i ] I N = [ x ] I Ax = b, () A is a coefficient matrix, [ ] M M ( N +) x is a solution to be estimated. Let [ ] b = b I is a data vector, and i B = A b = I be the augmented matrix to the system (). he system of linear equations may behave in any one of three possible ways: A he system has no solution if ( A) rank( B) rank <. B he system has a single unique solution if rank ( ) = rank( B) = N A. C he system has infinitely many solutions if rank ( ) = rank( B) < N A.
5 Solutions to linear systems he case C occurs for rank-deficient problems or under-determined problems. In spite of infinitely many solutions, a good approximation to a true solution can be obtained if some a priori knowledge about the nature of the true solution is accessible. he additional constraints are usually concerned with a degree of sparsity or smoothness of the true solution.
6 RREF Basic variables Free variables
7 Example x + 3x + x = a x x + x3 = b 3x+ 7x x3 = c Let: 3 [ A b] (Basic variables) (Free variable) 3 a 0 5 a 3b Gauss-Jordan = b 0 b+ a 3 7 c c a+ b he system is consistent (it has solutions) if c a+ b= 0 c= a b Solution: x 3 = free variable, x = a+ b x, x = a 3b+ 5 x3. 3
8 Example Let: x+ x + x3+ 3x4 = 4 x+ 4x + x3+ 3x4 = 5 3x+ 6x + x3+ 4x4 = 7 [ A b] Gauss-Jordan = x = x + x, x = free variable, Solution: 4 Consistent x = x, x 4 = free variable. 3 4
9 Example hus: x x x4 x x 0 0 = = + x + x 4 x 3 x 4 0 x x General solution Particular solution xgeneral = xparticular + xhomogeneous Homogenous solution N(A) he homogeneous solution is a solution to the system Ax = 0. Problem: How to choose free variables? Remark: Replacing free variables with zero-values is not always a good solution!
10 Sparseness constraints Regularized least-squares problem: min x { ( ) ( )} p Ax b +γ E x L p diversity measure (Gorodnitsky, Rao, 997): E ( p) N ( ) = sgn( p) x x p. = p
11 FOCUSS algorithm Let Stationary point: ( p) ( ) J( x) = Ax b +γ E x ( ) J ( x ) = A Ax A b+ λw x x = 0 x * * * * ( ) { p/ = x } W x* diag, p λ = γ Hence: ( ) ( ) Iterative updates: ( λ ) M x W x A AW x A I b * = * * +. ( λ ) ( ) ( ) x W x A AW x A I b = + k+ k k M.
12 FOCUSS algorithm [ 0,] p, λ > 0 (0) Randomly initialize x, For k = 0,,, until convergence do { / } ( k ) p x W = diag, k ( ) λ x = W A AW A + I b, ( k+ ) k k M
13 Wiener filtered FOCUSS [ 0,] p, λ > 0 algorithm ( 0 ) Randomly initialize x, For k = 0,,, until convergence do { / } ( k ) p x W = diag, k ( ) λ x = W A AW A + I b, ( k+ ) k k M x ( ( k ) ) F x ( k + ) + R. Zdunek, Z. He, A. Cichocki, Proc. IEEE ISBI, 008
14 Wiener filtered FOCUSS algorithm Markov Random Field (MRF) μ σ = ( k + ) x n L n = N ( k + ) ( ) x n L n N First and second-order interactions N μ = h,, +, + h, L = 9 h, + h, h +, + h + - local mean around -th pixel - local variance
15 Wiener filtered FOCUSS algorithm Update rule Mean noise variance x ν σ ν ( k+ ) ) μ + σ N = N = σ ( ( k+ x μ )
16 Limited-view tomographic imaging ( A ) rank = LSS ( ) 3 ( A ; b ) = { x R N : Ax b = min! } ( A; b) x N ( A) 0 0 α 0 0 α = 0 α α A 5 ( ) α = (Rank-deficient) { [ ] } N( A( ) ) = span,,,, LSS = LS + where x LS = x r = P ( ) x exact R A
17 omographic imaging example Phantom image Minimal l norm least squares solution: (LS algorithms: AR, SIR)
18 omographic imaging (noise-free data) FOCUSS algorithm (p =, k = 5, ) λ =0 8 Wiener filtered FOCUSS algorithm (p =, k = 5, ) λ =0 8
19 omographic imaging (noisy data SNR = 30 db) FOCUSS algorithm (p =, k = 5, = 40 ) Wiener filtered FOCUSS algorithm (p =, k = 5, ) λ λ = 40
20 omographic imaging (Normalized RMSE, noise-free)
21 omographic imaging (Normalized RMSE, p =, noisy data)
22 M-FOCUSS M N Let AX + N = B where R, A X [ x x ] ( A) M N, (under-determined) rank = M, N,,, N, M = R B R, > N [ n n ],, M =. R (Cotter, Rao, Engan, Kreutz-Delgado, 005) (additive noise) If M < N, the nullspace of A is non-trivial, thus additional constraints are necessary to select the right solution. he M-FOCUSS assumes sparse solutions. heorem: he sparse solution to the consistent system AX = B is unique, ( ) rank B =, M, if with any M columns of A are linearly independent (unique representation property (URP) condition), and for each t: x t has at most ( M + )/ nonzero entries, where is a ceil function.
23 M-FOCUSS he M-FOCUSS algorithm iteratively solves the following equality constrained problem: where p J ( ) ( ) ( p) min J, X N N ( ) p x t = = t= X s.t. AX = B, p/ X = x =. - l diversity measure for sparsity p 0 p degree of sparsity x - the -th row of X (LS solution with minimal l -norm) (l 0 norm solution NP-hard problem)
24 M-FOCUSS For inconsistent data, i.e. B R(A), Cotter at al. developped the regularized M-FOCUSS algorithm that solves the ikhonov regularized least-squares problem in a single iterative step: ( k) ( k ) k X = arg min Ψ( X X ), where ( ) X p/ diag ( w / ( k ) W = ), with w = ( xt ) = ( x ) Regularized M-FOCUSS: For k =,,... A ( k+ ) ( k = AW ), ( ) ( ) t= ( ) λ M ( k+ ) ( k ) ( k ) ( k ) ( k X = W A A A ) + I B ( ) Ψ X X = B AX + λ W X, ( k ) λ > 0 regularization parameter. F. F
Chapter 1. Vectors, Matrices, and Linear Spaces
1.6 Homogeneous Systems, Subspaces and Bases 1 Chapter 1. Vectors, Matrices, and Linear Spaces 1.6. Homogeneous Systems, Subspaces and Bases Note. In this section we explore the structure of the solution
More informationA ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT
Progress In Electromagnetics Research Letters, Vol. 16, 53 60, 2010 A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Y. P. Liu and Q. Wan School of Electronic Engineering University of Electronic
More informationEXAM. Exam #2. Math 2360 Summer II, 2000 Morning Class. Nov. 15, 2000 ANSWERS
EXAM Exam # Math 6 Summer II Morning Class Nov 5 ANSWERS i Problem Consider the matrix 6 pts A = 6 4 9 5 7 6 5 5 5 4 The RREF of A is the matrix R = A Find a basis for the nullspace of A Solve the homogeneous
More informationThe matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.
) Find all solutions of the linear system. Express the answer in vector form. x + 2x + x + x 5 = 2 2x 2 + 2x + 2x + x 5 = 8 x + 2x + x + 9x 5 = 2 2 Solution: Reduce the augmented matrix [ 2 2 2 8 ] to
More informationLecture 21: 5.6 Rank and Nullity
Lecture 21: 5.6 Rank and Nullity Wei-Ta Chu 2008/12/5 Rank and Nullity Definition The common dimension of the row and column space of a matrix A is called the rank ( 秩 ) of A and is denoted by rank(a);
More informationRow Reduction and Echelon Forms
Row Reduction and Echelon Forms 1 / 29 Key Concepts row echelon form, reduced row echelon form pivot position, pivot, pivot column basic variable, free variable general solution, parametric solution existence
More informationMATH 2360 REVIEW PROBLEMS
MATH 2360 REVIEW PROBLEMS Problem 1: In (a) (d) below, either compute the matrix product or indicate why it does not exist: ( )( ) 1 2 2 1 (a) 0 1 1 2 ( ) 0 1 2 (b) 0 3 1 4 3 4 5 2 5 (c) 0 3 ) 1 4 ( 1
More informationMidterm 1 Review. Written by Victoria Kala SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015
Midterm 1 Review Written by Victoria Kala vtkala@math.ucsb.edu SH 6432u Office Hours: R 12:30 1:30 pm Last updated 10/10/2015 Summary This Midterm Review contains notes on sections 1.1 1.5 and 1.7 in your
More informationA Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems
A Continuation Approach to Estimate a Solution Path of Mixed L2-L Minimization Problems Junbo Duan, Charles Soussen, David Brie, Jérôme Idier Centre de Recherche en Automatique de Nancy Nancy-University,
More informationEquivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 12, DECEMBER 2008 2009 Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation Yuanqing Li, Member, IEEE, Andrzej Cichocki,
More informationSolutions of Linear system, vector and matrix equation
Goals: Solutions of Linear system, vector and matrix equation Solutions of linear system. Vectors, vector equation. Matrix equation. Math 112, Week 2 Suggested Textbook Readings: Sections 1.3, 1.4, 1.5
More informationSimultaneous Sparsity
Simultaneous Sparsity Joel A. Tropp Anna C. Gilbert Martin J. Strauss {jtropp annacg martinjs}@umich.edu Department of Mathematics The University of Michigan 1 Simple Sparse Approximation Work in the d-dimensional,
More informationSECTION 3.3. PROBLEM 22. The null space of a matrix A is: N(A) = {X : AX = 0}. Here are the calculations of AX for X = a,b,c,d, and e. =
SECTION 3.3. PROBLEM. The null space of a matrix A is: N(A) {X : AX }. Here are the calculations of AX for X a,b,c,d, and e. Aa [ ][ ] 3 3 [ ][ ] Ac 3 3 [ ] 3 3 [ ] 4+4 6+6 Ae [ ], Ab [ ][ ] 3 3 3 [ ]
More informationLecture 3: Linear Algebra Review, Part II
Lecture 3: Linear Algebra Review, Part II Brian Borchers January 4, Linear Independence Definition The vectors v, v,..., v n are linearly independent if the system of equations c v + c v +...+ c n v n
More informationImproved FOCUSS Method With Conjugate Gradient Iterations
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 1, JANUARY 2009 399 Improved FOCUSS Method With Conjugate Gradient Iterations Zhaoshui He, Andrzej Cichocki, Rafal Zdunek, and Shengli Xie Abstract
More informationLinear Algebra I Lecture 10
Linear Algebra I Lecture 10 Xi Chen 1 1 University of Alberta January 30, 2019 Outline 1 Gauss-Jordan Algorithm ] Let A = [a ij m n be an m n matrix. To reduce A to a reduced row echelon form using elementary
More informationMATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix.
MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix. Basis Definition. Let V be a vector space. A linearly independent spanning set for V is called a basis.
More informationDM559 Linear and Integer Programming. Lecture 2 Systems of Linear Equations. Marco Chiarandini
DM559 Linear and Integer Programming Lecture Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. Outline 1. 3 A Motivating Example You are organizing
More informationLinear equations in linear algebra
Linear equations in linear algebra Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra Pearson Collections Samy T. Linear
More informationMath 314H EXAM I. 1. (28 points) The row reduced echelon form of the augmented matrix for the system. is the matrix
Math 34H EXAM I Do all of the problems below. Point values for each of the problems are adjacent to the problem number. Calculators may be used to check your answer but not to arrive at your answer. That
More informationMath 2030 Assignment 5 Solutions
Math 030 Assignment 5 Solutions Question 1: Which of the following sets of vectors are linearly independent? If the set is linear dependent, find a linear dependence relation for the vectors (a) {(1, 0,
More informationChapter 1. Vectors, Matrices, and Linear Spaces
1.4 Solving Systems of Linear Equations 1 Chapter 1. Vectors, Matrices, and Linear Spaces 1.4. Solving Systems of Linear Equations Note. We give an algorithm for solving a system of linear equations (called
More informationThe definition of a vector space (V, +, )
The definition of a vector space (V, +, ) 1. For any u and v in V, u + v is also in V. 2. For any u and v in V, u + v = v + u. 3. For any u, v, w in V, u + ( v + w) = ( u + v) + w. 4. There is an element
More informationb for the linear system x 1 + x 2 + a 2 x 3 = a x 1 + x 3 = 3 x 1 + x 2 + 9x 3 = 3 ] 1 1 a 2 a
Practice Exercises for Exam Exam will be on Monday, September 8, 7. The syllabus for Exam consists of Sections One.I, One.III, Two.I, and Two.II. You should know the main definitions, results and computational
More informationSolutions to Homework 5 - Math 3410
Solutions to Homework 5 - Math 34 (Page 57: # 489) Determine whether the following vectors in R 4 are linearly dependent or independent: (a) (, 2, 3, ), (3, 7,, 2), (, 3, 7, 4) Solution From x(, 2, 3,
More informationLecture 2 Systems of Linear Equations and Matrices, Continued
Lecture 2 Systems of Linear Equations and Matrices, Continued Math 19620 Outline of Lecture Algorithm for putting a matrix in row reduced echelon form - i.e. Gauss-Jordan Elimination Number of Solutions
More informationChapter 3. Directions: For questions 1-11 mark each statement True or False. Justify each answer.
Chapter 3 Directions: For questions 1-11 mark each statement True or False. Justify each answer. 1. (True False) Asking whether the linear system corresponding to an augmented matrix [ a 1 a 2 a 3 b ]
More informationType II variational methods in Bayesian estimation
Type II variational methods in Bayesian estimation J. A. Palmer, D. P. Wipf, and K. Kreutz-Delgado Department of Electrical and Computer Engineering University of California San Diego, La Jolla, CA 9093
More information1. Determine by inspection which of the following sets of vectors is linearly independent. 3 3.
1. Determine by inspection which of the following sets of vectors is linearly independent. (a) (d) 1, 3 4, 1 { [ [,, 1 1] 3]} (b) 1, 4 5, (c) 3 6 (e) 1, 3, 4 4 3 1 4 Solution. The answer is (a): v 1 is
More informationMATH 2050 Assignment 6 Fall 2018 Due: Thursday, November 1. x + y + 2z = 2 x + y + z = c 4x + 2z = 2
MATH 5 Assignment 6 Fall 8 Due: Thursday, November [5]. For what value of c does have a solution? Is it unique? x + y + z = x + y + z = c 4x + z = Writing the system as an augmented matrix, we have c R
More informationPseudoinverse & Moore-Penrose Conditions
ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego p. 1/1 Lecture 7 ECE 275A Pseudoinverse & Moore-Penrose Conditions ECE 275AB Lecture 7 Fall 2008 V1.0 c K. Kreutz-Delgado, UC San Diego
More informationYORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions
YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 222 3. M Test # July, 23 Solutions. For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For
More informationMATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.
MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:
More informationLinear Equations in Linear Algebra
Linear Equations in Linear Algebra.7 LINEAR INDEPENDENCE LINEAR INDEPENDENCE Definition: An indexed set of vectors {v,, v p } in n is said to be linearly independent if the vector equation x x x 2 2 p
More informationEUSIPCO
EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,
More informationFind the solution set of 2x 3y = 5. Answer: We solve for x = (5 + 3y)/2. Hence the solution space consists of all vectors of the form
Math 2 Homework #7 March 4, 2 7.3.3. Find the solution set of 2x 3y = 5. Answer: We solve for x = (5 + 3y/2. Hence the solution space consists of all vectors of the form ( ( ( ( x (5 + 3y/2 5/2 3/2 x =
More informationNon-Negative Matrix Factorization with Quasi-Newton Optimization
Non-Negative Matrix Factorization with Quasi-Newton Optimization Rafal ZDUNEK, Andrzej CICHOCKI Laboratory for Advanced Brain Signal Processing BSI, RIKEN, Wako-shi, JAPAN Abstract. Non-negative matrix
More informationReview Notes for Linear Algebra True or False Last Updated: February 22, 2010
Review Notes for Linear Algebra True or False Last Updated: February 22, 2010 Chapter 4 [ Vector Spaces 4.1 If {v 1,v 2,,v n } and {w 1,w 2,,w n } are linearly independent, then {v 1 +w 1,v 2 +w 2,,v n
More informationSIGNAL SEPARATION USING RE-WEIGHTED AND ADAPTIVE MORPHOLOGICAL COMPONENT ANALYSIS
TR-IIS-4-002 SIGNAL SEPARATION USING RE-WEIGHTED AND ADAPTIVE MORPHOLOGICAL COMPONENT ANALYSIS GUAN-JU PENG AND WEN-LIANG HWANG Feb. 24, 204 Technical Report No. TR-IIS-4-002 http://www.iis.sinica.edu.tw/page/library/techreport/tr204/tr4.html
More informationWeek #4: Midterm 1 Review
Week #4: Midterm Review April 5, NAMES: TARDIS : http://math.ucsb.edu/ kgracekennedy/spring 4A.html Week : Introduction to Systems of Linear Equations Problem.. What row operations are allowed and why?...
More informationSimultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors
Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,
More informationSection Gaussian Elimination
Section. - Gaussian Elimination A matrix is said to be in row echelon form (REF) if it has the following properties:. The first nonzero entry in any row is a. We call this a leading one or pivot one..
More informationMultiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing
Multiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing ibeltal F. Alem, Daniel H. Chae, and Rodney A. Kennedy The Australian National
More informationMTH 35, SPRING 2017 NIKOS APOSTOLAKIS
MTH 35, SPRING 2017 NIKOS APOSTOLAKIS 1. Linear independence Example 1. Recall the set S = {a i : i = 1,...,5} R 4 of the last two lectures, where a 1 = (1,1,3,1) a 2 = (2,1,2, 1) a 3 = (7,3,5, 5) a 4
More informationRECENTLY, there has been a great deal of interest in
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 47, NO. 1, JANUARY 1999 187 An Affine Scaling Methodology for Best Basis Selection Bhaskar D. Rao, Senior Member, IEEE, Kenneth Kreutz-Delgado, Senior Member,
More informationMatrices and systems of linear equations
Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.
More information1. Solve each linear system using Gaussian elimination or Gauss-Jordan reduction. The augmented matrix of this linear system is
Solutions to Homework Additional Problems. Solve each linear system using Gaussian elimination or Gauss-Jordan reduction. (a) x + y = 8 3x + 4y = 7 x + y = 3 The augmented matrix of this linear system
More informationChapter 6. Linear Independence. Chapter 6
Linear Independence Linear Dependence/Independence A set of vectors {v, v 2,..., v p } is linearly dependent if we can express the zero vector, 0, as a non-trivial linear combination of the vectors. α
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationROBUST BLIND CALIBRATION VIA TOTAL LEAST SQUARES
ROBUST BLIND CALIBRATION VIA TOTAL LEAST SQUARES John Lipor Laura Balzano University of Michigan, Ann Arbor Department of Electrical and Computer Engineering {lipor,girasole}@umich.edu ABSTRACT This paper
More informationχ 1 χ 2 and ψ 1 ψ 2 is also in the plane: αχ 1 αχ 2 which has a zero first component and hence is in the plane. is also in the plane: ψ 1
58 Chapter 5 Vector Spaces: Theory and Practice 57 s Which of the following subsets of R 3 are actually subspaces? (a The plane of vectors x =(χ,χ, T R 3 such that the first component χ = In other words,
More informationMATH 152 Exam 1-Solutions 135 pts. Write your answers on separate paper. You do not need to copy the questions. Show your work!!!
MATH Exam -Solutions pts Write your answers on separate paper. You do not need to copy the questions. Show your work!!!. ( pts) Find the reduced row echelon form of the matrix Solution : 4 4 6 4 4 R R
More informationSystem of Linear Equations
Chapter 7 - S&B Gaussian and Gauss-Jordan Elimination We will study systems of linear equations by describing techniques for solving such systems. The preferred solution technique- Gaussian elimination-
More informationOn the Role of the Properties of the Nonzero Entries on Sparse Signal Recovery
On the Role of the Properties of the Nonzero Entries on Sparse Signal Recovery Yuzhe Jin and Bhaskar D. Rao Department of Electrical and Computer Engineering, University of California at San Diego, La
More informationLinear System Equations
King Saud University September 24, 2018 Table of contents 1 2 3 4 Definition A linear system of equations with m equations and n unknowns is defined as follows: a 1,1 x 1 + a 1,2 x 2 + + a 1,n x n = b
More information1 Regression with High Dimensional Data
6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:
More informationNonnegative Tensor Factorization with Smoothness Constraints
Nonnegative Tensor Factorization with Smoothness Constraints Rafal ZDUNEK 1 and Tomasz M. RUTKOWSKI 2 1 Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology,
More informationDNNs for Sparse Coding and Dictionary Learning
DNNs for Sparse Coding and Dictionary Learning Subhadip Mukherjee, Debabrata Mahapatra, and Chandra Sekhar Seelamantula Department of Electrical Engineering, Indian Institute of Science, Bangalore 5612,
More informationl 0 -norm Minimization for Basis Selection
l 0 -norm Minimization for Basis Selection David Wipf and Bhaskar Rao Department of Electrical and Computer Engineering University of California, San Diego, CA 92092 dwipf@ucsd.edu, brao@ece.ucsd.edu Abstract
More informationReview for Chapter 1. Selected Topics
Review for Chapter 1 Selected Topics Linear Equations We have four equivalent ways of writing linear systems: 1 As a system of equations: 2x 1 + 3x 2 = 7 x 1 x 2 = 5 2 As an augmented matrix: ( 2 3 ) 7
More informationMath 369 Exam #2 Practice Problem Solutions
Math 369 Exam #2 Practice Problem Solutions 2 5. Is { 2, 3, 8 } a basis for R 3? Answer: No, it is not. To show that it is not a basis, it suffices to show that this is not a linearly independent set.
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More informationMA 242 LINEAR ALGEBRA C1, Solutions to First Midterm Exam
MA 242 LINEAR ALGEBRA C Solutions to First Midterm Exam Prof Nikola Popovic October 2 9:am - :am Problem ( points) Determine h and k such that the solution set of x + = k 4x + h = 8 (a) is empty (b) contains
More informationRobust multichannel sparse recovery
Robust multichannel sparse recovery Esa Ollila Department of Signal Processing and Acoustics Aalto University, Finland SUPELEC, Feb 4th, 2015 1 Introduction 2 Nonparametric sparse recovery 3 Simulation
More informationSparse Signal Recovery: Theory, Applications and Algorithms
Sparse Signal Recovery: Theory, Applications and Algorithms Bhaskar Rao Department of Electrical and Computer Engineering University of California, San Diego Collaborators: I. Gorodonitsky, S. Cotter,
More informationSGN Advanced Signal Processing Project bonus: Sparse model estimation
SGN 21006 Advanced Signal Processing Project bonus: Sparse model estimation Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 12 Sparse models Initial problem: solve
More informationDictionary Learning for L1-Exact Sparse Coding
Dictionary Learning for L1-Exact Sparse Coding Mar D. Plumbley Department of Electronic Engineering, Queen Mary University of London, Mile End Road, London E1 4NS, United Kingdom. Email: mar.plumbley@elec.qmul.ac.u
More informationCompressed Sensing and Robust Recovery of Low Rank Matrices
Compressed Sensing and Robust Recovery of Low Rank Matrices M. Fazel, E. Candès, B. Recht, P. Parrilo Electrical Engineering, University of Washington Applied and Computational Mathematics Dept., Caltech
More informationsquares based sparse system identification for the error in variables
Lim and Pang SpringerPlus 2016)5:1460 DOI 10.1186/s40064-016-3120-6 RESEARCH Open Access l 1 regularized recursive total least squares based sparse system identification for the error in variables Jun
More informationProblem Sheet 1 with Solutions GRA 6035 Mathematics
Problem Sheet 1 with Solutions GRA 6035 Mathematics BI Norwegian Business School 2 Problems 1. From linear system to augmented matrix Write down the coefficient matrix and the augmented matrix of the following
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationCOMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION. Aditya K. Jagannatham and Bhaskar D.
COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION Aditya K Jagannatham and Bhaskar D Rao University of California, SanDiego 9500 Gilman Drive, La Jolla, CA 92093-0407
More informationMath 54 HW 4 solutions
Math 54 HW 4 solutions 2.2. Section 2.2 (a) False: Recall that performing a series of elementary row operations A is equivalent to multiplying A by a series of elementary matrices. Suppose that E,...,
More informationEXAM. Exam #3. Math 2360 Fall 2000 Morning Class. Nov. 29, 2000 ANSWERS
EXAM Exam # Math 6 Fall Morning Class Nov. 9, ANSWERS i 5 pts. Problem. In each part you are given the augmented matrix of a system of linear equations, with the coefficent matrix in reduced row echelon
More informationMATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve:
MATH 2331 Linear Algebra Section 1.1 Systems of Linear Equations Finding the solution to a set of two equations in two variables: Example 1: Solve: x x = 3 1 2 2x + 4x = 12 1 2 Geometric meaning: Do these
More informationChapter 2 Notes, Linear Algebra 5e Lay
Contents.1 Operations with Matrices..................................1.1 Addition and Subtraction.............................1. Multiplication by a scalar............................ 3.1.3 Multiplication
More informationLecture 22: More On Compressed Sensing
Lecture 22: More On Compressed Sensing Scribed by Eric Lee, Chengrun Yang, and Sebastian Ament Nov. 2, 207 Recap and Introduction Basis pursuit was the method of recovering the sparsest solution to an
More informationIEOR 265 Lecture 3 Sparse Linear Regression
IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M
More informationLinear Independence Reading: Lay 1.7
Linear Independence Reading: Lay 17 September 11, 213 In this section, we discuss the concept of linear dependence and independence I am going to introduce the definitions and then work some examples and
More informationA randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors
A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017 2 Classification of hazardous
More informationNotes on Row Reduction
Notes on Row Reduction Francis J. Narcowich Department of Mathematics Texas A&M University September The Row-Reduction Algorithm The row-reduced form of a matrix contains a great deal of information, both
More informationLecture 6: Spanning Set & Linear Independency
Lecture 6: Elif Tan Ankara University Elif Tan (Ankara University) Lecture 6 / 0 Definition (Linear Combination) Let v, v 2,..., v k be vectors in (V,, ) a vector space. A vector v V is called a linear
More information10. Rank-nullity Definition Let A M m,n (F ). The row space of A is the span of the rows. The column space of A is the span of the columns.
10. Rank-nullity Definition 10.1. Let A M m,n (F ). The row space of A is the span of the rows. The column space of A is the span of the columns. The nullity ν(a) of A is the dimension of the kernel. The
More information1 - Systems of Linear Equations
1 - Systems of Linear Equations 1.1 Introduction to Systems of Linear Equations Almost every problem in linear algebra will involve solving a system of equations. ü LINEAR EQUATIONS IN n VARIABLES We are
More informationSections 1.5, 1.7. Ma 322 Fall Ma 322. Sept
Sections 1.5, 1.7 Ma 322 Fall 213 Ma 322 Sept. 9-13 Summary ˆ Solutions of homogeneous equations AX =. ˆ Using the rank. ˆ Parametric solution of AX = B. ˆ Linear dependence and independence of vectors
More informationAdaptive Filters. un [ ] yn [ ] w. yn n wun k. - Adaptive filter (FIR): yn n n w nun k. (1) Identification. Unknown System + (2) Inverse modeling
Adaptive Filters - Statistical digital signal processing: in many problems of interest, the signals exhibit some inherent variability plus additive noise we use probabilistic laws to model the statistical
More informationMath x + 3y 5z = 14 3x 2y + 3z = 17 4x + 3y 2z = 1
Math 210 1. Solve the system: x + y + z = 1 2x + 3y + 4z = 5 (a z = 2, y = 1 and x = 0 (b z =any value, y = 3 2z and x = z 2 (c z =any value, y = 3 2z and x = z + 2 (d z =any value, y = 3 + 2z and x =
More informationLecture 18: The Rank of a Matrix and Consistency of Linear Systems
Lecture 18: The Rank of a Matrix and Consistency of Linear Systems Winfried Just Department of Mathematics, Ohio University February 28, 218 Review: The linear span Definition Let { v 1, v 2,..., v n }
More informationLecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra
Lecture: Linear algebra. 1. Subspaces. 2. Orthogonal complement. 3. The four fundamental subspaces 4. Solutions of linear equation systems The fundamental theorem of linear algebra 5. Determining the fundamental
More information2 Regularized Image Reconstruction for Compressive Imaging and Beyond
EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement
More informationFast Algorithms for Sparse Recovery with Perturbed Dictionary
Fast Algorithms for Sparse Recovery with Perturbed Dictionary Xuebing Han, Hao Zhang, Gang Li arxiv:.637v3 cs.it May Abstract In this paper, we account for approaches of sparse recovery from large underdetermined
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationNumber of solutions of a system
Roberto s Notes on Linear Algebra Chapter 3: Linear systems and matrices Section 7 Number of solutions of a system What you need to know already: How to solve a linear system by using Gauss- Jordan elimination.
More informationLinear Combination. v = a 1 v 1 + a 2 v a k v k
Linear Combination Definition 1 Given a set of vectors {v 1, v 2,..., v k } in a vector space V, any vector of the form v = a 1 v 1 + a 2 v 2 +... + a k v k for some scalars a 1, a 2,..., a k, is called
More informationLinear Independence x
Linear Independence A consistent system of linear equations with matrix equation Ax = b, where A is an m n matrix, has a solution set whose graph in R n is a linear object, that is, has one of only n +
More informationNormed & Inner Product Vector Spaces
Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed
More informationRow Space, Column Space, and Nullspace
Row Space, Column Space, and Nullspace MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Introduction Every matrix has associated with it three vector spaces: row space
More informationRank and Nullity. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics
Rank and Nullity MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Objectives We have defined and studied the important vector spaces associated with matrices (row space,
More informationSections 1.5, 1.7. Ma 322 Spring Ma 322. Jan 24-28
Sections 1.5, 1.7 Ma 322 Spring 217 Ma 322 Jan 24-28 Summary ˆ Text: Solution Sets of Linear Systems (1.5),Linear Independence (1.7) ˆ Solutions of homogeneous equations AX =. ˆ Using the rank. ˆ Parametric
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More information