Reconstruction of Block-Sparse Signals by Using an l 2/p -Regularized Least-Squares Algorithm
|
|
- Caren Greene
- 5 years ago
- Views:
Transcription
1 Reconstruction of Block-Sparse Signals by Using an l 2/p -Regularized Least-Squares Algorithm Jeevan K. Pant, Wu-Sheng Lu, and Andreas Antoniou University of Victoria May 21, 2012 Compressive Sensing 1/23 University of Victoria
2 Outline Compressive Sensing Compressive Sensing 2/23 University of Victoria
3 Outline Compressive Sensing Signal Recovery by Using l 1 or l p Minimization Compressive Sensing 2/23 University of Victoria
4 Outline Compressive Sensing Signal Recovery by Using l 1 or l p Minimization Recovery of Block-Sparse Signals Compressive Sensing 2/23 University of Victoria
5 Outline Compressive Sensing Signal Recovery by Using l 1 or l p Minimization Recovery of Block-Sparse Signals Block-Sparse Signal Recovery by Using l 2/p Minimization Compressive Sensing 2/23 University of Victoria
6 Outline Compressive Sensing Signal Recovery by Using l 1 or l p Minimization Recovery of Block-Sparse Signals Block-Sparse Signal Recovery by Using l 2/p Minimization Performance Evaluation Compressive Sensing 2/23 University of Victoria
7 Outline Compressive Sensing Signal Recovery by Using l 1 or l p Minimization Recovery of Block-Sparse Signals Block-Sparse Signal Recovery by Using l 2/p Minimization Performance Evaluation Conclusions Compressive Sensing 2/23 University of Victoria
8 Compressive Sensing A signal x(n) of length N is K-sparse if it contains K nonzero components with K N. Compressive Sensing 3/23 University of Victoria
9 Compressive Sensing A signal x(n) of length N is K-sparse if it contains K nonzero components with K N. A signal is near K-sparse if it contains K significant components. Compressive Sensing 3/23 University of Victoria
10 Compressive Sensing A signal x(n) of length N is K-sparse if it contains K nonzero components with K N. A signal is near K-sparse if it contains K significant components. Example: an image with near sparse wavelet coefficients: An image of An equivalent Wavelet coefficients Lena 1-D signal Compressive Sensing 3/23 University of Victoria
11 Compressive Sensing, cont d Compressive sensing (CS) is a data acquisition process whereby a sparse signal x(n) represented by a vector x of length N is determined using a small number of projections represented by a matrix Φ of dimension M N. Compressive Sensing 4/23 University of Victoria
12 Compressive Sensing, cont d Compressive sensing (CS) is a data acquisition process whereby a sparse signal x(n) represented by a vector x of length N is determined using a small number of projections represented by a matrix Φ of dimension M N. In such a process, measurement vector y and signal vector x are interrelated by the equation y = Φ x = 16 measurements projection matrix 4-sparse signal of size of length 30 Compressive Sensing 4/23 University of Victoria
13 Signal Recovery by Using l 1 or l p Minimization The inverse problem of recovering signal x from measurement y such that Φ x = y M N N 1 M 1 is an ill-posed problem. Compressive Sensing 5/23 University of Victoria
14 Signal Recovery by Using l 1 or l p Minimization The inverse problem of recovering signal x from measurement y such that Φ x = y M N N 1 M 1 is an ill-posed problem. l 2 minimization often fails to yield a sparse x, i.e., a signal obtained as x = arg x is often not sparse. min x 2 subject to Φx = y Compressive Sensing 5/23 University of Victoria
15 Signal Recovery by Using l 1 or l p Minimization The inverse problem of recovering signal x from measurement y such that Φ x = y M N N 1 M 1 is an ill-posed problem. l 2 minimization often fails to yield a sparse x, i.e., a signal obtained as x = arg x is often not sparse. min x 2 subject to Φx = y A sparse x can be recovered using l 1 minimization as x = arg x min x 1 subject to Φx = y Compressive Sensing 5/23 University of Victoria
16 Signal Recovery by Using l 1 or l p Minimization, cont d Recently, l p minimization based algorithms have been shown to recover sparse signals using fewer measurements. In these algorithms, the signal is recovered by using the optimization problem where p < 1. minimize x subject to x p p = N i=1 x i p Φx = y Compressive Sensing 6/23 University of Victoria
17 Signal Recovery by Using l 1 or l p Minimization, cont d Recently, l p minimization based algorithms have been shown to recover sparse signals using fewer measurements. In these algorithms, the signal is recovered by using the optimization problem where p < 1. minimize x subject to x p p = N i=1 x i p Φx = y Note that the objective function x p p in the above problem is nonconvex and nondifferentiable. Compressive Sensing 6/23 University of Victoria
18 Signal Recovery by Using l 1 or l p Minimization, cont d Recently, l p minimization based algorithms have been shown to recover sparse signals using fewer measurements. In these algorithms, the signal is recovered by using the optimization problem where p < 1. minimize x subject to x p p = N i=1 x i p Φx = y Note that the objective function x p p in the above problem is nonconvex and nondifferentiable. Despite this, it has been shown in the literature that if the above problem is solved with sufficient care, improved reconstruction performance can be achieved. Compressive Sensing 6/23 University of Victoria
19 Signal Recovery by Using l 1 or l p Minimization, cont d Example: N = 256, K = 35, M = 100. Compressive Sensing 7/23 University of Victoria
20 Recovery of Block-Sparse Signals Let N, d, and N/d be positive integers such that d < N and N/d < N. Compressive Sensing 8/23 University of Victoria
21 Recovery of Block-Sparse Signals Let N, d, and N/d be positive integers such that d < N and N/d < N. A signal x of length N can be divided into N/d blocks as x = [ x 1 x 2 x N/d ] T where x i = [ x (i 1)d+1 x (i 1)d+1 x (i 1)d+d ] T for i = 1, 2,..., N/d. Compressive Sensing 8/23 University of Victoria
22 Recovery of Block-Sparse Signals Let N, d, and N/d be positive integers such that d < N and N/d < N. A signal x of length N can be divided into N/d blocks as x = [ x 1 x 2 x N/d ] T where x i = [ x (i 1)d+1 x (i 1)d+1 x (i 1)d+d ] T for i = 1, 2,..., N/d. Signal x is said to be K-block sparse if it has K nonzero blocks with K N/d. Compressive Sensing 8/23 University of Victoria
23 Recovery of Block-Sparse Signals Let N, d, and N/d be positive integers such that d < N and N/d < N. A signal x of length N can be divided into N/d blocks as x = [ x 1 x 2 x N/d ] T where x i = [ x (i 1)d+1 x (i 1)d+1 x (i 1)d+d ] T for i = 1, 2,..., N/d. Signal x is said to be K-block sparse if it has K nonzero blocks with K N/d. Note that the definition of K-sparse in the conventional CS is the special case of K-block sparse with d = 1. Compressive Sensing 8/23 University of Victoria
24 Recovery of Block-Sparse Signals, cont d Block-sparsity naturally arises in various signals such as speech signals, multiband signals, and some images. Compressive Sensing 9/23 University of Victoria
25 Recovery of Block-Sparse Signals, cont d Block-sparsity naturally arises in various signals such as speech signals, multiband signals, and some images. Speech signal: Compressive Sensing 9/23 University of Victoria
26 Recovery of Block-Sparse Signals, cont d Block-sparsity naturally arises in various signals such as speech signals, multiband signals, and some images. Speech signal: Multiband spectrum: Compressive Sensing 9/23 University of Victoria
27 Recovery of Block-Sparse Signals, cont d An image of Jupiter: An image of Jupiter An equivalent 1-D signal The 1-D signal from n = to Compressive Sensing 10/23 University of Victoria
28 Recovery of Block-Sparse Signals, cont d The block sparsity of a signal can be measured using the l 2/0 -pseudonorm which is given by N/d x 2/0 = ( x i 2 ) 0 i=1 Compressive Sensing 11/23 University of Victoria
29 Recovery of Block-Sparse Signals, cont d The block sparsity of a signal can be measured using the l 2/0 -pseudonorm which is given by N/d x 2/0 = ( x i 2 ) 0 The value of function x 2/0 is equal to the number of blocks of x which have all-nonzero components. i=1 Compressive Sensing 11/23 University of Victoria
30 Recovery of Block-Sparse Signals, cont d The block sparsity of a signal can be measured using the l 2/0 -pseudonorm which is given by N/d x 2/0 = ( x i 2 ) 0 The value of function x 2/0 is equal to the number of blocks of x which have all-nonzero components. A block-sparse signal can therefore be recovered by solving the optimization problem minimize x subject to i=1 x 2/0 Φx = y Compressive Sensing 11/23 University of Victoria
31 Recovery of Block-Sparse Signals, cont d The block sparsity of a signal can be measured using the l 2/0 -pseudonorm which is given by N/d x 2/0 = ( x i 2 ) 0 The value of function x 2/0 is equal to the number of blocks of x which have all-nonzero components. A block-sparse signal can therefore be recovered by solving the optimization problem minimize x subject to i=1 x 2/0 Φx = y Unfortunately, this problem is nonconvex with combinatorial complexity. Compressive Sensing 11/23 University of Victoria
32 Recovery of Block-Sparse Signals, cont d A practical method for recovering a block sparse signal is to solve the problem minimize x 2/1 x subject to Φx = y where N/d x 2/1 = x i 2 i=1 Compressive Sensing 12/23 University of Victoria
33 Recovery of Block-Sparse Signals, cont d A practical method for recovering a block sparse signal is to solve the problem minimize x 2/1 x subject to Φx = y where N/d x 2/1 = x i 2 i=1 Note that function x 2/1 is the l 1 norm of the vector [ x1 2 x 2 2 x N/d 2 ] T, which essentially gives a measure of the inter-block sparsity of x. Compressive Sensing 12/23 University of Victoria
34 Recovery of Block-Sparse Signals, cont d A practical method for recovering a block sparse signal is to solve the problem minimize x 2/1 x subject to Φx = y where N/d x 2/1 = x i 2 i=1 Note that function x 2/1 is the l 1 norm of the vector [ x1 2 x 2 2 x N/d 2 ] T, which essentially gives a measure of the inter-block sparsity of x. The above problem is a convex programming problem which can be solved using a semidefinite-programming or a second-order cone-programming (SOCP) solver. Compressive Sensing 12/23 University of Victoria
35 Recovery of Block-Sparse Signals, cont d Example: N = 512, d = 8, K = 5, M = 100. Compressive Sensing 13/23 University of Victoria
36 Block-Sparse Signal Recovery by Using l 2/p Minimization We propose reconstructing a block-sparse signal x from measurement y by solving the l 2/p -regularized least-squares problem minimize x with p < 1 for a small ɛ where F ɛ (x) = 1 2 Φx y λ x p 2/p,ɛ (P) N/d x p 2/p,ɛ = ( xi ɛ 2) p/2 i=1 Compressive Sensing 14/23 University of Victoria
37 Block-Sparse Signal Recovery by Using l 2/p Minimization We propose reconstructing a block-sparse signal x from measurement y by solving the l 2/p -regularized least-squares problem minimize x F ɛ (x) = 1 2 Φx y λ x p 2/p,ɛ (P) with p < 1 for a small ɛ where N/d x p 2/p,ɛ = ( xi ɛ 2) p/2 i=1 Note that lim ɛ 0 x p 2/p,ɛ = x p 2/p lim p 0 x p 2/p = x 2/0 Compressive Sensing 14/23 University of Victoria
38 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Good signal reconstruction performance is expected when problem P on slide 14 is solved with a sufficiently small ɛ. Compressive Sensing 15/23 University of Victoria
39 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Good signal reconstruction performance is expected when problem P on slide 14 is solved with a sufficiently small ɛ. However, for small ɛ the objective function F ɛ (x) becomes highly nonconvex and nearly nondifferentiable. Compressive Sensing 15/23 University of Victoria
40 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Good signal reconstruction performance is expected when problem P on slide 14 is solved with a sufficiently small ɛ. However, for small ɛ the objective function F ɛ (x) becomes highly nonconvex and nearly nondifferentiable. The larger the ɛ, the easier the optimization of F ɛ (x). Compressive Sensing 15/23 University of Victoria
41 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Good signal reconstruction performance is expected when problem P on slide 14 is solved with a sufficiently small ɛ. However, for small ɛ the objective function F ɛ (x) becomes highly nonconvex and nearly nondifferentiable. The larger the ɛ, the easier the optimization of F ɛ (x). Therefore, we propose to solve problem P on slide 14 by using the following sequential optimization procedure: Choose a sufficiently large value of ɛ and solve problem P using Fletcher-Reeves conjugate-gradient (CG) algorithm. Set the solution to x. Reduce the value of ɛ, use x as an initializer, and solve problem P again. Repeat this procedure until problem P is solved for a sufficiently small value of ɛ. Output the final solution and stop. Compressive Sensing 15/23 University of Victoria
42 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d In the kth iteration of Fletcher-Reeves CG algorithm, iterate x k is updated as x k+1 = x k + α k d k where d k = g k + β k 1 d k 1 β k 1 = g k 2 2 g k Compressive Sensing 16/23 University of Victoria
43 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d In the kth iteration of Fletcher-Reeves CG algorithm, iterate x k is updated as x k+1 = x k + α k d k where d k = g k + β k 1 d k 1 β k 1 = g k 2 2 g k Given x k and d k, the step size α k is obtained by solving the optimization problem minimize α f (α) = F ɛ (x k + αd k ) Compressive Sensing 16/23 University of Victoria
44 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d By setting the first derivative of f (α) to zero, we get where α = G(α) d T k Φ T (Φx k y) + λ p G(α) = N/d Φd k λ p γ i γ i = ( x i + α d i 22 + ɛ 2 ) p/2 1 i=1 N/d γ i ( x Tki d ) ki i=1 ( d T ki d ki ) Compressive Sensing 17/23 University of Victoria
45 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d By setting the first derivative of f (α) to zero, we get where α = G(α) d T k Φ T (Φx k y) + λ p G(α) = N/d Φd k λ p γ i γ i = ( x i + α d i 22 + ɛ 2 ) p/2 1 i=1 N/d γ i ( x Tki d ) ki i=1 ( d T ki d ki ) In the above equations, x ki and d ki are the ith blocks of vectors x k and d k, respectively. Compressive Sensing 17/23 University of Victoria
46 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Step size α k is determined by using the recursive relation α l+1 = G(α l ) for l = 1, 2,... Compressive Sensing 18/23 University of Victoria
47 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Step size α k is determined by using the recursive relation α l+1 = G(α l ) for l = 1, 2,... According to Banach s fixed-point theorem, if dg(α)/dα < 1 then function G(α) is a contraction mapping, i.e., G(α 1 ) G(α 2 ) η α 1 α 2 with η < 1 and, as a consequence, the above recursion converges to a solution. Compressive Sensing 18/23 University of Victoria
48 Block-Sparse Signal Recovery by Using l 2/p Minimization, cont d Step size α k is determined by using the recursive relation α l+1 = G(α l ) for l = 1, 2,... According to Banach s fixed-point theorem, if dg(α)/dα < 1 then function G(α) is a contraction mapping, i.e., G(α 1 ) G(α 2 ) η α 1 α 2 with η < 1 and, as a consequence, the above recursion converges to a solution. Extensive experimental results have shown that function G(α) for function f (α) is, in practice, a contraction mapping. Compressive Sensing 18/23 University of Victoria
49 Performance Evaluation Number of perfectly recovered instances with N = 512, M = 100, and d = 8 over 100 runs. l 2/p -RLS: Proposed l 2/p -Regularized Least-Squares l 2/1 -SOCP: l 2/1 Second-Order Cone-Programming (Eldar and Mishali, 2009) BOMP: Block Orthogonal Matching Pursuit (Eldar et. al., 2010) Compressive Sensing 19/23 University of Victoria
50 Performance Evaluation, cont d Average CPU time with M = N/2, K = M/2.5d, and d = 8 over 100 runs. l 2/p -RLS: Proposed l 2/p -Regularized Least-Squares l 2/1 -SOCP: l 2/1 Second-Order Cone-Programming (Eldar and Mishali, 2009) BOMP: Block Orthogonal Matching Pursuit (Eldar et. al., 2010) Compressive Sensing 20/23 University of Victoria
51 Performance Evaluation, cont d Example: N = 512, d = 8, K = 9, M = 100. Compressive Sensing 21/23 University of Victoria
52 Conclusions Compressive sensing is an effective sampling technique for sparse signals. Compressive Sensing 22/23 University of Victoria
53 Conclusions Compressive sensing is an effective sampling technique for sparse signals. l 1 -minimization and l p -minimization with p < 1 work well for the reconstruction of sparse signals. Compressive Sensing 22/23 University of Victoria
54 Conclusions Compressive sensing is an effective sampling technique for sparse signals. l 1 -minimization and l p -minimization with p < 1 work well for the reconstruction of sparse signals. l 2/1 -minimization offers improved reconstruction performance for block-sparse signals. Compressive Sensing 22/23 University of Victoria
55 Conclusions Compressive sensing is an effective sampling technique for sparse signals. l 1 -minimization and l p -minimization with p < 1 work well for the reconstruction of sparse signals. l 2/1 -minimization offers improved reconstruction performance for block-sparse signals. The proposed l 2/p -regularized least-squares algorithm offers improved reconstruction performance for block-sparse signals relative to the l 2/1 -SOCP and BOMP algorithms. Compressive Sensing 22/23 University of Victoria
56 Thank you for your attention. This presentation can be downloaded from: andreas/rlectures/iscas2012-jeevan-pres.pdf Compressive Sensing 23/23 University of Victoria
Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm
Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm J. K. Pant, W.-S. Lu, and A. Antoniou University of Victoria August 25, 2011 Compressive Sensing 1 University
More informationSPARSE signal representations have gained popularity in recent
6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying
More informationDesign of Projection Matrix for Compressive Sensing by Nonsmooth Optimization
Design of Proection Matrix for Compressive Sensing by Nonsmooth Optimization W.-S. Lu T. Hinamoto Dept. of Electrical & Computer Engineering Graduate School of Engineering University of Victoria Hiroshima
More informationDesigns of Orthogonal Filter Banks and Orthogonal Cosine-Modulated Filter Banks
1 / 45 Designs of Orthogonal Filter Banks and Orthogonal Cosine-Modulated Filter Banks Jie Yan Department of Electrical and Computer Engineering University of Victoria April 16, 2010 2 / 45 OUTLINE 1 INTRODUCTION
More informationMLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT
MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net
More informationLarge-Scale L1-Related Minimization in Compressive Sensing and Beyond
Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March
More informationEstimating Unknown Sparsity in Compressed Sensing
Estimating Unknown Sparsity in Compressed Sensing Miles Lopes UC Berkeley Department of Statistics CSGF Program Review July 16, 2014 early version published at ICML 2013 Miles Lopes ( UC Berkeley ) estimating
More informationDirect Design of Orthogonal Filter Banks and Wavelets
Direct Design of Orthogonal Filter Banks and Wavelets W.-S. Lu T. Hinamoto Dept. of Electrical & Computer Engineering Graduate School of Engineering University of Victoria Hiroshima University Victoria,
More informationConjugate Gradient algorithm. Storage: fixed, independent of number of steps.
Conjugate Gradient algorithm Need: A symmetric positive definite; Cost: 1 matrix-vector product per step; Storage: fixed, independent of number of steps. The CG method minimizes the A norm of the error,
More informationLecture Notes 9: Constrained Optimization
Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form
More informationRecent Developments in Compressed Sensing
Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationCompressed Sensing: Extending CLEAN and NNLS
Compressed Sensing: Extending CLEAN and NNLS Ludwig Schwardt SKA South Africa (KAT Project) Calibration & Imaging Workshop Socorro, NM, USA 31 March 2009 Outline 1 Compressed Sensing (CS) Introduction
More informationConvex Optimization and l 1 -minimization
Convex Optimization and l 1 -minimization Sangwoon Yun Computational Sciences Korea Institute for Advanced Study December 11, 2009 2009 NIMS Thematic Winter School Outline I. Convex Optimization II. l
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationNear Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing
Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar
More informationNew Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit
New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence
More informationAn Introduction to Sparse Approximation
An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,
More informationA New Estimate of Restricted Isometry Constants for Sparse Solutions
A New Estimate of Restricted Isometry Constants for Sparse Solutions Ming-Jun Lai and Louis Y. Liu January 12, 211 Abstract We show that as long as the restricted isometry constant δ 2k < 1/2, there exist
More informationElaine T. Hale, Wotao Yin, Yin Zhang
, Wotao Yin, Yin Zhang Department of Computational and Applied Mathematics Rice University McMaster University, ICCOPT II-MOPTA 2007 August 13, 2007 1 with Noise 2 3 4 1 with Noise 2 3 4 1 with Noise 2
More informationTowards Global Design of Orthogonal Filter Banks and Wavelets
Towards Global Design of Orthogonal Filter Banks and Wavelets Jie Yan and Wu-Sheng Lu Department of Electrical and Computer Engineering University of Victoria Victoria, BC, Canada V8W 3P6 jyan@ece.uvic.ca,
More informationIEOR 265 Lecture 3 Sparse Linear Regression
IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M
More informationGeneralized Orthogonal Matching Pursuit- A Review and Some
Generalized Orthogonal Matching Pursuit- A Review and Some New Results Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur, INDIA Table of Contents
More informationCompressive sensing in the analog world
Compressive sensing in the analog world Mark A. Davenport Georgia Institute of Technology School of Electrical and Computer Engineering Compressive Sensing A D -sparse Can we really acquire analog signals
More informationOptimization methods
Optimization methods Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda /8/016 Introduction Aim: Overview of optimization methods that Tend to
More informationCompressed Sensing. 1 Introduction. 2 Design of Measurement Matrices
Compressed Sensing Yonina C. Eldar Electrical Engineering Department, Technion-Israel Institute of Technology, Haifa, Israel, 32000 1 Introduction Compressed sensing (CS) is an exciting, rapidly growing
More informationAlgorithms for sparse analysis Lecture I: Background on sparse approximation
Algorithms for sparse analysis Lecture I: Background on sparse approximation Anna C. Gilbert Department of Mathematics University of Michigan Tutorial on sparse approximations and algorithms Compress data
More informationMismatch and Resolution in CS
Mismatch and Resolution in CS Albert Fannjiang, UC Davis Coworker: Wenjing Liao, UC Davis SPIE Optics + Photonics, San Diego, August 2-25, 20 Mismatch: gridding error Band exclusion Local optimization
More informationThree Generalizations of Compressed Sensing
Thomas Blumensath School of Mathematics The University of Southampton June, 2010 home prev next page Compressed Sensing and beyond y = Φx + e x R N or x C N x K is K-sparse and x x K 2 is small y R M or
More informationSOCP Relaxation of Sensor Network Localization
SOCP Relaxation of Sensor Network Localization Paul Tseng Mathematics, University of Washington Seattle University of Vienna/Wien June 19, 2006 Abstract This is a talk given at Univ. Vienna, 2006. SOCP
More informationEE 381V: Large Scale Optimization Fall Lecture 24 April 11
EE 381V: Large Scale Optimization Fall 2012 Lecture 24 April 11 Lecturer: Caramanis & Sanghavi Scribe: Tao Huang 24.1 Review In past classes, we studied the problem of sparsity. Sparsity problem is that
More informationSparse Recovery of Streaming Signals Using. M. Salman Asif and Justin Romberg. Abstract
Sparse Recovery of Streaming Signals Using l 1 -Homotopy 1 M. Salman Asif and Justin Romberg arxiv:136.3331v1 [cs.it] 14 Jun 213 Abstract Most of the existing methods for sparse signal recovery assume
More informationSparse analysis Lecture III: Dictionary geometry and greedy algorithms
Sparse analysis Lecture III: Dictionary geometry and greedy algorithms Anna C. Gilbert Department of Mathematics University of Michigan Intuition from ONB Key step in algorithm: r, ϕ j = x c i ϕ i, ϕ j
More informationOptimization methods
Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method Lecture 5, Continuous Optimisation Oxford University Computing Laboratory, HT 2006 Notes by Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The notion of complexity (per iteration)
More informationsparse and low-rank tensor recovery Cubic-Sketching
Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru
More informationBayesian Paradigm. Maximum A Posteriori Estimation
Bayesian Paradigm Maximum A Posteriori Estimation Simple acquisition model noise + degradation Constraint minimization or Equivalent formulation Constraint minimization Lagrangian (unconstraint minimization)
More informationPart 4: IIR Filters Optimization Approach. Tutorial ISCAS 2007
Part 4: IIR Filters Optimization Approach Tutorial ISCAS 2007 Copyright 2007 Andreas Antoniou Victoria, BC, Canada Email: aantoniou@ieee.org July 24, 2007 Frame # 1 Slide # 1 A. Antoniou Part4: IIR Filters
More informationSparse Approximation of Signals with Highly Coherent Dictionaries
Sparse Approximation of Signals with Highly Coherent Dictionaries Bishnu P. Lamichhane and Laura Rebollo-Neira b.p.lamichhane@aston.ac.uk, rebollol@aston.ac.uk Support from EPSRC (EP/D062632/1) is acknowledged
More informationEUSIPCO
EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,
More information2 Regularized Image Reconstruction for Compressive Imaging and Beyond
EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement
More informationA Unified Approach to the Design of Interpolated and Frequency Response Masking FIR Filters
A Unified Approach to the Design of Interpolated and Frequency Response Masking FIR Filters Wu Sheng Lu akao Hinamoto University of Victoria Hiroshima University Victoria, Canada Higashi Hiroshima, Japan
More informationPre-weighted Matching Pursuit Algorithms for Sparse Recovery
Journal of Information & Computational Science 11:9 (214) 2933 2939 June 1, 214 Available at http://www.joics.com Pre-weighted Matching Pursuit Algorithms for Sparse Recovery Jingfei He, Guiling Sun, Jie
More informationCOMPRESSIVE SAMPLING USING EM ALGORITHM. Technical Report No: ASU/2014/4
COMPRESSIVE SAMPLING USING EM ALGORITHM ATANU KUMAR GHOSH, ARNAB CHAKRABORTY Technical Report No: ASU/2014/4 Date: 29 th April, 2014 Applied Statistics Unit Indian Statistical Institute Kolkata- 700018
More informationCompressive Imaging by Generalized Total Variation Minimization
1 / 23 Compressive Imaging by Generalized Total Variation Minimization Jie Yan and Wu-Sheng Lu Department of Electrical and Computer Engineering University of Victoria, Victoria, BC, Canada APCCAS 2014,
More informationFast Hard Thresholding with Nesterov s Gradient Method
Fast Hard Thresholding with Nesterov s Gradient Method Volkan Cevher Idiap Research Institute Ecole Polytechnique Federale de ausanne volkan.cevher@epfl.ch Sina Jafarpour Department of Computer Science
More informationThe Conjugate Gradient Algorithm
Optimization over a Subspace Conjugate Direction Methods Conjugate Gradient Algorithm Non-Quadratic Conjugate Gradient Algorithm Optimization over a Subspace Consider the problem min f (x) subject to x
More informationTRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS
TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS Martin Kleinsteuber and Simon Hawe Department of Electrical Engineering and Information Technology, Technische Universität München, München, Arcistraße
More informationSource Reconstruction for 3D Bioluminescence Tomography with Sparse regularization
1/33 Source Reconstruction for 3D Bioluminescence Tomography with Sparse regularization Xiaoqun Zhang xqzhang@sjtu.edu.cn Department of Mathematics/Institute of Natural Sciences, Shanghai Jiao Tong University
More informationCompressive Sensing (CS)
Compressive Sensing (CS) Luminita Vese & Ming Yan lvese@math.ucla.edu yanm@math.ucla.edu Department of Mathematics University of California, Los Angeles The UCLA Advanced Neuroimaging Summer Program (2014)
More informationPrimal Dual Pursuit A Homotopy based Algorithm for the Dantzig Selector
Primal Dual Pursuit A Homotopy based Algorithm for the Dantzig Selector Muhammad Salman Asif Thesis Committee: Justin Romberg (Advisor), James McClellan, Russell Mersereau School of Electrical and Computer
More informationOptimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30
Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained
More informationInfo-Greedy Sequential Adaptive Compressed Sensing
Info-Greedy Sequential Adaptive Compressed Sensing Yao Xie Joint work with Gabor Braun and Sebastian Pokutta Georgia Institute of Technology Presented at Allerton Conference 2014 Information sensing for
More informationSymmetries in Experimental Design and Group Lasso Kentaro Tanaka and Masami Miyakawa
Symmetries in Experimental Design and Group Lasso Kentaro Tanaka and Masami Miyakawa Workshop on computational and algebraic methods in statistics March 3-5, Sanjo Conference Hall, Hongo Campus, University
More informationSolution Recovery via L1 minimization: What are possible and Why?
Solution Recovery via L1 minimization: What are possible and Why? Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Eighth US-Mexico Workshop on Optimization
More informationScale Mixture Modeling of Priors for Sparse Signal Recovery
Scale Mixture Modeling of Priors for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Jason Palmer, Zhilin Zhang and Ritwik Giri Outline Outline Sparse
More informationRobust Sparse Recovery via Non-Convex Optimization
Robust Sparse Recovery via Non-Convex Optimization Laming Chen and Yuantao Gu Department of Electronic Engineering, Tsinghua University Homepage: http://gu.ee.tsinghua.edu.cn/ Email: gyt@tsinghua.edu.cn
More informationLecture 16: Multiresolution Image Analysis
Lecture 16: Multiresolution Image Analysis Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology rhody@cis.rit.edu November 9, 2004 Abstract Multiresolution analysis
More informationBlind Compressed Sensing
1 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE arxiv:1002.2586v2 [cs.it] 28 Apr 2010 Abstract The fundamental principle underlying compressed sensing is that a signal,
More informationIterative regularization of nonlinear ill-posed problems in Banach space
Iterative regularization of nonlinear ill-posed problems in Banach space Barbara Kaltenbacher, University of Klagenfurt joint work with Bernd Hofmann, Technical University of Chemnitz, Frank Schöpfer and
More informationMessage Passing Algorithms for Compressed Sensing: II. Analysis and Validation
Message Passing Algorithms for Compressed Sensing: II. Analysis and Validation David L. Donoho Department of Statistics Arian Maleki Department of Electrical Engineering Andrea Montanari Department of
More informationSparsity Regularization
Sparsity Regularization Bangti Jin Course Inverse Problems & Imaging 1 / 41 Outline 1 Motivation: sparsity? 2 Mathematical preliminaries 3 l 1 solvers 2 / 41 problem setup finite-dimensional formulation
More informationMotivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble
Bayesian Methods for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Zhilin Zhang and Ritwik Giri Motivation Sparse Signal Recovery is an interesting
More informationCoSaMP. Iterative signal recovery from incomplete and inaccurate samples. Joel A. Tropp
CoSaMP Iterative signal recovery from incomplete and inaccurate samples Joel A. Tropp Applied & Computational Mathematics California Institute of Technology jtropp@acm.caltech.edu Joint with D. Needell
More informationPhase recovery with PhaseCut and the wavelet transform case
Phase recovery with PhaseCut and the wavelet transform case Irène Waldspurger Joint work with Alexandre d Aspremont and Stéphane Mallat Introduction 2 / 35 Goal : Solve the non-linear inverse problem Reconstruct
More informationOptimization for Compressed Sensing
Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve
More informationCME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.
CME342 Parallel Methods in Numerical Analysis Matrix Computation: Iterative Methods II Outline: CG & its parallelization. Sparse Matrix-vector Multiplication. 1 Basic iterative methods: Ax = b r = b Ax
More information1 Conjugate gradients
Notes for 2016-11-18 1 Conjugate gradients We now turn to the method of conjugate gradients (CG), perhaps the best known of the Krylov subspace solvers. The CG iteration can be characterized as the iteration
More information5742 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 12, DECEMBER /$ IEEE
5742 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 12, DECEMBER 2009 Uncertainty Relations for Shift-Invariant Analog Signals Yonina C. Eldar, Senior Member, IEEE Abstract The past several years
More information1 Sparsity and l 1 relaxation
6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the
More informationIterative Reweighted Minimization Methods for l p Regularized Unconstrained Nonlinear Programming
Iterative Reweighted Minimization Methods for l p Regularized Unconstrained Nonlinear Programming Zhaosong Lu October 5, 2012 (Revised: June 3, 2013; September 17, 2013) Abstract In this paper we study
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationarxiv: v1 [math.na] 26 Nov 2009
Non-convexly constrained linear inverse problems arxiv:0911.5098v1 [math.na] 26 Nov 2009 Thomas Blumensath Applied Mathematics, School of Mathematics, University of Southampton, University Road, Southampton,
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationSOCP Relaxation of Sensor Network Localization
SOCP Relaxation of Sensor Network Localization Paul Tseng Mathematics, University of Washington Seattle IWCSN, Simon Fraser University August 19, 2006 Abstract This is a talk given at Int. Workshop on
More informationApplied Machine Learning for Biomedical Engineering. Enrico Grisan
Applied Machine Learning for Biomedical Engineering Enrico Grisan enrico.grisan@dei.unipd.it Data representation To find a representation that approximates elements of a signal class with a linear combination
More informationModel-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk
Model-Based Compressive Sensing for Signal Ensembles Marco F. Duarte Volkan Cevher Richard G. Baraniuk Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional
More informationSparse linear models
Sparse linear models Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 2/22/2016 Introduction Linear transforms Frequency representation Short-time
More informationHomotopy methods based on l 0 norm for the compressed sensing problem
Homotopy methods based on l 0 norm for the compressed sensing problem Wenxing Zhu, Zhengshan Dong Center for Discrete Mathematics and Theoretical Computer Science, Fuzhou University, Fuzhou 350108, China
More informationAnalog-to-Information Conversion
Analog-to-Information Conversion Sergiy A. Vorobyov Dept. Signal Processing and Acoustics, Aalto University February 2013 Winter School on Compressed Sensing, Ruka 1/55 Outline 1 Compressed Sampling (CS)
More informationSparse signals recovered by non-convex penalty in quasi-linear systems
Cui et al. Journal of Inequalities and Applications 018) 018:59 https://doi.org/10.1186/s13660-018-165-8 R E S E A R C H Open Access Sparse signals recovered by non-conve penalty in quasi-linear systems
More informationRobust multichannel sparse recovery
Robust multichannel sparse recovery Esa Ollila Department of Signal Processing and Acoustics Aalto University, Finland SUPELEC, Feb 4th, 2015 1 Introduction 2 Nonparametric sparse recovery 3 Simulation
More informationSparse Approximation via Penalty Decomposition Methods
Sparse Approximation via Penalty Decomposition Methods Zhaosong Lu Yong Zhang February 19, 2012 Abstract In this paper we consider sparse approximation problems, that is, general l 0 minimization problems
More informationParticle Filtered Modified-CS (PaFiMoCS) for tracking signal sequences
Particle Filtered Modified-CS (PaFiMoCS) for tracking signal sequences Samarjit Das and Namrata Vaswani Department of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/
More informationSensing systems limited by constraints: physical size, time, cost, energy
Rebecca Willett Sensing systems limited by constraints: physical size, time, cost, energy Reduce the number of measurements needed for reconstruction Higher accuracy data subject to constraints Original
More informationCompressed Sensing and Linear Codes over Real Numbers
Compressed Sensing and Linear Codes over Real Numbers Henry D. Pfister (joint with Fan Zhang) Texas A&M University College Station Information Theory and Applications Workshop UC San Diego January 31st,
More informationOslo Class 6 Sparsity based regularization
RegML2017@SIMULA Oslo Class 6 Sparsity based regularization Lorenzo Rosasco UNIGE-MIT-IIT May 4, 2017 Learning from data Possible only under assumptions regularization min Ê(w) + λr(w) w Smoothness Sparsity
More informationBlock-sparse Solutions using Kernel Block RIP and its Application to Group Lasso
Block-sparse Solutions using Kernel Block RIP and its Application to Group Lasso Rahul Garg IBM T.J. Watson research center grahul@us.ibm.com Rohit Khandekar IBM T.J. Watson research center rohitk@us.ibm.com
More informationConjugate Gradients I: Setup
Conjugate Gradients I: Setup CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Conjugate Gradients I: Setup 1 / 22 Time for Gaussian Elimination
More informationLU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark
DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline
More informationThe Pros and Cons of Compressive Sensing
The Pros and Cons of Compressive Sensing Mark A. Davenport Stanford University Department of Statistics Compressive Sensing Replace samples with general linear measurements measurements sampled signal
More informationSGN Advanced Signal Processing Project bonus: Sparse model estimation
SGN 21006 Advanced Signal Processing Project bonus: Sparse model estimation Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 12 Sparse models Initial problem: solve
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 21: Sensitivity of Eigenvalues and Eigenvectors; Conjugate Gradient Method Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis
More informationSparse Approximation and Variable Selection
Sparse Approximation and Variable Selection Lorenzo Rosasco 9.520 Class 07 February 26, 2007 About this class Goal To introduce the problem of variable selection, discuss its connection to sparse approximation
More informationExact Topology Identification of Large-Scale Interconnected Dynamical Systems from Compressive Observations
Exact Topology Identification of arge-scale Interconnected Dynamical Systems from Compressive Observations Borhan M Sanandaji, Tyrone Vincent, and Michael B Wakin Abstract In this paper, we consider the
More informationLecture 16: Compressed Sensing
Lecture 16: Compressed Sensing Introduction to Learning and Analysis of Big Data Kontorovich and Sabato (BGU) Lecture 16 1 / 12 Review of Johnson-Lindenstrauss Unsupervised learning technique key insight:
More informationConstrained optimization
Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained
More informationCompressed Sensing: a Subgradient Descent Method for Missing Data Problems
Compressed Sensing: a Subgradient Descent Method for Missing Data Problems ANZIAM, Jan 30 Feb 3, 2011 Jonathan M. Borwein Jointly with D. Russell Luke, University of Goettingen FRSC FAAAS FBAS FAA Director,
More information