Robust Sparse Recovery via Non-Convex Optimization

Similar documents
EUSIPCO

Pre-weighted Matching Pursuit Algorithms for Sparse Recovery

Introduction to Compressed Sensing

Solving Corrupted Quadratic Equations, Provably

Generalized Orthogonal Matching Pursuit- A Review and Some

Elaine T. Hale, Wotao Yin, Yin Zhang

Analysis of Greedy Algorithms

Subspace Projection Matrix Completion on Grassmann Manifold

The Analysis Cosparse Model for Signals and Images

Designing Information Devices and Systems I Discussion 13B

SPARSE signal representations have gained popularity in recent

of Orthogonal Matching Pursuit

Greedy Signal Recovery and Uniform Uncertainty Principles

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Three Generalizations of Compressed Sensing

Nonconvex Sparse Logistic Regression with Weakly Convex Regularization

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk

Robust Principal Component Analysis

Multipath Matching Pursuit

Sparsity in Underdetermined Systems

Sparsity Regularization

Stability and Robustness of Weak Orthogonal Matching Pursuits

Simultaneous Sparsity

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

A new method on deterministic construction of the measurement matrix in compressed sensing

CSC 576: Variants of Sparse Learning

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm

A Review of Sparsity-Based Methods for Analysing Radar Returns from Helicopter Rotor Blades

An Introduction to Sparse Approximation

Linear Convergence of Adaptively Iterative Thresholding Algorithms for Compressed Sensing

Compressive Sensing and Beyond

A NEW FRAMEWORK FOR DESIGNING INCOHERENT SPARSIFYING DICTIONARIES

Self-Calibration and Biconvex Compressive Sensing

A New Estimate of Restricted Isometry Constants for Sparse Solutions

SCRIBERS: SOROOSH SHAFIEEZADEH-ABADEH, MICHAËL DEFFERRARD

MODIFIED DISTRIBUTED ITERATIVE HARD THRESHOLDING

Stochastic geometry and random matrix theory in CS

Recent Developments in Compressed Sensing

When Dictionary Learning Meets Classification

MIT 9.520/6.860, Fall 2018 Statistical Learning Theory and Applications. Class 08: Sparsity Based Regularization. Lorenzo Rosasco

Sparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda

Robust multichannel sparse recovery

Machine Learning for Signal Processing Sparse and Overcomplete Representations. Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 6, JUNE

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT

ACCORDING to Shannon s sampling theorem, an analog

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Covariance Sketching via Quadratic Sampling

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France

GREEDY SIGNAL RECOVERY REVIEW

Generalized Approximate Message Passing for Unlimited Sampling of Sparse Signals

Constrained optimization

Machine Learning for Signal Processing Sparse and Overcomplete Representations

Optimization for Compressed Sensing

Sketching for Large-Scale Learning of Mixture Models

Exact Topology Identification of Large-Scale Interconnected Dynamical Systems from Compressive Observations

Subspace Projection Matrix Recovery from Incomplete Information

Sparse Least Mean Square Algorithm for Estimation of Truncated Volterra Kernels

Strengthened Sobolev inequalities for a random subspace of functions

c 2011 International Press Vol. 18, No. 1, pp , March DENNIS TREDE

Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Final Presentation

Sparse Solutions of Systems of Equations and Sparse Modelling of Signals and Images

The Pros and Cons of Compressive Sensing

The Iteration-Tuned Dictionary for Sparse Representations

Sparse linear models

EE 381V: Large Scale Optimization Fall Lecture 24 April 11

A Tutorial on Compressive Sensing. Simon Foucart Drexel University / University of Georgia

COMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

Bhaskar Rao Department of Electrical and Computer Engineering University of California, San Diego

Complementary Matching Pursuit Algorithms for Sparse Approximation

Optimization methods

PHASE TRANSITION OF JOINT-SPARSE RECOVERY FROM MULTIPLE MEASUREMENTS VIA CONVEX OPTIMIZATION

Stopping Condition for Greedy Block Sparse Signal Recovery

Recent developments on sparse representation

Combining geometry and combinatorics

A Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems

Recovery of Low Rank and Jointly Sparse. Matrices with Two Sampling Matrices

ROBUST BLIND SPIKES DECONVOLUTION. Yuejie Chi. Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 43210

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Local Strong Convexity of Maximum-Likelihood TDOA-Based Source Localization and Its Algorithmic Implications

Compressed Sensing. 1 Introduction. 2 Design of Measurement Matrices

Info-Greedy Sequential Adaptive Compressed Sensing

A tutorial on sparse modeling. Outline:

Sparse Solutions of an Undetermined Linear System

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit

Sparse Covariance Selection using Semidefinite Programming

Sparse analysis Lecture VII: Combining geometry and combinatorics, sparse matrices for sparse signal recovery

Hard Thresholding Pursuit Algorithms: Number of Iterations

On Rank Awareness, Thresholding, and MUSIC for Joint Sparse Recovery

Reconstruction of Block-Sparse Signals by Using an l 2/p -Regularized Least-Squares Algorithm

CoSaMP: Greedy Signal Recovery and Uniform Uncertainty Principles

sparse and low-rank tensor recovery Cubic-Sketching

Transcription:

Robust Sparse Recovery via Non-Convex Optimization Laming Chen and Yuantao Gu Department of Electronic Engineering, Tsinghua University Homepage: http://gu.ee.tsinghua.edu.cn/ Email: gyt@tsinghua.edu.cn August 23, 2014

Contents 1 Introduction 2 Preliminary 3 Main Contribution 4 Simulation 5 Summary

Contents 1 Introduction 2 Preliminary 3 Main Contribution 4 Simulation 5 Summary

Introduction Sparse Recovery Problem Sparse recovery problem is expressed as argmin x 0, subject to y = Ax x where A R M N is a fat sensing matrix. Computationally intractable Sparse Recovery Algorithms Greedy pursuits (OMP, CoSaMP, SP, etc) Convex optimization based algorithms Non-convex optimization based algorithms Better performance tends to be derived Still lacking complete convergence analysis

Introduction Current Convergence Analysis For p-irls (0 < p < 1), the convergence is guaranteed in a sufficiently small neighborhood of the sparse signal. For MM subspace algorithm, it is shown that the generated sequence will converge to a critical point. For SL0, the complete convergence analysis is done due to the local convexity of the penalties, and the algorithm needs to solve a sequence of optimization problems rather than a single one to guarantee convergence. Objective To provide theoretical convergence guarantees of a non-convex approach for sparse recovery from the initial solution to the global optimal.

Contents 1 Introduction 2 Preliminary 3 Main Contribution 4 Simulation 5 Summary

Preliminary Problem Setup Non-convex optimization problem is formulated as argminj(x) = x N F (x i ), subject to y = Ax i=1 F ( ) belongs to a class of sparseness measures (proposed by Rémi Gribonval et al.) 2 1.5 1 No. 1 No. 2 No. 3 No. 4 No. 5 No. 6 0.5 0 2 1.5 1 0.5 0 0.5 1 1.5 2

Preliminary Null Space Property Define γ(j, A, K) as the smallest quantity such that J(z S ) γ(j, A, K)J(z S c) holds for any set S {1, 2,..., N} with #S K and for any z N (A). Performance analysis with γ(j, A, K) (provided by Rémi Gribonval et al.) If γ(j, A, K) < 1, for any K-sparse x and y = Ax, the optimization problem returns x ; If γ(j, A, K) > 1, there exists a K-sparse x and y = Ax such that the problem returns a signal differs from x ; γ(l 0, A, K) γ(j, A, K) γ(l 1, A, K).

Preliminary Weak Convexity F ( ) is ρ-convex if and only if ρ is the largest quantity such that there exists convex H( ) and F (t) = H(t) + ρt 2. F ( ) is strongly convex if ρ > 0; convex if ρ = 0; weakly convex (also known as semi-convex) if ρ < 0. Generalized gradient (a generalization of subgradient for convex functions) for weakly convex functions can be seen as any element in the set F (t) := H(t) + 2ρt.

Contents 1 Introduction 2 Preliminary 3 Main Contribution 4 Simulation 5 Summary

Main Contribution Sparsity-inducing Penalty F ( ) is a weakly convex sparseness measure; Vast majority of non-convex sparsity-inducing penalties are formed by weakly convex sparseness measures; There exists α such that F (t) α t. Define ρ/α as the non-convexity of the penalty (a measure of how quickly the generalized gradient of F ( ) decreases). 2 1.5 ρ/α = 1/3 ρ/α = 1/2 ρ/α = 1 1 0.5 0 2 1.5 1 0.5 0 0.5 1 1.5 2

Main Contribution Projected Generalized Gradient (PGG) Method Initialization: Calculate A, x(0) = A y, n = 0; Repeat: x(n + 1) = x(n) κ J(x(n)); x(n + 1) = x(n + 1) + A (y A x(n + 1)); n = n + 1; Until: Stopping criterion satisfied; Gradient Update x* x(0) y=ax Projection

Main Contribution Performance of PGG Theorem 1. Assume y = Ax + e and x 0 K. For any tuple (J, A, K) with J( ) formed by weakly convex sparseness measure F ( ) and γ(j, A, K) < 1, and for any positive constant M 0, if the non-convexity of J( ) satisfies ρ α 1 M 0 1 γ(j, A, K) 5 + 3γ(J, A, K), the recovered solution ˆx by PGG satisfies ˆx x 2 4α2 N C 1 κ + 8C 2 e 2 provided that x(0) x 2 M 0.

Main Contribution Approximate PGG (APGG) Method To reduce the computational burden of calculating A = A T (AA T ) 1, assume we adopt A T B to approximate A, and let ζ = I AA T B 2. Theorem 2. Under the same assumptions as Theorem 1, if the non-convexity of J( ) satisfies ρ α 1 M 0 1 γ(j, A, K) 5 + 3γ(J, A, K), and ζ < 1, the recovered solution ˆx by APGG satisfies ˆx x 2 2C 3 (ζ)κ + 2C 4 (ζ) e 2 provided that x(0) x 2 M 0.

Contents 1 Introduction 2 Preliminary 3 Main Contribution 4 Simulation 5 Summary

Simulation Parameter Setting Matrix A: 200 1000; independent and identically distributed Gaussian entries; Vector x : nonzero entries independently satisfy Gaussian distribution; normalized to have unit l 2 norm; The approximate A is calculated using an iterative method (introduced by Ben-Israel et al.);

Simulation First Experiment Recovery performance of the PGG method is tested in the noiseless scenario with different sparsity-inducing penalties and different choices of non-convexity. Kmax 70 60 50 40 30 No.1 No.2 No.3 No.4 No.5 No.6 20 10 0 10 1 10 0 10 1 10 2 Non convexity

Simulation Second Experiment The recovery performance of the (A)PGG method is compared in the noiseless scenario with some typical sparse recovery algorithms. 1 0.9 Successful Recovery Probability 0.8 0.7 0.6 0.5 0.4 OMP 0.3 L1 RL1 0.2 ISL0 IRLS 0.1 PGG APGG 0 20 30 40 50 60 70 80 90 100 K

Simulation Third Experiment The recovery precisions of the (A)PGG method are simulated under different settings of step size and measurement noise. Average RSNR 100 90 80 70 60 50 40 30 20 10 PGG APGG MSNR = 20dB MSNR = 30dB MSNR = 40dB MSNR = 50dB MSNR = infinity 0 10 2 10 4 Step Size κ 10 6

Contents 1 Introduction 2 Preliminary 3 Main Contribution 4 Simulation 5 Summary

Summary Conclusion A class of weakly convex sparseness measures is adopted to constitute the sparsity-inducing penalties; The convergence analysis of the (A)PGG method reveals that when the non-convexity is below a threshold, the recovery error is linear in both the step size and the noise term; As for the APGG method, the influence of the approximate projection is reflected in the coefficients instead of an additional error term.

Related Works [1] Laming Chen and Yuantao Gu, The Convergence Guarantees of a Non-convex Approach for Sparse Recovery, IEEE Transactions on Signal Processing, 62(15):3754-3767, 2014. [2] Laming Chen and Yuantao Gu, The Convergence Guarantees of a Non-convex Approach for Sparse Recovery Using Regularized Least Squares, IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 3374-3378, May 4-9, 2014, Florence, Italy. [3] Laming Chen and Yuantao Gu, Robust Recovery of Low-Rank Matrices via Non-Convex Optimization, International Conference on Digital Signal Processing (DSP), 355-360, Aug. 20-23, 2014, Hong Kong, China. http://gu.ee.tsinghua.edu.cn/

Thanks!

Supplementary Weakly Convex Sparseness Measures with Parameter ρ No. F (t) ρ 1. t 0 2. (p 1)σ p 2 t ( t +σ) 1 p 3. 1 e σ t σ 2 /2 4. ln(1 + σ t ) σ 2 /2 5. atan(σ t ) 3 3σ 2 /16 6. (2σ t σ 2 t 2 )X t 1 σ + X t > 1 σ σ 2