Recovering overcomplete sparse representations from structured sensing

Similar documents
Strengthened Sobolev inequalities for a random subspace of functions

On the coherence barrier and analogue problems in compressed sensing

Exponential decay of reconstruction error from binary measurements of sparse signals

Beyond incoherence and beyond sparsity: compressed sensing in the real world

Sparse Legendre expansions via l 1 minimization

Interpolation via weighted l 1 -minimization

Interpolation via weighted l 1 minimization

Sparse recovery for spherical harmonic expansions

Lecture 22: More On Compressed Sensing

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

Interpolation via weighted l 1 -minimization

Signal Recovery, Uncertainty Relations, and Minkowski Dimension

Lecture Notes 9: Constrained Optimization

Compressed sensing and imaging

COMPRESSED Sensing (CS) is a method to recover a

Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit

CoSaMP: Greedy Signal Recovery and Uniform Uncertainty Principles

Greedy Signal Recovery and Uniform Uncertainty Principles

Breaking the coherence barrier - A new theory for compressed sensing

COMPRESSED SENSING IN PYTHON

arxiv: v9 [cs.cv] 12 Mar 2013

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Rui ZHANG Song LI. Department of Mathematics, Zhejiang University, Hangzhou , P. R. China

Compressed Sensing and Sparse Recovery

Reconstruction from Anisotropic Random Measurements

GREEDY SIGNAL RECOVERY REVIEW

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

CoSaMP. Iterative signal recovery from incomplete and inaccurate samples. Joel A. Tropp

Sparse Optimization Lecture: Sparse Recovery Guarantees

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Compressive Sensing and Beyond

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Sparsity Regularization

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

AN INTRODUCTION TO COMPRESSIVE SENSING

Lecture 3. Random Fourier measurements

Constrained optimization

The Analysis Cosparse Model for Signals and Images

SGD and Randomized projection algorithms for overdetermined linear systems

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5

Sensing systems limited by constraints: physical size, time, cost, energy

IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 9, SEPTEMBER

Thresholds for the Recovery of Sparse Solutions via L1 Minimization

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Super-resolution via Convex Programming

Recent Developments in Compressed Sensing

Stability and robustness of l 1 -minimizations with Weibull matrices and redundant dictionaries

Interpolation via weighted l 1 minimization

Compressed Sensing and Redundant Dictionaries

Compressed Sensing and Redundant Dictionaries

An algebraic perspective on integer sparse recovery

Uniqueness Conditions for A Class of l 0 -Minimization Problems

Sparsifying Transform Learning for Compressed Sensing MRI

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Recovering any low-rank matrix, provably

Fast Hard Thresholding with Nesterov s Gradient Method

Compressive Sensing (CS)

Signal Sparsity Models: Theory and Applications

Interpolation-Based Trust-Region Methods for DFO

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit

A Few Basic Problems in Compressed Sensing

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design

Near-optimal Compressed Sensing Guarantees for Total Variation Minimization

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

An Introduction to Sparse Approximation

Near Optimal Signal Recovery from Random Projections

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit

SPECTRAL COMPRESSIVE SENSING WITH POLAR INTERPOLATION. Karsten Fyhn, Hamid Dadkhahi, Marco F. Duarte

Sparse and Low Rank Recovery via Null Space Properties

Solving Underdetermined Linear Equations and Overdetermined Quadratic Equations (using Convex Programming)

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

LEARNING DATA TRIAGE: LINEAR DECODING WORKS FOR COMPRESSIVE MRI. Yen-Huan Li and Volkan Cevher

Combining geometry and combinatorics

Optimization-based sparse recovery: Compressed sensing vs. super-resolution

Stable Signal Recovery from Incomplete and Inaccurate Measurements

Towards a Mathematical Theory of Super-resolution

6 Compressed Sensing and Sparse Recovery

Super-resolution by means of Beurling minimal extrapolation

One condition for all: solution uniqueness and robustness of l 1 -synthesis and l 1 -analysis minimizations

Abstract This paper is about the efficient solution of large-scale compressed sensing problems.

THEORY OF COMPRESSIVE SENSING VIA l 1 -MINIMIZATION: A NON-RIP ANALYSIS AND EXTENSIONS

The Sparsity Gap. Joel A. Tropp. Computing & Mathematical Sciences California Institute of Technology

Conditions for Robust Principal Component Analysis

WIRELESS sensor networks (WSNs) have attracted

arxiv: v1 [cs.it] 21 Feb 2013

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Minimizing the Difference of L 1 and L 2 Norms with Applications

TRACKING SOLUTIONS OF TIME VARYING LINEAR INVERSE PROBLEMS

Introduction to Compressed Sensing

of Orthogonal Matching Pursuit

Part IV Compressed Sensing

Compressed Sensing and Neural Networks

Compressive Sensing Theory and L1-Related Optimization Algorithms

Phase Transition Phenomenon in Sparse Approximation

Sparse Interactions: Identifying High-Dimensional Multilinear Systems via Compressed Sensing

Acceleration of Randomized Kaczmarz Method

Compressed sensing (CS), introduced by Candès, Romberg &

An iterative hard thresholding estimator for low rank matrix recovery

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1

Transcription:

Recovering overcomplete sparse representations from structured sensing Deanna Needell Claremont McKenna College Feb. 2015 Support: Alfred P. Sloan Foundation and NSF CAREER #1348721. Joint work with Felix Krahmer (Univ. of Göttingen) and Rachel Ward (UT Austin)

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 2 / 29

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 3 / 29

Imaging via Fourier measurements Magnetic Resonance Imaging (MRI): Imaging method for medical diagnostics Mathematical model: For an image f L 2 ([0, 1] 2 ), MRI measures 2D-Fourier series coefficients Ff (ω 1, ω 2 ) = f (x, y) exp(2πi(ω 1 x + ω 2 y))dxdy. Discretize to obtain expansion of a discrete image f C N N in the discrete Fourier basis consisting of the vectors ϕ ω1,ω 2 (t 1, t 2 ) = 1 N ei2π(t 1ω 1 +t 2 ω 2 )/N, N/2 + 1 t 1, t 2 N/2. Model can also be used for other applications. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 4 / 29

Compressed sensing for discrete images Model assumption: the image x C N2, is approximately s-sparse in a representation system {b i }, i.e., x s j=1 x k j b kj. Suitable systems: Wavelets, shearlets,... Intuition: Low dimensionality due to image structure, but nonlinear. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 5 / 29

Compressed sensing for discrete images Model assumption: the image x C N2, is approximately s-sparse in a representation system {b i }, i.e., x s j=1 x k j b kj. Suitable systems: Wavelets, shearlets,... Intuition: Low dimensionality due to image structure, but nonlinear. Goal: Reconstruction of x from m N 2 linear measurements, that is, from y = Ax, where A C m N2. Underdetermined system Many solutions. Sufficient condition for robust recovery via convex optimization: m s log α N random measurements Incoherence between measurements and basis elements. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 5 / 29

Important tool: The Restricted Isometry Property Definition (Candès-Romberg-Tao (2006)) A matrix A C m n has the Restricted Isometry Property of order s and level δ (0, 1) (in short, (s, δ)-rip), if one has (1 δ) x 2 2 Ax 2 2 (1 + δ) x 2 2 for all s-sparse x C n (1) The Restricted Isometry Constant δ s (A) is the smallest δ satisfying (1). Idea: Any submatrix of s columns is well-conditioned. Typical result: If A C m n has the (2k, δ)-rip with δ 1 2 and the equation y = Ax has a k-sparse solution x #, then one has x # = argmin Az=y z 1 (e.g., Cai et al. (2014)). Guarantee only works for basis representations Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 6 / 29

Compressed Sensing MRI - the basis case Donoho-Lustig-Pauly (2007): Use compressed sensing to reduce number of MRI measurements needed. Main issue to address: Lack of incoherence between Fourier measurements and good bases for image representation In this talk: address this issue, provide reconstruction guarantees Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 7 / 29

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 8 / 29

Dictionaries Definition A set D = {d 1,..., d N } C n is called a tight frame if there exists A > 0 such that for all x C n N i=1 d i, x 2 = A x 2 2 Interpretation: Dictionary, the d i s represent different features of the signal. Only few features active: Sparsity Redundancy N > n allows for sparser representations. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 9 / 29

Analysis vs. Synthesis Sparsity A signal x C n is synthesis s-sparse with respect to a dictionary D if there exist {z i } N i=1 such that x = N i=1 z id i. A signal x C n is analysis s-sparse with respect to a dictionary D if D x is s-sparse in C N. For many tight frames that appear in applications, synthesis sparse signals are also approximately analysis sparse. Definition For a dictionary D C n N and a sparsity level s, we define the localization factor as D Dz η s,d = η def = sup 1. Dz 2 =1, z 0 s s Assumption: Localization factor not too large. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 10 / 29

D-RIP If D is coherent, that is some d i are very close, it may be impossible to recover z even from complete knowledge of Dz. Consequently, RIP based guarantees cannot work. Important observation: No need to find z, x = Dz suffices. Definition (Candès-Eldar-N-Randall (2010)) Fix a dictionary D C n N and matrix A C m n. The matrix A satisfies the D-RIP with parameters δ and s if for all s-sparse vectors x C n. (1 δ) Dx 2 2 ADx 2 2 (1 + δ) Dx 2 2 Examples of D-RIP matrices: Subgaussian matrices [Candès-Eldar-N-Randall (2010)], RIP matrices with random column signs [Krahmer-Ward (2011)] Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 11 / 29

D-RIP based recovery guarantees Theorem (Candès-Eldar-N-Randall (2010)) For A that has the D-RIP and a measurement vector y = Ax = ADz consider the minimization problem Then the minimizer ˆx satisfies ˆx = argmin D x 1 subject to A x = y. x C n ˆx x 2 C D x (D x) s 1 s Small localization factor small error bound In this talk: provide D-RIP constructions closer to applications Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 12 / 29

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 13 / 29

Uniform sampling The mutual coherence of two bases {ϕ k } and {b j } is defined to be µ = sup b j, ϕ k. j,k Theorem (Rudelson-Vershynin (2006), Rauhut (2007)) Consider the matrix A = Φ Ω B C m N with entries A l,k = ϕ jl, b k, l [m], k [N], (2) where the ϕ jl are independent samples drawn uniformly at random from an ONB {ϕ j } N j=1 incoherent with the sparsity basis {b j} in the sense that µ KN 1/2. Then once, for some s log(n), m Cδ 2 K 2 s log 3 (s) log(n), (3) with probability at least 1 N γ log3 (s), the restricted isometry constant δ s of 1 m A satisfies δ s δ. The constants C, γ > 0 are universal. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 14 / 29

Variable density sampling [Lustig-Donoho-Pauly (2007)]: For a better performance with real images, one should be undersampling less near the k-space origin and more in the periphery of k-space. For example, one may choose samples randomly with sampling density scaling according to a power of distance from the origin. Idea by Puy-Vandergheynst-Wiaux (2011): Variable density sampling can reduce coherence. Strategy: Find optimal weights using convex optimization. Work with problem specific discretization level. No theoretical recovery guarantees. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 15 / 29

Local coherence Empirical observation of Puy et al.: Often only few Fourier basis vectors have high coherence with the sparsity basis. Changing the weights can compensate for this inhomogeneity. We introduce the local coherence to address this issue. Definition (Local coherence) The local coherence of an ONB {ϕ j } N j=1 of CN with respect to another ONB {ψ k } N k=1 of CN is the function µ loc (j) = sup 1 k N ϕ j, ψ k. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 16 / 29

RIP for variable density subsampling Theorem (Consequence of Rauhut-Ward 12) Assume the local coherence of an ONB Φ = {ϕ j } N j=1 with respect to an ONB Ψ = {ψ k } N k=1 is pointwise bounded by the function κ, that is, ϕ j, ψ k κ j. Consider the matrix A C m N with entries sup 1 k N A l,k = ϕ jl, ψ k, j [m], k [N], (4) where the j l are drawn independently according to ν l = P(l j = l) = Suppose that κ2 l κ 2 2. m Cδ 2 κ 2 2s log 3 (s) log(n), (5) and let D = diag(d j,j ), where d j,j = κ 2 /κ j. Then with probability at least 1 N γ log3 (s), the preconditioned matrix 1 m DA has a restricted isometry constant δ s δ. The constants C, γ > 0 are universal. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 17 / 29

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 18 / 29

Total variation Discrete image x = (x j,k ) C n2, (j, k) {1, 2,..., n} 2 := [n] 2 Discrete directional derivatives (x u ) j,k =x j,k+1 x j,k, (x v ) j,k =x j+1,k x j,k, Discrete gradient x = (x u, x v ) is very close to sparse. Not a basis representation, does not allow stable image reconstruction Total variation (TV): x TV = x l 1. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 19 / 29

Compressed sensing via TV minimization Proposition (N-Ward (2012)) Let (a j ) be an orthonormal basis for C n2 bivariate Haar basis (w j ), that is incoherent with the sup a j, w k C/n. j,k Let A u : C n2 C m consist of m s log 6 (n) uniformly subsampled bases a j as rows. Then with high probability, the following holds for all X C n2 : If y = A u X + ξ with ξ 2 ε and ˆX = argmin Z TV such that F u Z y 2 ε Z then ˆX X 2 ( X ( X ) s 1 s + ε Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 20 / 29

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 21 / 29

Local incoherence between Fourier and Haar systems Theorem (Krahmer-Ward (2014)) Fix δ (0, 1/3) and integers N = 2 p, m, and s such that m s log 3 s log 5 N. Select m discrete frequencies ( Ω j 1, 2) Ωj independently according to µ(ω 1, ω 2) := P [ ( (Ω j 1, Ωj 2 ) = (ω1, ω2)] min 1, C ω 2 1 +ω2 2 ), N 2 + 1 ω1, ω2 N 2, and let F Ω : C N2 C m be the DFT matrix restricted to { ( Ω j 1, Ωj 2) }. Then with high probability, the following holds for all f C N N. Given measurements y = F Ω f, the TV-minimizer f # = argmin g C N N g TV such that F Ω g = y, approximates f to within the best s-term approximation error of f : f f # 2 C f ( f )s 1 s. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 22 / 29

Numerical Simulations (a) Original image (b) Lowest frequencies only (c) Sample (k1 2 + k2 2 ) 1 (d) Sample (k1 2 + k2 2 ) 3/2 Figure : Reconstruction using m = 12, 000 noiseless partial DFT measurements with frequencies Ω = (k 1, k 2) sampled from various distributions. Relative reconstruction errors (b).18, (c).21, and (d).19 Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 23 / 29

Which sampling density to choose? Reconstruction error 1 0.5 0 0 1 2 3 4 5 6 inf Power used in sampling strategy (a) Reconstruction errors by various power-law density sampling at low noise (filled line) and high noise (dashed line) Are all the reconstructions of comparable quality? Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 24 / 29

Really all the same? 100 100 150 150 200 200 250 250 300 300 350 350 400 100 150 200 250 300 350 400 400 100 150 200 250 300 350 400 (b) The wet paint reconstructions indicated by the circled errors on the error plot, zoomed in on the paint sign. At high noise level, inverse quadratic-law sampling (left) still reconstructs fine details of the image better than low frequency-only sampling (right). Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 25 / 29

Theorem (Krahmer-N-Ward (2014)) Fix a sparsity level s < N, and constant 0 < δ < 1. Let D C n N be a tight frame, let A = {a 1,..., a n } be an ONB of C n, and κ R n + an entrywise upper bound of the local coherence, that is, µ loc i (A, D) = sup a i, d j κ i. j [N] Consider the localization energy η. Construct à C m n by sampling vectors from A at random according to the probability distribution ν given by ν(i) = κ2 i κ 2 2 and normalizing by n/m. Then as long as m Cδ 2 s κ 2 2(η) 2 log 3 (sη 2 ) log(n), and m Cδ 2 s κ 2 2η 2 log(1/γ) (6) then with probability 1 γ, à satisfies the D-RIP with parameters s and δ. Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 26 / 29

Consequences Recovery guarantees for Fourier measurements and Haar wavelet frames of redundancy 2 by previous local coherence analysis. Constant local coherence: Implies incoherence based guarantees (for example for oversampled Fourier dictionary). Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 27 / 29

Outline 1 Compressed Sensing and MRI 2 Compressed Sensing with Dictionaries 3 Beyond Incoherence 4 Total variation minimization 5 Main Results 6 Discussion Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 28 / 29

Summary and open questions Compressive imaging via variable density sampling Uniform recovery guarantees for approximately Haar- and Gradient-sparse images First guarantees for Fourier measurements and dictionary sparsity Goals: Basis-independent error bounds via continuous total variation. Why is cubic decay better than quadratic? Sharper error bounds using shearlets, curvelets What is the right noise model? Deanna Needell (CMC) Compressed Sensing with Dictionaries Feb. 2015 29 / 29