Sensing systems limited by constraints: physical size, time, cost, energy

Similar documents
New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

An Introduction to Sparse Approximation

Sparse analysis Lecture III: Dictionary geometry and greedy algorithms

Introduction to Compressed Sensing

Greedy Signal Recovery and Uniform Uncertainty Principles

Compressive Sensing and Beyond

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Strengthened Sobolev inequalities for a random subspace of functions

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Randomness-in-Structured Ensembles for Compressed Sensing of Images

Combining geometry and combinatorics

Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5

Sparse analysis Lecture VII: Combining geometry and combinatorics, sparse matrices for sparse signal recovery

Multipath Matching Pursuit

Sparse analysis Lecture V: From Sparse Approximation to Sparse Signal Recovery

Lecture 22: More On Compressed Sensing

Machine Learning for Signal Processing Sparse and Overcomplete Representations. Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013

CoSaMP. Iterative signal recovery from incomplete and inaccurate samples. Joel A. Tropp

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit

of Orthogonal Matching Pursuit

A new method on deterministic construction of the measurement matrix in compressed sensing

Machine Learning for Signal Processing Sparse and Overcomplete Representations

Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France

CoSaMP: Greedy Signal Recovery and Uniform Uncertainty Principles

Signal Recovery from Permuted Observations

AN INTRODUCTION TO COMPRESSIVE SENSING

Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit

Stability and Robustness of Weak Orthogonal Matching Pursuits

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk

Sparse Optimization Lecture: Sparse Recovery Guarantees

Recovering overcomplete sparse representations from structured sensing

The Pros and Cons of Compressive Sensing

Stable Signal Recovery from Incomplete and Inaccurate Measurements

Simultaneous Sparsity

Generalized Orthogonal Matching Pursuit- A Review and Some

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Designing Information Devices and Systems I Discussion 13B

Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Final Presentation

Pre-weighted Matching Pursuit Algorithms for Sparse Recovery

Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France.

Sparse & Redundant Signal Representation, and its Role in Image Processing

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

COMPRESSIVE SAMPLING USING EM ALGORITHM. Technical Report No: ASU/2014/4

Part IV Compressed Sensing

c 2011 International Press Vol. 18, No. 1, pp , March DENNIS TREDE

L-statistics based Modification of Reconstruction Algorithms for Compressive Sensing in the Presence of Impulse Noise

Sparse Solutions of Systems of Equations and Sparse Modelling of Signals and Images

Compressed sensing techniques for hyperspectral image recovery

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit

Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images!

Compressed Sensing. 1 Introduction. 2 Design of Measurement Matrices

Sparsity in Underdetermined Systems

A NEW FRAMEWORK FOR DESIGNING INCOHERENT SPARSIFYING DICTIONARIES

EE 381V: Large Scale Optimization Fall Lecture 24 April 11

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

SPARSE signal representations have gained popularity in recent

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise

Primal Dual Pursuit A Homotopy based Algorithm for the Dantzig Selector

New Applications of Sparse Methods in Physics. Ra Inta, Centre for Gravitational Physics, The Australian National University

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

Sparse linear models

Signal Sparsity Models: Theory and Applications

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Compressed Sensing and Related Learning Problems

GREEDY SIGNAL RECOVERY REVIEW

Signal Recovery, Uncertainty Relations, and Minkowski Dimension

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 9, SEPTEMBER

Compressed Sensing and Sparse Recovery

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Interpolation via weighted l 1 minimization

Sparse Solutions of an Undetermined Linear System

The Analysis Cosparse Model for Signals and Images

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Compressed Sensing Using Reed- Solomon and Q-Ary LDPC Codes

Compressed Sensing with Very Sparse Gaussian Random Projections

COMPRESSED SENSING IN PYTHON

Exact Topology Identification of Large-Scale Interconnected Dynamical Systems from Compressive Observations

Sparse recovery using sparse random matrices

Recent Developments in Compressed Sensing

Minimizing the Difference of L 1 and L 2 Norms with Applications

Constrained optimization

Compressive Sensing Applied to Full-wave Form Inversion

COMPRESSIVE SAMPLING OF SOUND FIELDS USING MOVING MICROPHONES. Fabrice Katzberg, Radoslaw Mazur, Marco Maass, Philipp Koch, and Alfred Mertins

Computable Performance Analysis of Sparsity Recovery with Applications

Sparse Legendre expansions via l 1 minimization

Lecture 16: Compressed Sensing

A Tutorial on Compressive Sensing. Simon Foucart Drexel University / University of Georgia

A tutorial on sparse modeling. Outline:

Super-resolution via Convex Programming

Block-sparse Solutions using Kernel Block RIP and its Application to Group Lasso

On Rank Awareness, Thresholding, and MUSIC for Joint Sparse Recovery

On the Minimization Over Sparse Symmetric Sets: Projections, O. Projections, Optimality Conditions and Algorithms

Optimization-based sparse recovery: Compressed sensing vs. super-resolution

Recovery of Sparse Signals Using Multiple Orthogonal Least Squares

Interpolation via weighted l 1 -minimization

Transcription:

Rebecca Willett

Sensing systems limited by constraints: physical size, time, cost, energy Reduce the number of measurements needed for reconstruction Higher accuracy data subject to constraints

Original Scene Downsampled Reconstruction from ¼ as many measurements

Original Scene Downsampled Reconstruction from ¼ as many measurements

= + y = f + n Each observation is a measurement of ONE pixel

y 1 = f, I 1 =, y 2 = f, I 2 =,. y N = f, I N =, Each observation is a measurement of ONE pixel

Images are compressible Measuring all pixels inherently wasteful

y 1 = f, r 1 =, Measure sum of half the pixels Narrow down star location

y 1 = f, r 1 =, y 2 = f, r 2 =,. y M = f, r M =, Each observation is a measurement of half the pixels

y 1 = f, r 1 =, y 2 = f, r 2 =,. y M = f, r M =, These ideas extend to multiple stars and random combinations of pixels

y = Rf + n Observations All k random projections True image Noise

= + n y R f System is underdetermined: infinitely many solutions

Assume f is K-sparse or β-compressible in some basis Ψ. That is, and either f = N i=1 θ i ψ i θ 0 K or f f K K β where f K is the best K-term approximation of f in the basis Ψ.

= f Ψ θ

Combining y = Rf + n with f = Ψθ: = + n y RΨ A θ

θ =argmin θ f = Ψ θ y RΨθ 2 2 data fit + τθ 1 sparsity Key theory: If R meets certain conditions and f is sparse or compressible in Ψ, then f will be very accurate even when the number of measurements is small relative to N.

CONVENTIONAL SENSING Noisy Image COMPRESSED SENSING Random Projections Smaller Less Data Cheaper

Definition: Restricted Isometry Property. The matrix A satisfies the Restricted Isometry Property of order K with parameter δ K [0, 1) if (1 δ K )θ 2 2 Aθ2 2 (1 + δ K)θ 2 2 holds simultaneously for all K-sparse vectors θ. Matrices with this property are denoted RIP(K, δ K ). Candes and Tao (2006)

For example, if the entries of A are independent and identically distributed according to A i,j N 0, 1 M or A i,j = M 1/2 with probability 1 M 1/2 with probability 1 then A satisfies RIP(K, δ K ) with high probability for any integer K = O(M/ log N).

Matrices which satisfy the RIP combined with sparse recovery algorithms are guaranteed to yield accurate estimates of the underlying function f, as specified by the following theorem. Theorem: Noisy Sparse Recovery with RIP Matrices. Let A be a matrix satisfying RIP(2K, δ 2K ) with δ 2K < 2 1, and let y = Aθ + n be a vector of noisy observations of any signal θ R N, where the n is a noise or error term with n 2. Let θ K be the best K-sparse approximation of θ. Then the estimate obeys θ =argmin θ θ 1 subject to y Aθ 2 θ θ2 C 1,K + C 2,K θ θ K 1 K, where C 1,K and C 2,K are constants which depend on K but not on N or M. Candes (2006)

θ =argmin θ y A θ 2 2 + τ θ 1. This estimate can be computed in a variety of ways. Many off-the-shelf optimization software packages are unsuitable Can t handle large N Our objective isn t differentiable Don t exploit fast transforms (e.g. Fourier and wavelet) Gradient projection methods Introduce additional variables and recast problem as constrained optimization with differentiable objective Projection onto constraint set can be done with thresholding More robust to noise Orthogonal matching pursuits (OMP) Start with estimate = 0 Greedily choose elements of estimate to have non-zero magnitude by iteratively processing residual errors Very fast when little noise

Our objective is θ =argmin θ y A θ 2 2 + τ θ 1. The first term can be re-written as and its gradient is y T y 2 θ T A T y + θ T A T A θ 2A T (y A θ). This suggests a simple strategy for computing θ: start with an initial estimate θ, update it by adding a step in the negative gradient direction, then apply thresholding!

Start with some initial estimate θ (0) ; see how well it fits y: y Aθ (0). Use this residual to update the initial estimate: θ (0) + A T y Aθ (0). Impose sparsity via thresholding this estimate: θ (1) = threshold θ (0) + A T y Aθ (0) Repeat until y Aθ (i) is small: θ (i+1) = threshold θ (i) + A T y Aθ (i).

Original θ, with K = 15 f (blue) and y (red circles); M = 30 perfect reconstruction!

Matrices which satisfy the RIP combined with sparse recovery algorithms are guaranteed to yield accurate estimates of the underlying function f, as specified by the following theorem. Theorem: Noisy Sparse Recovery with RIP Matrices. Let A be a matrix satisfying RIP(2K, δ 2K ) with δ 2K < 2 1, and let y = Aθ + n be a vector of noisy observations of any signal θ R N, where the n is a noise or error term with n 2. Let θ K be the best K-sparse approximation of θ. Then the estimate obeys θ =argmin θ θ 1 subject to y Aθ 2 θ θ2 C 1,K + C 2,K θ θ K 1 K, where C 1,K and C 2,K are constants which depend on K but not on N or M. Candes (2006), Candes, Romberg & Tao (2006)

Let h = θ θ be our error vector. Let T 0 be the indicies of the largest K elements of θ, T 1 be the indicies of the largest K elements of h T, T c 2 be the indicies of the next K largest 0 elements of h T, and so on. For a vector x, let x c 0 T j be defined via xi, i T j x Tj,i =. 0, i / T j Then h = h T0 + h T1 + h T2 +... There are two main steps to our proof: θ θ2 = h 2 h T0 T 1 2 + h (T0 T 1 ) c 2 (STEP 1) Ch T0 T 1 2 + CK 1/2 θ θ K 1 (STEP 2) C + CK 1/2 θ θ K 1 C will represent constants which may depend on K but not N or M.

h (T0 T 1 ) c 2 = j 2 h Tj 2 j 2 h Tj 2 (remember me later!!) j 2 K 1/2 h Tj j 2 K 1/2 h Tj 1 1 /K = K 1/2 (h T1 1 + h T2 1 +...) = K 1/2 h T c 0 1 how big??

First note Rearranging terms we find θ 1 θ1 = θ + h 1 Putting everything together we have θ T0 1 h T0 1 + h T c 0 θ T c 0 1 h T c 0 1 h T0 1 +2θ T c 0 1 = h T0 1 +2θ θ K 1 h (T0 T 1 ) c 2 K 1/2 (h T0 1 +2θ θ K 1 ) as desired for Step 1. h T0 T 1 2 +2K 1/2 θ θ K 1

We now need to bound h T0 T 1 2. Note (1 δ 2K )h T0 T 1 2 2 Ah T0 T 1 2 2 = Ah T0 T 1,Ah Ah T0 T 1, j 2 Ah Tj For the first term Ah T0 T 1,Ah Ah T0 T 1 2 Ah 2 ( 1+δ 2K h T0 T 1 2 )A( θ θ)2 ( 1+δ 2K h T0 T 1 2 )(A θ y2 + y Aθ 2 ) ( 1+δ 2K h T0 T 1 2 )2 The second term is bounded similarly by Ah T0 T 1, Ah Tj 2δ 2K h Tj 2 h T0 T 1 2 j 2 j 2

Thus (1 δ 2K )h T0 T 1 2 2 h T0 T 1 2 2 1+δ 2K + 2δ 2K h Tj 2 h T0 T 1 2 C + CK 1/2 θ θ K 1. Putting it all together we have j 2 as desired. θ θ2 C + CK 1/2 θ θ K 1

In other words, the accuracy of the reconstruction of a general image f from measurements collected using a system which satisfies the RIP depends on (a) the amount of noise present and (b) how well f may be approximated by an image sparse in Ψ. If we have no noise ( =0) and our signal is K-sparse, then we have θ = θ; i.e., we can perfectly reconstruct the original signal!

1 0.8 Unsolvable; too little data or too little sparsity ρ=k/m 0.6 0.4 0.2 Solvable; sufficient data and sparsity 0 0 0.2 0.4 0.6 0.8 1 δ=m/n Donoho and Tanner (2010)

Consider the worst-case coherence of A RΨ. Formally, one denotes the Gram matrix G = A T A and let µ(a) = max 1 i,j i=j G i,j be the largest off-diagonal element of the Gram matrix. A good goal in designing a sensing matrix is to therefore choose R and Ψ so that µ is as close as possible to N 1/2. Theorem: Noisy Sparse Recovery with Incoherent Matrices. Let y = Aθ + n be a vector of noisy observations of any K-sparse signal θ R N, where K (µ(a) 1 + 1)/4 and the n is a noise or error term with n 2. Then our estimate obeys θ θ 2 2 4 2 1 µ(a)(4k 1). Tropp (2004), Donoho, Elad & Temlyakov (2006), Donoho & Huo (2001), Gilbert, Muthukrishnan & Strauss (2003)

Sparse vector Projection vectors Signal is locally concentrated, measurements are global Each measurement contains a little information about each component

What are the major open problems and areas of research? In what ways can these concepts be generalized to other problem domains?