Sparse Legendre expansions via l 1 minimization

Similar documents
Sparse recovery for spherical harmonic expansions

Interpolation via weighted l 1 -minimization

Interpolation via weighted l 1 -minimization

Interpolation via weighted l 1 minimization

Recovering overcomplete sparse representations from structured sensing

AN INTRODUCTION TO COMPRESSIVE SENSING

Strengthened Sobolev inequalities for a random subspace of functions

Compressed Sensing and Sparse Recovery

Sparse solutions of underdetermined systems

Interpolation via weighted l 1 minimization

Interpolation-Based Trust-Region Methods for DFO

Greedy Signal Recovery and Uniform Uncertainty Principles

Lecture Notes 9: Constrained Optimization

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Reconstruction from Anisotropic Random Measurements

Lecture 22: More On Compressed Sensing

Constrained optimization

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

An Introduction to Sparse Approximation

The uniform uncertainty principle and compressed sensing Harmonic analysis and related topics, Seville December 5, 2008

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

GREEDY SIGNAL RECOVERY REVIEW

Sparse Recovery with Pre-Gaussian Random Matrices

Sparsity Regularization

Lecture 3. Random Fourier measurements

Sparse and Low Rank Recovery via Null Space Properties

Recent Developments in Compressed Sensing

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit

Combining geometry and combinatorics

Rapidly Computing Sparse Chebyshev and Legendre Coefficient Expansions via SFTs

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

Sparse analysis Lecture VII: Combining geometry and combinatorics, sparse matrices for sparse signal recovery

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

Random Coding for Fast Forward Modeling

Exponential decay of reconstruction error from binary measurements of sparse signals

Introduction to Compressed Sensing

Recall that any inner product space V has an associated norm defined by

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

6 Compressed Sensing and Sparse Recovery

CoSaMP. Iterative signal recovery from incomplete and inaccurate samples. Joel A. Tropp

CoSaMP: Greedy Signal Recovery and Uniform Uncertainty Principles

Geometry of log-concave Ensembles of random matrices

arxiv: v3 [math.na] 7 Dec 2018

Compressed Sensing and Redundant Dictionaries

Signal Recovery from Permuted Observations

Limitations in Approximating RIP

On the coherence barrier and analogue problems in compressed sensing

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design

Thresholds for the Recovery of Sparse Solutions via L1 Minimization

Extracting structured dynamical systems using sparse optimization with very few samples

Compressed Sensing and Redundant Dictionaries

Methods for sparse analysis of high-dimensional data, II

Toeplitz Compressed Sensing Matrices with. Applications to Sparse Channel Estimation

Sensing systems limited by constraints: physical size, time, cost, energy

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation

Sparse Optimization Lecture: Sparse Recovery Guarantees

ROBUST BLIND SPIKES DECONVOLUTION. Yuejie Chi. Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 43210

Sparse analysis Lecture V: From Sparse Approximation to Sparse Signal Recovery

Solution Recovery via L1 minimization: What are possible and Why?

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

An algebraic perspective on integer sparse recovery

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Primal Dual Pursuit A Homotopy based Algorithm for the Dantzig Selector

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina

Methods for sparse analysis of high-dimensional data, II

A Tutorial on Compressive Sensing. Simon Foucart Drexel University / University of Georgia

COMPRESSED SENSING IN PYTHON

(Part 1) High-dimensional statistics May / 41

STAT 200C: High-dimensional Statistics

Algorithms for sparse analysis Lecture I: Background on sparse approximation

Compressed Sensing - Near Optimal Recovery of Signals from Highly Incomplete Measurements

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing

Z Algorithmic Superpower Randomization October 15th, Lecture 12

Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit

Sparse Astronomical Data Analysis. Jean-Luc Starck. Collaborators: J. Bobin., F. Sureau. CEA Saclay

Signal Recovery, Uncertainty Relations, and Minkowski Dimension

Compressive Sensing Theory and L1-Related Optimization Algorithms

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Beyond incoherence and beyond sparsity: compressed sensing in the real world

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1

Multiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing

Reconstruction of sparse Legendre and Gegenbauer expansions

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

Optimization-based sparse recovery: Compressed sensing vs. super-resolution

Sparse Interactions: Identifying High-Dimensional Multilinear Systems via Compressed Sensing

Solution-recovery in l 1 -norm for non-square linear systems: deterministic conditions and open questions

Stability and robustness of l 1 -minimizations with Weibull matrices and redundant dictionaries

The Pros and Cons of Compressive Sensing

COMPRESSED Sensing (CS) is a method to recover a

An iterative hard thresholding estimator for low rank matrix recovery

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Fast Reconstruction Algorithms for Deterministic Sensing Matrices and Applications

Tutorial: Sparse Signal Recovery

Uncertainty quantification for sparse solutions of random PDEs

A class of non-convex polytopes that admit no orthonormal basis of exponentials

III. Quantum ergodicity on graphs, perspectives

Random regular digraphs: singularity and spectrum

Transcription:

Sparse Legendre expansions via l 1 minimization Rachel Ward, Courant Institute, NYU Joint work with Holger Rauhut, Hausdorff Center for Mathematics, Bonn, Germany. June 8, 2010

Outline Sparse recovery for bounded orthonormal systems Sparse recovery for (certain) unbounded orthonormal systems From compressive sensing to function approximation Applications

Outline Sparse recovery for bounded orthonormal systems Sparse recovery for (certain) unbounded orthonormal systems From compressive sensing to function approximation Applications

Problem set-up Consider the general problem: reconstruct an s-sparse vector x C N or x R N (or a compressible vector) from its vector of m measurements y = Ax Suppose that the measurements correspond to underdetermined system: s < m < N, so that the system y = Ax has infinitely many solutions. We want a sparse solution. Preferably we would like to have a fast algorithm that performs the sparse reconstruction.

5 Algorithms for sparse recovery l 0 -minimization: x 0 := supp x, min x 0 subject to Ax = y. x C N Problem: l 0 -minimization is NP hard. l 1 minimization: min x x 1 = N x j subject to Ax = y j=1 This is the convex relaxation of l 0 -minimization problem, can be solved efficiently.

A sufficient condition for sparse recovery The restricted isometry constant δ s of a matrix A C m N is defined as the smallest δ s such that for all s-sparse x C N. (1 δ s ) x 2 2 Ax 2 2 (1 + δ s ) x 2 2 (Candes, Romberg, Tao 2005, Foucart/Lai 2009): If A R m N satisfies RIP of order 2s, and if δ 2s.46..., then any vector x R N may be approximated from y = Ax + ξ with ξ 2 ɛ by up to error x = min z z 1 subject to Az y 2 ɛ x x 2 C 1 x x s 1 s + C 2 ɛ.

Which matrices satisfy the restricted isometry property? If A R m N is a Gaussian or Bernoulli random matrix, m = O(s ln(n/s)) measurements ensure that A satisfy the restricted isometry property of order s with high probability. In contrast, all available deterministic constructions run into quadratic bottleneck: require m = O(s 2 log N) measurements. Compromise between completely random and deterministic: structured random matrices.

Structured random matrices: the random partial Fourier matrix Consider the discrete Fourier matrix F C N N with entries F l,k = 1 N e 2πilk/N, l, k = 1,..., N, Randomly subsample m < N rows from F to get the m N random partial Fourier matrix. The measurements y l = (Fx) l are a random subsample of the DFT of x. Theorem (Candes, Romberg, Tao 2005; Rudelson, Vershynin 2008) If m O(s ln 4 (N)), then with high probability, the m N random partial Fourier matrix satisfies the restricted isometry property of order s.

The random partial Fourier matrix: Function approximation interpretation Random partial Fourier measurements correspond to samples of a trigonometric polynomial: y l = (Fx) l = g x (t l ), t l = 2πj l /N, where g x (t) := N 1 k=0 x ke ikt. We say that g x (t) is an s-sparse trigonometric polynomial of degree N 1 if the coefficient vector (x k ) N 1 k=0 is s-sparse. RIP of the random partial Fourier matrix means that all s-sparse trigonometric polynomials of degree N can be recovered from their values on a fixed subset of m = O(s log 4 N) measurements of the form t l = 2πl/N.

Outline Sparse recovery for bounded orthonormal systems Sparse recovery for (certain) unbounded orthonormal systems From compressive sensing to function approximation Applications

Sparse recovery for more general orthonormal systems? We may consider the recovery of functions f (u) = N 1 k=0 x k ψ k (u), u D R d that are sparse relative to a function system ψ 1,..., ψ N : D C that is orthonormal with respect to measure ν on D, from a number m < N of point samples y l = f (u l ) = N x k ψ k (u l ), u 1,..., u m ν. k=0 where m is proportional to the sparsity level s. Whether sparse recovery is possible depends on the structure of the sampling matrix A l,k = ψ k (u l ) associated to the orthonormal system

Sparse recovery for general bounded orthonormal systems Sampling matrix: A l,k = ψ k (u l ), l [m], k [N] Theorem (Rauhut 09) Suppose that ψ k K for all k [N]. Let A C m N be the sampling matrix associated to the bounded orthonormal system {ψ j } N 1 j=0 with sampling points u 1,..., u m D chosen i.i.d. from the orthogonalization measure ν. If m O(K 2 s ln 4 (N)), then with probability exceeding 1 N log3 N, the matrix 1 m A satisfies RIP of order s. Examples of uniformly bounded orthonormal systems: The complex exponential basis (with uniform orthogonalization measure) and Chebyshev polynomial basis, orthonormal on [ 1, 1] w.r.t. the Chebyshev measure.

The Legendre polynomials The Legendre polynomials P l are orthonormal on D = [ 1, 1] with respect to Lebesgue measure. They are generated via Gram-Schmidt orthogonalization on the monomial system 1, x, x 2,... They are used in schemes for solving several classical PDEs, such as Laplace s equation Fast Legendre polynomial transform: requires only O(N log N) arithmetic operations

4 The Legendre polynomials The Legendre polynomials satisfy P l = 2l + 1, so K = sup 0 l N 1 P l = 2N 1. From existing theory: s-sparse polynomials of the form f (t) = N k=0 x kp k (t) may be recovered from its values at m = O(sK 2 ln 4 (N)) = O(sN ln 4 N) sampling points. this is trivial!

The Legendre polynomials The Legendre polynomials satisfy P l = 2l + 1, so K = sup 0 l N 1 P l = 2N 1. From existing theory: s-sparse polynomials of the form f (t) = N k=0 x kp k (t) may be recovered from its values at m = O(sK 2 ln 4 (N)) = O(sN ln 4 N) sampling points. this is trivial!

Sparse recovery for Legendre orthonormal systems Theorem (Rauhut, W 10) Let N and s be given. Suppose that m = O(s log 4 N) sampling points t 1,..., t m are drawn from the Chebyshev measure dν = π 1 (1 t 2 ) 1/2 dt on [ 1, 1]. Then with probability exceeding 1 N log3 N, the following holds for all polynomials f (t) = N 1 k=0 x kp k (t): Suppose that noisy sample values (f (t 1 ) + η 1,..., f (t m ) + η m ) are observed, and η 2 mɛ. Then the coefficient vector x = (x 0, x 1,..., x N 1 ) is approximated by x := min z z 1 subject to Az y 2 ɛ up to error x x 2 C 1 x x s 1 s + C 2 ɛ.

17 Numerical Example A sparse Legendre polynomial with sparsity s = 5, maximal degree N = 80, and n = 20 i.i.d. sampling points from the Chebyshev measure. Reconstruction by l 1 -minimization is exact! 3 2 1 0 1 2 3 1 0.5 0 0.5 1

Numerical Example 3 2 1 0 1 2 3 1 0.5 0 0.5 1 The same sparse Legendre polynomial (black) subject to noisy random measurements, and the stable approximation obtained by l 1 -minimization.

19 Key idea in proof: Premultiplication The Legendre polynomials grow uniformly as they approach the endpoints: P n (t) < 2π 1/2 (1 t 2 ) 1/4, 1 t 1; (Constant due to S. Bernstein)

20 Key idea in proof: Premultiplication The Legendre polynomials grow uniformly as they approach the endpoints: P n (t) < 2π 1/2 2π 1/2 (1 t 2 ) 1/4, 1 t 1; (Constant due to S. Bernstein) So the functions Q n (t) = π 2 (1 t2 ) 1/4 P n (t) form a uniformly bounded system (and Q n 2). Also, the Q n s are orthonormal with respect to the Chebyshev measure dν = π 1 (1 t 2 ) 1/2 dt on [ 1, 1]: 1 1 Q n (t)q m (t)π 1 (1 t 2 ) 1/2 dt = 1 1 P n (t)p m (t)dt = δ n,m.

Outline Sparse recovery for bounded orthonormal systems Sparse recovery for (certain) unbounded orthonormal systems From compressive sensing to function approximation Applications

22 Universality of the Chebyshev measure Consider a probability measure ν(t)dt on [ 1, 1]. (Szego): If ν(t) satisfies a mild continuity condition, then the polynomials p ν n that are orthonormal w.r.t. ν will still satisfy a uniform growth condition: p ν n(t) C ν (1 t 2 ) 1/4 ν(t) 1/2, as before, q ν n(t) = (1 t 2 ) 1/4 ν(t) 1/2 p ν n(t) is a uniformly bounded system and orthonormal w.r.t. Chebyshev measure (1 t 2 ) 1/2 dt.

From compressive sensing to function approximation Consider the weighted norm on continuous functions in [ 1, 1], f,w = and the error sup f (t) w(t), w(t) = (1 t 2 ) 1/4 ν(t), t [ 1,1] N 1 σ N,s (f ),w = inf{ f x k pk ν,w : (x k ) R N, x 0 s} k=0 Theorem (Rauhut and W, 10) Let N, m, s be given with m = O(s log 4 N). Then there exist sampling points t 1,..., t m (chosen according to Chebyshev distribution) and an efficient reconstruction procedure (l 1 minimization) such that for any continuous function f, the polynomial P of degree at most N 1 reconstructed from f (t 1 ),..., f (t m ) satisfies 23 f P,w C w s σn,s (f ),w.

From compressive sensing to function approximation Consider the weighted norm on continuous functions in [ 1, 1], f,w = and the error sup f (t) w(t), w(t) = (1 t 2 ) 1/4 ν(t), t [ 1,1] N 1 σ N,s (f ),w = inf{ f x k pk ν,w : (x k ) R N, x 0 s} k=0 Theorem (Rauhut and W, 10) Let N, m, s be given with m = O(s log 4 N). Then there exist sampling points t 1,..., t m (chosen according to Chebyshev distribution) and an efficient reconstruction procedure (l 1 minimization) such that for any continuous function f, the polynomial P of degree at most N 1 reconstructed from f (t 1 ),..., f (t m ) satisfies 24 f P,w C w s σn,s (f ),w.

Extension to spherical harmonics Re(Y 0 0 ), Re(Y 0 1 ), Re(Y 0 2 ). Re(Y 1 2 ), Re(Y 0 3 ), Re(Y 1 3 ). Y m l (φ, θ) = C l,mp m l ( cos(θ) ) e imφ, l m l, l = 0, 1,... P m l s are the associated Legendre functions. Fast Spherical harmonic transform: requires only O(N(log N) 2 ) arithmetic operations

26 Extension to spherical harmonics The associated Legendre functions Pl m may be written in terms of orthonormal polynomials with respect to the positive weights ν α (x) = (1 x 2 ) α, α = 0, 1,..., n. Krasikov 08: (1 x 2 ) 1/4 ν α (x) 1/2 p α n (x) C α O(α 1/4 ) from this one can derive... Corollary Let N, m, s be given with m = O(sN 1/4 log 4 N). Then there exist sampling points t 1,..., t m (chosen according to the spherical Chebyshev distribution) and an efficient reconstruction procedure (l 1 minimization) such that any function on the sphere which is an s-sparse expansion in the first N spherical harmonic basis functions may be exactly recovered from these sampling points.

Outline Sparse recovery for bounded orthonormal systems Sparse recovery for (certain) unbounded orthonormal systems From compressive sensing to function approximation Applications

8 Application: Cosmic Microwave Background Radiation (CMB) map WMAP full-sky temperature anisotropy map (Source: http://map.gsfc.nasa.gov) T (θ, φ) = l l=0 m= l a l,myl m (θ, φ) where the Y s are the spherical harmonics. Red band: measurements are corrupted, interference with galactic foreground signal.

29 CMB map is compressible in spherical harmonics Source: Lyman Page, http://ophelia.princeton.edu/ a a s 2 / a 2, s = 1,..., n 2 Consider the coefficient vector a = (a l,m ) in T (θ, φ) n l l=0 m= l a l,myl m (θ, φ). This vector is predicted and observed to be compressible in the spherical harmonic basis.

CMB map is compressible in spherical harmonics (J. Starck et. al., 08): 1 Propose full-sky CMB map inpainting from partial CMB measurements T (θ j, φ k ) = m j,k. Obtain the coefficients a = (a l,m ) by solving the l 1 minimization problem, a = arg min N l l=0 m= l z l,m s.t. N l l=0 m= l z l,m Y l,m (θ j, φ k ) = m j,k 1 Source: P. Abrial, Y. Moudden, J. Starck, J. Fadili, J. Delabrouille, and M. Nguyen. CMB data analysis and sparsity. Stat. Methodol., 2008.

Final remarks Our results provide some theoretical justification for the good inpainting results obtained for the CMB map. Open questions: 1. Sparse recovery in spherical harmonic basis: can we get better recovery guarantees? 2. Can we generalize these results to functions sparse in bases that are eigenfunction solutions to a more general class of PDE THANK YOU

Final remarks Our results provide some theoretical justification for the good inpainting results obtained for the CMB map. Open questions: 1. Sparse recovery in spherical harmonic basis: can we get better recovery guarantees? 2. Can we generalize these results to functions sparse in bases that are eigenfunction solutions to a more general class of PDE THANK YOU