Self-Calibration and Biconvex Compressive Sensing

Similar documents
Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

ROBUST BLIND SPIKES DECONVOLUTION. Yuejie Chi. Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 43210

Solving Corrupted Quadratic Equations, Provably

Blind Deconvolution Using Convex Programming. Jiaming Cao

Strengthened Sobolev inequalities for a random subspace of functions

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

Solving Underdetermined Linear Equations and Overdetermined Quadratic Equations (using Convex Programming)

Recovery of Low Rank and Jointly Sparse. Matrices with Two Sampling Matrices

Robust multichannel sparse recovery

Introduction to Compressed Sensing

CSC 576: Variants of Sparse Learning

Recovery of Simultaneously Structured Models using Convex Optimization

Covariance Sketching via Quadratic Sampling

Sparse Solutions of an Undetermined Linear System

Performance Analysis for Strong Interference Remove of Fast Moving Target in Linear Array Antenna

Robust Principal Component Analysis

SPARSE signal representations have gained popularity in recent

arxiv: v1 [cs.it] 25 Jan 2015

Fast and Robust Phase Retrieval

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation

Massive MIMO: Signal Structure, Efficient Processing, and Open Problems II

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

arxiv: v2 [cs.it] 7 Jan 2017

Recent Developments in Compressed Sensing

Phase Retrieval from Local Measurements: Deterministic Measurement Constructions and Efficient Recovery Algorithms

Minimizing the Difference of L 1 and L 2 Norms with Applications

Coherent imaging without phases

Tractable Upper Bounds on the Restricted Isometry Constant

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Recovering overcomplete sparse representations from structured sensing

Robust Sparse Recovery via Non-Convex Optimization

The Pros and Cons of Compressive Sensing

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Coprime Coarray Interpolation for DOA Estimation via Nuclear Norm Minimization

Nonconvex Methods for Phase Retrieval

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Compressed Sensing and Neural Networks

Compressive Sensing under Matrix Uncertainties: An Approximate Message Passing Approach

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

A New Estimate of Restricted Isometry Constants for Sparse Solutions

Real-Valued Khatri-Rao Subspace Approaches on the ULA and a New Nested Array

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Structured matrix factorizations. Example: Eigenfaces

Solution Recovery via L1 minimization: What are possible and Why?

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University

SCRIBERS: SOROOSH SHAFIEEZADEH-ABADEH, MICHAËL DEFFERRARD

CoSaMP. Iterative signal recovery from incomplete and inaccurate samples. Joel A. Tropp

Towards a Mathematical Theory of Super-resolution

ADAPTIVE ANTENNAS. SPATIAL BF

Research Article Support Recovery of Greedy Block Coordinate Descent Using the Near Orthogonality Property

Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil

MAXIMUM A POSTERIORI ESTIMATION OF SIGNAL RANK. PO Box 1500, Edinburgh 5111, Australia. Arizona State University, Tempe AZ USA

Self-Calibration and Biconvex Compressive Sensing

GREEDY SIGNAL RECOVERY REVIEW

Enhanced Compressive Sensing and More

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Sparse and Low Rank Recovery via Null Space Properties

Compressed Sensing and Sparse Recovery

Pre-weighted Matching Pursuit Algorithms for Sparse Recovery

arxiv: v1 [cs.it] 21 Feb 2013

Nonlinear Analysis with Frames. Part III: Algorithms

Exact Joint Sparse Frequency Recovery via Optimization Methods

Super-resolution via Convex Programming

EUSIPCO

MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING. Kaitlyn Beaudet and Douglas Cochran

SIGNALS with sparse representations can be recovered

Statistical Issues in Searches: Photon Science Response. Rebecca Willett, Duke University

Gauge optimization and duality

EXTENSION OF NESTED ARRAYS WITH THE FOURTH-ORDER DIFFERENCE CO-ARRAY ENHANCEMENT

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

THE estimation of covariance matrices is a crucial component

Multichannel Sparse Blind Deconvolution on the Sphere

Joint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

Sensing systems limited by constraints: physical size, time, cost, energy

Least squares regularized or constrained by L0: relationship between their global minimizers. Mila Nikolova

An iterative hard thresholding estimator for low rank matrix recovery

Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach

Sparse Recovery Beyond Compressed Sensing

Exploiting Sparsity for Wireless Communications

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Characterization of Gradient Dominance and Regularity Conditions for Neural Networks

An algebraic perspective on integer sparse recovery

Compressed Sensing and Robust Recovery of Low Rank Matrices

Recovery Guarantees for Rank Aware Pursuits

Recovering any low-rank matrix, provably

Robust Subspace DOA Estimation for Wireless Communications

A New High-Resolution and Stable MV-SVD Algorithm for Coherent Signals Detection

Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 26, 2013

Conditions for Robust Principal Component Analysis

Blind Compressed Sensing

Jointly Low-Rank and Bisparse Recovery: Questions and Partial Answers

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison

X. Zhang, G. Feng, and D. Xu Department of Electronic Engineering Nanjing University of Aeronautics & Astronautics Nanjing , China

Optimization for Compressed Sensing

Compressive Sensing of Sparse Tensor

Phase recovery with PhaseCut and the wavelet transform case

Transcription:

Self-Calibration and Biconvex Compressive Sensing Shuyang Ling Department of Mathematics, UC Davis July 12, 2017 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 1 / 22

Acknowledgements Research in collaboration with: Prof.Thomas Strohmer (UC Davis) This work is sponsored by NSF-DMS and DARPA. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 2 / 22

Outline (a) Self-calibration and mathematical framework (b) Biconvex compressive sensing in array signal processing (c) SparseLift: a convex approach towards biconvex compressive sensing (d) Theory and numerical simulations Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 3 / 22

Calibration Calibration: Calibration is to adjust one device with the standard one. Why? To reduce or eliminate bias and inaccuracy. Difficult or even impossible to calibrate high-performance hardware. Self-calibration: Equip sensors with a smart algorithm which takes care of calibration automatically. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 4 / 22

Calibration Calibration: Calibration is to adjust one device with the standard one. Why? To reduce or eliminate bias and inaccuracy. Difficult or even impossible to calibrate high-performance hardware. Self-calibration: Equip sensors with a smart algorithm which takes care of calibration automatically. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 4 / 22

Calibration realized by machine? Uncalibrated device leads to imperfect sensing We encounter imperfect sensing all the time: the sensing matrix A(h) depending on an unknown calibration parameter h, y = A(h)x + w. This is too general to solve for h and x jointly. Examples: Phase retrieval problem: h is the unknown phase of the Fourier transform of x. Cryo-electron microscopy images: h can be the unknown orientation of a protein molecule and x is the particle. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 5 / 22

Calibration realized by machine? Uncalibrated device leads to imperfect sensing We encounter imperfect sensing all the time: the sensing matrix A(h) depending on an unknown calibration parameter h, y = A(h)x + w. This is too general to solve for h and x jointly. Examples: Phase retrieval problem: h is the unknown phase of the Fourier transform of x. Cryo-electron microscopy images: h can be the unknown orientation of a protein molecule and x is the particle. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 5 / 22

A simplified but important model Our focus: One special case is to assume A(h) to be of the form A(h) = D(h)A where D(h) is an unknown diagonal matrix. However, this seemingly simple model is very useful and mathematically nontrivial to analyze. Phase and gain calibration in array signal processing Blind deconvolution (image deblurring; joint channel and signal estimation, etc.) Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 6 / 22

A simplified but important model Our focus: One special case is to assume A(h) to be of the form A(h) = D(h)A where D(h) is an unknown diagonal matrix. However, this seemingly simple model is very useful and mathematically nontrivial to analyze. Phase and gain calibration in array signal processing Blind deconvolution (image deblurring; joint channel and signal estimation, etc.) Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 6 / 22

Self-calibration in array signal processing Calibration in the DOA (direction of arrival estimation) One calibration issue comes from the unknown gains of the antennae caused by temperature or humidity. Consider s signals impinging on an array of L antennae. Antenna elements θ " θ ' θ # θ $ θ % θ & y = s DA( θ k )x k + w k=1 where D is an unknown diagonal matrix and d ii is the unknown gain for i-th sensor. A(θ): array manifold. θ k : unknown direction of arrival. {x k } s k=1 are the impinging signals. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 7 / 22

Self-calibration in array signal processing Calibration in the DOA (direction of arrival estimation) One calibration issue comes from the unknown gains of the antennae caused by temperature or humidity. Consider s signals impinging on an array of L antennae. Antenna elements θ " θ ' θ # θ $ θ % θ & y = s DA( θ k )x k + w k=1 where D is an unknown diagonal matrix and d ii is the unknown gain for i-th sensor. A(θ): array manifold. θ k : unknown direction of arrival. {x k } s k=1 are the impinging signals. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 7 / 22

How is it related to compressive sensing? Discretize the manifold function A(θ) over [ π θ < π] on N grid points. y = DAx + w where A = A(θ 1 ) A(θ N ) C L N To achieve high resolution, we usually have L N. x C N 1 is s-sparse. Its s nonzero entries correspond to the directions of signals. Moreover, we don t know the locations of nonzero entries. Subspace constraint: assume D = diag(bh) where B is a known L K matrix and K < L. Number of constraints: L; number of unknowns: K + s. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 8 / 22

How is it related to compressive sensing? Discretize the manifold function A(θ) over [ π θ < π] on N grid points. y = DAx + w where A = A(θ 1 ) A(θ N ) C L N To achieve high resolution, we usually have L N. x C N 1 is s-sparse. Its s nonzero entries correspond to the directions of signals. Moreover, we don t know the locations of nonzero entries. Subspace constraint: assume D = diag(bh) where B is a known L K matrix and K < L. Number of constraints: L; number of unknowns: K + s. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 8 / 22

Self-calibration and biconvex compressive sensing Goal: Find (h, x) s.t. y = diag(bh)ax + w and x is sparse. Biconvex compressive sensing We are solving a biconvex (not convex) optimization problem to recover sparse signal x and calibrating parameter h. min h,x diag(bh)ax y 2 + λ x 1 A C L N and B C L K are known. h C K 1 and x C N 1 are unknown. x is sparse. Remark: If h is known, x can be recovered; if x is known, we can find h as well. Regarding identifiability issue, See [Lee, etc. 15]. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 9 / 22

Biconvex compressive sensing Goal: we want to find h and a sparse x from y, B and A. y: L 1 B: L K h: K 1 A: L N x: N 1, s-sparse w: L 1 = + Unknown parameters Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 10 / 22

Convex approach and lifting Two-step convex approach (a) Lifting: convert bilinear to linear constraints Widely used in phase retrieval [Candès, etc, 11], blind deconvolution [Ahmed, etc, 11], etc... (b) Solving a convex relaxation (semi-definite program) to recover h 0 x 0. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 11 / 22

Lifting: from bilinear to linear Step 1: lifting Let a i be the i-th column of A and b i be the i-th column of B. y i = (Bh 0 ) i x 0a i + w i = b i h 0 x 0a i + w i. Let X 0 := h 0 x 0 and define the linear operator A : CK N C L as, Then, there holds A(Z) := {b i Za i } L i=1 = { Z, b i a i } L i=1. y = A(X 0 ) + w. In this way, A (z) = L i=1 z ib i a i : C L C K N. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 12 / 22

Rank-1 matrix recovery Lifting: recovery of a rank - 1 and row-sparse matrix Find Z s.t. rank(z) = 1 A(Z) = A(X 0 ) Z has sparse rows X 0 0 = Ks where X 0 = h 0 x 0, h 0 C K and x 0 C N with x 0 0 = s. 0 0 h 1 x i1 0 0 h 1 x is 0 0 0 0 h 2 x i1 0 0 h 2 x is 0 0 Z =.............. 0 0 h K x i1 0 0 h K x is 0 0 An NP-hard problem to find such a rank-1 and sparse matrix. K N Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 13 / 22

Rank-1 matrix recovery Lifting: recovery of a rank - 1 and row-sparse matrix Find Z s.t. rank(z) = 1 A(Z) = A(X 0 ) Z has sparse rows X 0 0 = Ks where X 0 = h 0 x 0, h 0 C K and x 0 C N with x 0 0 = s. 0 0 h 1 x i1 0 0 h 1 x is 0 0 0 0 h 2 x i1 0 0 h 2 x is 0 0 Z =.............. 0 0 h K x i1 0 0 h K x is 0 0 An NP-hard problem to find such a rank-1 and sparse matrix. K N Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 13 / 22

SparseLift Z : nuclear norm and Z 1 : l 1 -norm of vectorized Z. A popular way: nuclear norm + l 1 - minimization min Z 1 + λ Z s.t. A(Z) = A(X 0 ), λ 0. However, combination of multiple norms may not do any better. [Oymak, etc. 12]. SparseLift min Z 1 s.t. A(Z) = A(X 0 ). Idea: Lift the recovery problem of two unknown vectors to a matrix-valued problem and exploit sparsity through l 1 -minimization. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 14 / 22

SparseLift Z : nuclear norm and Z 1 : l 1 -norm of vectorized Z. A popular way: nuclear norm + l 1 - minimization min Z 1 + λ Z s.t. A(Z) = A(X 0 ), λ 0. However, combination of multiple norms may not do any better. [Oymak, etc. 12]. SparseLift min Z 1 s.t. A(Z) = A(X 0 ). Idea: Lift the recovery problem of two unknown vectors to a matrix-valued problem and exploit sparsity through l 1 -minimization. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 14 / 22

Main theorem Theorem: [Ling-Strohmer, 2015] Recall the model: where y = DAx, D = diag(bh), (a) B is an L K DFT tall matrix with B B = I K (b) A is an L N real Gaussian random matrix or a random Fourier matrix. Then SparseLift recovers X 0 exactly with high probability if L = O( }{{} K dimension of h }{{} s log 2 L) level of sparsity where Ks = X 0 0. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 15 / 22

Comments It is shown that l 1-2 minimization also works (exploit group sparsity) [Flinth, 17]. min X fails if L < N. min X L = O(K + N) min X 1 L = O(Ks log KN) Solving l 1 -minimization is easier and cheaper than solving SDP. Compared with Compressive Sensing Compressive Sensing L = O(s log N) Our Case L = O(Ks log KN) Believed to be optimal if one uses the Lifting technique. It is unknown whether any algorithm would work for L = O(K + s). Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 16 / 22

Phase transition: SparseLift vs. 1 + λ min 1 + λ does not do any better than min 1. White: Success, Black: Failure Empirical prob of success: L = 128, N = 256 1 Empirical prob of success: L = 128, N = 256 1 14 0.9 14 0.9 12 0.8 12 0.8 0.7 0.7 10 0.6 10 0.6 k:1 to 15 8 6 0.5 0.4 k:1 to 15 8 6 0.5 0.4 0.3 0.3 4 0.2 4 0.2 2 0.1 2 0.1 2 4 6 8 10 12 14 s:1 to 15 (Gaussian Case: Performance of Sparselift) Figure: SparseLift 0 0 2 4 6 8 10 12 14 s:1 to 15 (Gaussian Case and min 1 + 0.1 solve Figure: min 1 + 0.1 L = 128, N = 256. A: Gaussian and B: Non-random partial Fourier matrix. 10 experiments for each pair (K, s), 1 K, s 15. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 17 / 22

Minimal L is nearly proportional to Ks L : 10 to 400; N = 512; A: Gaussian random matrices; B: first K columns of a DFT matrix. 400 Empirical prob of success: N = 512, k = 5 1 400 Empirical prob of success: N = 512, s = 5 1 350 0.9 350 0.9 0.8 0.8 300 0.7 300 0.7 L:10 to 400 250 200 150 0.6 0.5 0.4 L:10 to 400 250 200 150 0.6 0.5 0.4 0.3 0.3 100 0.2 100 0.2 50 0.1 50 0.1 2 4 6 8 10 12 14 n:1 to 15 0 2 4 6 8 10 12 14 k:1 to 15 0 Figure: Fix K = 5 Figure: Fix s = 5 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 18 / 22

Stability theory Assume that y is contaminated by noise, namely, y = A(X 0 ) + w with w η, we solve the following program to recover X 0, min Z 1 s.t. A(Z) y η. Theorem If A is either a Gaussian random matrix or a random Fourier matrix, ˆX X 0 F (C 0 + C 1 Ks)η with high probability. L satisfies the condition in the noiseless case. Both C 0 and C 1 are constants. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 19 / 22

Stability theory Assume that y is contaminated by noise, namely, y = A(X 0 ) + w with w η, we solve the following program to recover X 0, min Z 1 s.t. A(Z) y η. Theorem If A is either a Gaussian random matrix or a random Fourier matrix, ˆX X 0 F (C 0 + C 1 Ks)η with high probability. L satisfies the condition in the noiseless case. Both C 0 and C 1 are constants. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 19 / 22

Numerical example: relative error vs SNR 0 L = 128, N = 256, K = 5, s = 5 0 L = 128, N = 256, K = 5, s = 5 Average Relative Error of 10 Samples(dB) 10 20 30 40 50 60 70 80 Average Relative Error of 10 Samples(dB) 10 20 30 40 50 60 70 80 90 90 0 10 20 30 40 50 60 70 80 90 0 10 20 30 40 50 60 70 80 90 SNR(dB): Gaussian Case SNR(dB): Fourier Case Figure: A: Gaussian matrix Figure: A: random Fourier matrix Remarks: L = 128, N = 256, K = s = 5. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 20 / 22

DOA with a single snapshot Assume we have L = 64 sensors on one circle (circular array) and the gain d = Bh where B C 64 4 and h is complex Gaussian. The discretization of the angle consists of N = 180 grid points over [ 89 o, 90 o ] and SNR is 25dB. Self calibration for DOA Estimaion 1 true angles estimated angles 0.8 0.6 0.4 0.2 0 80 60 40 20 0 20 40 60 80 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 21 / 22

Outlook and Conclusion Conclusion: Is it possible to recover (h, x) with L = O(K + s) measurements? Consider multiple snapshots instead of one single snapshots. See details: 1 Self-calibration and biconvex compressive sensing. Inverse Problems 31 (11), 115002 2 Blind deconvolution meets blind demixing: algorithms and performance bounds, IEEE Transactions on Information Theory 63 (7), 4497-4520 3 Rapid, robust, and reliable blind deconvolution via nonconvex optimization, arxiv:1606.04933. 4 Regularized gradient descent: a nonconvex recipe for fast joint blind deconvolution and demixing arxiv:1703.08642. Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 22 / 22