Three Recent Examples of The Effectiveness of Convex Programming in the Information Sciences

Size: px
Start display at page:

Download "Three Recent Examples of The Effectiveness of Convex Programming in the Information Sciences"

Transcription

1 Three Recent Examples of The Effectiveness of Convex Programming in the Information Sciences Emmanuel Candès International Symposium on Information Theory, Istanbul, July 2013

2 Three stories about a theme Often have missing information: (1) Missing phase (phase retrieval) (2) Missing and/or corrupted entries in data matrix (robust PCA) (3) Missing high-frequency spectrum (super-resolution) Makes signal/data recovery difficult This lecture Convex programming usually (but not always) returns the right answer!

3 Story # 1: Phase Retrieval Collaborators: Y. Eldar, X. Li, T. Strohmer, V. Voroninski

4 X-ray crystallography Method for determining atomic structure within a crystal principle typical setup 10 Nobel Prizes in X-ray crystallography, and counting...

5 Importance principle Franklin s photograph

6 Missing phase problem Detectors only record intensities of diffracted rays magnitude measurements only! Fraunhofer diffraction intensity of electrical field 2 ˆx(f 1, f 2 ) 2 = x(t 1, t 2 )e i2π(f1t1+f2t2) dt 1 dt 2

7 Missing phase problem Detectors only record intensities of diffracted rays magnitude measurements only! Fraunhofer diffraction intensity of electrical field 2 ˆx(f 1, f 2 ) 2 = x(t 1, t 2 )e i2π(f1t1+f2t2) dt 1 dt 2 Phase retrieval problem (inversion) How can we recover the phase (or equivalently signal x(t 1, t 2 )) from ˆx(f 1, f 2 )?

8 About the importance of phase...

9 About the importance of phase... DFT DFT

10 About the importance of phase... DFT keep magnitude swap phase DFT

11 About the importance of phase... DFT keep magnitude swap phase DFT

12 X-ray imaging: now and then Röntgen (1895) Dierolf (2010)

13 Ultrashort X-ray pulses Imaging single large protein complexes

14 Discrete mathematical model Phaseless measurements about x 0 C n b k = a k, x 0 2 k {1,..., m} = [m] Phase retrieval is feasibility problem find x subject to a k, x 2 = b k k [m] Solving quadratic equations is NP hard in general

15 Discrete mathematical model Phaseless measurements about x 0 C n b k = a k, x 0 2 k {1,..., m} = [m] Phase retrieval is feasibility problem find x subject to a k, x 2 = b k k [m] Solving quadratic equations is NP hard in general Nobel Prize for Hauptman and Karle ( 85): make use of very specific prior knowledge

16 Discrete mathematical model Phaseless measurements about x 0 C n b k = a k, x 0 2 k {1,..., m} = [m] Phase retrieval is feasibility problem find x subject to a k, x 2 = b k k [m] Solving quadratic equations is NP hard in general Nobel Prize for Hauptman and Karle ( 85): make use of very specific prior knowledge Standard approach: Gerchberg Saxton (or Fienup) iterative algorithm Sometimes works well Sometimes does not

17 PhaseLift a k, x 2 = b k k [m]

18 PhaseLift a k, x 2 = b k k [m] Lifting: X = xx a k, x 2 = Tr(x a k a kx) = Tr(a k a kxx ) := Tr(A k X) a k a k = A k Turns quadratic measurements into linear measurements about xx

19 PhaseLift a k, x 2 = b k k [m] Lifting: X = xx a k, x 2 = Tr(x a k a kx) = Tr(a k a kxx ) := Tr(A k X) a k a k = A k Turns quadratic measurements into linear measurements about xx Phase retrieval: equivalent formulation find X s. t. Tr(A k X) = b k k [m] X 0, rank(x) = 1 Combinatorially hard min rank(x) s. t. Tr(A k X) = b k k [m] X 0

20 PhaseLift a k, x 2 = b k k [m] Lifting: X = xx a k, x 2 = Tr(x a k a kx) = Tr(a k a kxx ) := Tr(A k X) a k a k = A k Turns quadratic measurements into linear measurements about xx PhaseLift: tractable semidefinite relaxation minimize Tr(X) subject to Tr(A k X) = b k k [m] X 0 This is a semidefinite program (SDP) Trace is convex proxy for rank

21 Semidefinite programming (SDP) Special class of convex optimization problems Relatively natural extension of linear programming (LP) Efficient numerical solvers (interior point methods) LP (std. form): x R n SDP (std. form): X R n n minimize c, x subject to a T k x = b k k = 1,... x 0 minimize C, X subject to A k, X = b k k = 1,... X 0 Standard inner product: C, X = Tr(C X)

22 From overdetermined to highly underdetermined Quadratic equations Lift b k = a k, x 2 k [m] b = A(xx ) minimize subject to Tr(X) A(X) = b X 0 Have we made things worse? overdetermined (m > n) highly underdetermined (m n 2 )

23 This is not really new... Relaxation of quadratically constrained QP s Shor (87) [Lower bounds on nonconvex quadratic optimization problems] Goemans and Williamson (95) [MAX-CUT] Ben-Tal and Nemirovskii (01) [Monograph]... Similar approach for array imaging: Chai, Moscoso, Papanicolaou (11)

24 Quadratic equations: geometric view I

25 Quadratic equations: geometric view I

26 Quadratic equations: geometric view I

27 Quadratic equations: geometric view I

28 Quadratic equations: geometric view I

29 Quadratic equations: geometric view I

30 Quadratic equations: geometric view II

31 Exact phase retrieval via SDP Quadratic equations b k = a k, x 2 k [m] b = A(xx ) Simplest model: a k independently and uniformly sampled on unit sphere of C n if x C n (complex-valued problem) of R n if x R n (real-valued problem)

32 Exact phase retrieval via SDP Quadratic equations b k = a k, x 2 k [m] b = A(xx ) Simplest model: a k independently and uniformly sampled on unit sphere of C n if x C n (complex-valued problem) of R n if x R n (real-valued problem) Theorem (C. and Li ( 12); C., Strohmer and Voroninski ( 11)) Assume m n. With prob. 1 O(e γm ), for all x C n, only point in feasible set {X : A(X) = b and X 0} is xx

33 Exact phase retrieval via SDP Quadratic equations b k = a k, x 2 k [m] b = A(xx ) Simplest model: a k independently and uniformly sampled on unit sphere of C n if x C n (complex-valued problem) of R n if x R n (real-valued problem) Theorem (C. and Li ( 12); C., Strohmer and Voroninski ( 11)) Assume m n. With prob. 1 O(e γm ), for all x C n, only point in feasible set {X : A(X) = b and X 0} is xx Injectivity if m 4n 2 (Balan, Bodmann, Casazza, Edidin 09)

34 Numerical results: noiseless recovery (a) Smooth signal (real part) (b) Random signal (real part) Figure: Recovery (with reweighting) of n-dimensional complex signal (2n unknowns) from 4n quadratic measurements (random binary masks)

35 How is this possible? How can feasible set {X 0} {A(X) = b} have a unique point? [ ] x y Intersection of 0 with affine space y z

36 Correct representation Rank-1 matrices are on the boundary (extreme rays) of PSD cone

37 My mental representation

38 My mental representation

39 My mental representation

40 My mental representation

41 My mental representation

42 With noise b k x, a k 2 k [m] Noise aware recovery (SDP) minimize A(X) b 1 = k Tr(a ka k X) b k subject to X 0 Signal ˆx obtained by extracting first eigenvector (PC) of solution matrix

43 With noise b k x, a k 2 k [m] Noise aware recovery (SDP) minimize A(X) b 1 = k Tr(a ka k X) b k subject to X 0 Signal ˆx obtained by extracting first eigenvector (PC) of solution matrix In same setup as before and for realistic noise models, no method whatsoever can possibly yield a fundamentally smaller recovery error [C. and Li (2012)]

44 Numerical results: noisy recovery Figure: SNR versus relative MSE on a db-scale for different numbers of illuminations with binary masks

45 Numerical results: noiseless 2D images original image Gaussian masks Courtesy S. Marchesini (LBL) binary masks error with 8 binary masks

46 Story #2: Robust Principal Component Analysis Collaborators: X. Li, Y. Ma, J. Wright

47 The separation problem (Chandrasekahran et al.) M = L + S M: data matrix (observed) L: low-rank (unobserved) S: sparse (unobserved)

48 The separation problem (Chandrasekahran et al.) M = L + S M: data matrix (observed) L: low-rank (unobserved) S: sparse (unobserved) Problem: can we recover L and S accurately? Again, missing information

49 Motivation: robust principal component analysis (RPCA) PCA sensitive to outliers: breaks down with one (badly) corrupted data point x 11 x x 1n x 21 x x 2n x d1 x d2... x dn

50 Motivation: robust principal component analysis (RPCA) PCA sensitive to outliers: breaks down with one (badly) corrupted data point x 11 x x 1n x 11 x x 1n x 21 x x 2n x 21 x x 2n.... = x d1 x d2... x dn x d1 A... x dn

51 Robust PCA Data increasingly high dimensional Gross errors frequently occur in many applications Image processing Web data analysis Bioinformatics... Occlusions Malicious tampering Sensor failures... Important to make PCA robust

52 Gross errors Observe corrupted entries Users Movies A A A Y ij = L ij + S ij (i, j) Ω obs L low-rank matrix S entries that have been tampered with (impulsive noise) Problem Recover L from missing and corrupted samples

53 The L+S model (Partial) information y = A(M) about M }{{} object = L }{{} low rank + S }{{} sparse A????? A?????????? A? A????? A?? RPCA Dynamic MR data = low-dimensional structure + corruption video seq. = static background + sparse innovation Graphical modeling with hidden variables: Chandrasekaran, Sanghavi, Parrilo, Willsky ( 09, 11) marginal inverse covariance of observed variables = low-rank + sparse

54 When does separation make sense? Low-rank component cannot be sparse: L =

55 When does separation make sense? Low-rank component cannot be sparse: L = Sparse component cannot be low rank: S =

56 Low-rank component cannot be sparse L = Incoherent condition [C. and Recht ( 08)]: column and row spaces not aligned with coordinate axes (singular vectors are not sparse)

57 Low-rank component cannot be sparse A A M = Incoherent condition [C. and Recht ( 08)]: column and row spaces not aligned with coordinate axes (singular vectors are not sparse)

58 Sparse component cannot be low-rank x 1 x 2 x n 1 x n x 2 x n 1 x n x 1 x 2 x n 1 x n A x 2 x n 1 x n L = L + S = x 1 x 2 x n 1 x n A x }{{} 2 x n 1 x n 1x Sparsity pattern will be assumed (uniform) random

59 Demixing by convex programming L unknown (rank unknown) M = L + S S unknown (# of entries 0, locations, magnitudes all unknown)

60 Demixing by convex programming L unknown (rank unknown) M = L + S S unknown (# of entries 0, locations, magnitudes all unknown) Recovery via SDP minimize ˆL + λ Ŝ 1 subject to ˆL + Ŝ = M See also Chandrasekaran, Sanghavi, Parrilo, Willsky ( 09) nuclear norm: L = i σ i(l) (sum of sing. values) l 1 norm: S 1 = ij S ij (sum of abs. values)

61 Exact recovery via SDP min ˆL + λ Ŝ 1 s. t. ˆL + Ŝ = M

62 Exact recovery via SDP min ˆL + λ Ŝ 1 s. t. ˆL + Ŝ = M Theorem L is n n of rank(l) ρ r n (log n) 2 and incoherent S is n n, random sparsity pattern of cardinality at most ρ s n 2 Then with probability 1 O(n 10 ), SDP with λ = 1/ n is exact: ˆL = L, Ŝ = S Same conclusion for rectangular matrices with λ = 1/ max dim

63 Exact recovery via SDP min ˆL + λ Ŝ 1 s. t. ˆL + Ŝ = M Theorem L is n n of rank(l) ρ r n (log n) 2 and incoherent S is n n, random sparsity pattern of cardinality at most ρ s n 2 Then with probability 1 O(n 10 ), SDP with λ = 1/ n is exact: ˆL = L, Ŝ = S Same conclusion for rectangular matrices with λ = 1/ max dim No tuning parameter! Whatever the magnitudes of L and S A A A A A A A A A A A A

64 Phase transitions in probability of success (a) PCP, Random Signs (b) PCP, Coherent Signs (c) Matrix Completion L = XY T is a product of independent n r i.i.d. N (0, 1/n) matrices

65 Missing and corrupted RPCA min ˆL + λ Ŝ 1 s. t. ˆLij + Ŝij = L ij + S ij (i, j) Ω obs A????? A?????????? A? A????? A??

66 Missing and corrupted RPCA min ˆL + λ Ŝ 1 s. t. ˆLij + Ŝij = L ij + S ij (i, j) Ω obs A????? A?????????? A? A????? A?? Theorem L as before Ω obs random set of size 0.1n 2 (missing frac. is arbitrary) Each observed entry corrupted with prob. τ τ 0 Then with prob. 1 O(n 10 ), PCP with λ = 1/ 0.1n is exact: ˆL = L Same conclusion for rectangular matrices with λ = 1/ 0.1max dim

67 Background subtraction

68 With noise With Li, Ma, Wright & Zhou ( 10) Z stochastic or deterministic perturbation Y ij = L ij + S ij + Z ij (i, j) Ω minimize ˆL + λ Ŝ 1 subject to (i,j) Ω (M ij ˆL ij Ŝij) 2 δ 2 When perfect (noiseless) separation occurs = noisy variant is stable

69 Story #3: Super-resolution Collaborator: C. Fernandez-Granda

70 Limits of resolution In any optical imaging system, diffraction imposes fundamental limit on resolution The physical phenomenon called diffraction is of the utmost importance in the theory of optical imaging systems (Joseph Goodman)

71 Bandlimited imaging systems (Fourier optics) f obs (t) = (h f)(t) h : point spread function (PSF) ˆf obs (ω) = ĥ(ω) ˆf(ω) ĥ : transfer function (TF) Bandlimited system ω > Ω ĥ(ω) = 0 ˆf obs (ω) = ĥ(ω) ˆf(ω) suppresses all high-frequency components

72 Bandlimited imaging systems (Fourier optics) f obs (t) = (h f)(t) h : point spread function (PSF) ˆf obs (ω) = ĥ(ω) ˆf(ω) ĥ : transfer function (TF) Bandlimited system ω > Ω ĥ(ω) = 0 ˆf obs (ω) = ĥ(ω) ˆf(ω) suppresses all high-frequency components Example: coherent imaging ĥ(ω) = 1 P (ω) indicator of pupil element TF PSF cross-section (PSF) Pupil Airy disk

73 Rayleigh resolution limit Lord Rayleigh

74 The super-resolution problem objective data Retrieve fine scale information from low-pass data Equivalent description: extrapolate spectrum (ill posed)

75 Random vs. low-frequency sampling Random sampling (CS) Low-frequency sampling (SR) Compressive sensing: spectrum interpolation Super-resolution: spectrum extrapolation

76 Super-resolving point sources This lecture Signal of interest is superposition of point sources Celestial bodies in astronomy Line spectra in speech analysis Fluorescent molecules in single-molecule microscopy Many applications Radar Spectroscopy Medical imaging Astronomy Geophysics...

77 Single molecule imaging Microscope receives light from fluorescent molecules Problem Resolution is much coarser than size of individual molecules (low-pass data) Can we beat the diffraction limit and super-resolve those molecules? Higher molecule density faster imaging

78 Mathematical model Signal x = j a jδ τj a j C, τ j T [0, 1] Data y = F n x: n = 2f lo + 1 low-frequency coefficients (Nyquist sampling) y(k) = 1 0 e i2πkt x(dt) = j a j e i2πkτj k Z, k f lo Resolution limit: (λ lo /2 is Rayleigh distance) 1/f lo = λ lo

79 Mathematical model Signal x = j a jδ τj a j C, τ j T [0, 1] Data y = F n x: n = 2f lo + 1 low-frequency coefficients (Nyquist sampling) y(k) = 1 0 e i2πkt x(dt) = j a j e i2πkτj k Z, k f lo Resolution limit: (λ lo /2 is Rayleigh distance) 1/f lo = λ lo Question Can we resolve the signal beyond this limit? Swap time and frequency spectral estimation

80 Can you find the spikes? Low-frequency data about spike train

81 Can you find the spikes? Low-frequency data about spike train

82 Recovery by minimum total-variation Recovery by cvx prog. min x TV subject to F n x = y x TV = x(dt) is continuous analog of l 1 norm x = j a j δ τj = x TV = j a j With noise min 1 2 y F n x 2 l 2 + λ x TV

83 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo

84 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices

85 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision!

86 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes

87 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes Can recover (2λ lo ) 1 = f lo /2 = n/4 spikes from n low-freq. samples

88 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes Can recover (2λ lo ) 1 = f lo /2 = n/4 spikes from n low-freq. samples Cannot go below λ lo

89 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes Can recover (2λ lo ) 1 = f lo /2 = n/4 spikes from n low-freq. samples Cannot go below λ lo Essentially same result in higher dimensions

90 About separation: sparsity is not enough! CS: sparse signals are away from null space of sampling operator Super-res: this is not the case Signal Spectrum

91 About separation: sparsity is not enough! CS: sparse signals are away from null space of sampling operator Super-res: this is not the case Signal Spectrum

92 Analysis via prolate spheroidal functions David Slepian If distance between spikes less than λ lo /2 (Rayleigh), problem hopelessly ill posed

93 The super-resolution factor (SRF) Have data at resolution λ lo Super-resolution factor Wish resolution λ hi SRF = λ lo λ hi

94 The super-resolution factor (SRF): frequency viewpoint Observe spectrum up to f lo Wish to extrapolate up to f hi Super-resolution factor SRF = f hi f lo

95 With noise y = F n x + noise F n x = 1 0 e i2πkt x(dt) k f lo At native resolution (ˆx x) ϕ λlo TV noise level

96 With noise y = F n x + noise F n x = 1 0 e i2πkt x(dt) k f lo At finer resolution λ hi = λ lo /SRF, convex programming achieves (ˆx x) ϕ λhi TV SRF 2 noise level

97 With noise y = F n x + noise F n x = 1 0 e i2πkt x(dt) k f lo At finer resolution λ hi = λ lo /SRF, convex programming achieves (ˆx x) ϕ λhi TV SRF 2 noise level Modulus of continuity studies for super-resolution: Donoho ( 92)

98 Formulation as a finite-dimensional problem Primal problem min x TV s. t. F n x = y Dual problem max Re y, c s. t. F nc 1 Infinite-dimensional variable x Finitely many constraints Finite-dimensional variable c Infinitely many constraints (Fn c)(t) = c k e i2πkt k f lo

99 Formulation as a finite-dimensional problem Primal problem min x TV s. t. F n x = y Dual problem max Re y, c s. t. F nc 1 Infinite-dimensional variable x Finitely many constraints Finite-dimensional variable c Infinitely many constraints (Fn c)(t) = c k e i2πkt k f lo Semidefinite representability (F n c)(t) 1 for all t [0, 1] equivalent to (1) there is Q Hermitian s. t. [ ] Q c c 0 1 (2) Tr(Q) = 1 (3) sums along superdiagonals vanish: n j i=1 Q i,i+j = 0 for 1 j n 1

100 SDP formulation Dual as an SDP maximize Re y, c subject to [ ] Q c c 0 1 n j i=1 Q i,i+j = δ j 0 j n 1 Dual solution c: coeffs. of low-pass trig. polynomial k c ke i2πkt interpolating the sign of the primal solution

101 SDP formulation Dual as an SDP maximize Re y, c subject to [ ] Q c c 0 1 n j i=1 Q i,i+j = δ j 0 j n 1 Dual solution c: coeffs. of low-pass trig. polynomial k c ke i2πkt interpolating the sign of the primal solution To recover spike locations (1) Solve dual (2) Check when polynomial takes on magnitude 1

102 Noisy example SNR: 14 db Measurements

103 Noisy example SNR: 14 db Measurements Low res signal

104 Noisy example SNR: 14 db Measurements High res signal

105 Noisy example Average localization error: High res Signal Estimate

106 Summary Three important problems with missing data Phase retrieval Matrix completion/rpca Super-resolution Three simple and model-free recovery procedures via convex programming Three near-perfect solutions

107 Apologies: things I have not talked about Algorithms Applications Avalanche of related works

108 A small sample of papers I have greatly enjoyed Phase retrieval Netrapalli, Jain, Sanghavi, Phase retrieval using alternating minimization ( 13) Waldspurger, d Aspremont, Mallat, Phase recovery, MaxCut and complex semidefinite programming ( 12) Robust PCA Gross, Recovering low-rank matrices from few coefficients in any basis ( 09) Chandrasekaran, Parrilo and Willsky, Latent variable graphical model selection via convex optimization ( 11) Hsu, Kakade and Zhang, Robust matrix decomposition with outliers ( 11) Super-resolution Kahane, Analyse et synthèse harmoniques ( 11) Slepian, Prolate spheroidal wave functions, Fourier analysis, and uncertainty. V - The discrete case ( 78)

Super-resolution via Convex Programming

Super-resolution via Convex Programming Super-resolution via Convex Programming Carlos Fernandez-Granda (Joint work with Emmanuel Candès) Structure and Randomness in System Identication and Learning, IPAM 1/17/2013 1/17/2013 1 / 44 Index 1 Motivation

More information

Towards a Mathematical Theory of Super-resolution

Towards a Mathematical Theory of Super-resolution Towards a Mathematical Theory of Super-resolution Carlos Fernandez-Granda www.stanford.edu/~cfgranda/ Information Theory Forum, Information Systems Laboratory, Stanford 10/18/2013 Acknowledgements This

More information

Fast and Robust Phase Retrieval

Fast and Robust Phase Retrieval Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National

More information

Support Detection in Super-resolution

Support Detection in Super-resolution Support Detection in Super-resolution Carlos Fernandez-Granda (Stanford University) 7/2/2013 SampTA 2013 Support Detection in Super-resolution C. Fernandez-Granda 1 / 25 Acknowledgements This work was

More information

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization Agenda Applications of semidefinite programming 1 Control and system theory 2 Combinatorial and nonconvex optimization 3 Spectral estimation & super-resolution Control and system theory SDP in wide use

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

Phase recovery with PhaseCut and the wavelet transform case

Phase recovery with PhaseCut and the wavelet transform case Phase recovery with PhaseCut and the wavelet transform case Irène Waldspurger Joint work with Alexandre d Aspremont and Stéphane Mallat Introduction 2 / 35 Goal : Solve the non-linear inverse problem Reconstruct

More information

The Phase Problem: A Mathematical Tour from Norbert Wiener to Random Matrices and Convex Optimization

The Phase Problem: A Mathematical Tour from Norbert Wiener to Random Matrices and Convex Optimization The Phase Problem: A Mathematical Tour from Norbert Wiener to Random Matrices and Convex Optimization Thomas Strohmer Department of Mathematics University of California, Davis Norbert Wiener Colloquium

More information

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation UIUC CSL Mar. 24 Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation Yuejie Chi Department of ECE and BMI Ohio State University Joint work with Yuxin Chen (Stanford).

More information

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the

More information

Statistical Issues in Searches: Photon Science Response. Rebecca Willett, Duke University

Statistical Issues in Searches: Photon Science Response. Rebecca Willett, Duke University Statistical Issues in Searches: Photon Science Response Rebecca Willett, Duke University 1 / 34 Photon science seen this week Applications tomography ptychography nanochrystalography coherent diffraction

More information

The convex algebraic geometry of rank minimization

The convex algebraic geometry of rank minimization The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

ROBUST BLIND SPIKES DECONVOLUTION. Yuejie Chi. Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 43210

ROBUST BLIND SPIKES DECONVOLUTION. Yuejie Chi. Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 43210 ROBUST BLIND SPIKES DECONVOLUTION Yuejie Chi Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 4 ABSTRACT Blind spikes deconvolution, or blind super-resolution, deals with

More information

Sparse Recovery Beyond Compressed Sensing

Sparse Recovery Beyond Compressed Sensing Sparse Recovery Beyond Compressed Sensing Carlos Fernandez-Granda www.cims.nyu.edu/~cfgranda Applied Math Colloquium, MIT 4/30/2018 Acknowledgements Project funded by NSF award DMS-1616340 Separable Nonlinear

More information

Nonlinear Analysis with Frames. Part III: Algorithms

Nonlinear Analysis with Frames. Part III: Algorithms Nonlinear Analysis with Frames. Part III: Algorithms Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD July 28-30, 2015 Modern Harmonic Analysis and Applications

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Optimization-based sparse recovery: Compressed sensing vs. super-resolution

Optimization-based sparse recovery: Compressed sensing vs. super-resolution Optimization-based sparse recovery: Compressed sensing vs. super-resolution Carlos Fernandez-Granda, Google Computational Photography and Intelligent Cameras, IPAM 2/5/2014 This work was supported by a

More information

Provable Alternating Minimization Methods for Non-convex Optimization

Provable Alternating Minimization Methods for Non-convex Optimization Provable Alternating Minimization Methods for Non-convex Optimization Prateek Jain Microsoft Research, India Joint work with Praneeth Netrapalli, Sujay Sanghavi, Alekh Agarwal, Animashree Anandkumar, Rashish

More information

Low-Rank Matrix Recovery

Low-Rank Matrix Recovery ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery

More information

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Convex relaxation. In example below, we have N = 6, and the cut we are considering Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution

More information

Going off the grid. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison

Going off the grid. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Going off the grid Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang We live in a continuous world... But we work

More information

Coherent imaging without phases

Coherent imaging without phases Coherent imaging without phases Miguel Moscoso Joint work with Alexei Novikov Chrysoula Tsogka and George Papanicolaou Waves and Imaging in Random Media, September 2017 Outline 1 The phase retrieval problem

More information

Structured matrix factorizations. Example: Eigenfaces

Structured matrix factorizations. Example: Eigenfaces Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix

More information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department

More information

Tractable Upper Bounds on the Restricted Isometry Constant

Tractable Upper Bounds on the Restricted Isometry Constant Tractable Upper Bounds on the Restricted Isometry Constant Alex d Aspremont, Francis Bach, Laurent El Ghaoui Princeton University, École Normale Supérieure, U.C. Berkeley. Support from NSF, DHS and Google.

More information

arxiv: v2 [cs.it] 20 Sep 2011

arxiv: v2 [cs.it] 20 Sep 2011 Phase Retrieval via Matrix Completion Emmanuel J. Candès, Yonina C. Eldar, Thomas Strohmer and Vladislav Voroninski August 2011 Abstract arxiv:1109.0573v2 [cs.it] 20 Sep 2011 This paper develops a novel

More information

Gauge optimization and duality

Gauge optimization and duality 1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel

More information

Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil

Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil Yuejie Chi Departments of ECE and BMI The Ohio State University Colorado School of Mines December 9, 24 Page Acknowledgement Joint work

More information

Sparse Solutions of an Undetermined Linear System

Sparse Solutions of an Undetermined Linear System 1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research

More information

Super-Resolution of Point Sources via Convex Programming

Super-Resolution of Point Sources via Convex Programming Super-Resolution of Point Sources via Convex Programming Carlos Fernandez-Granda July 205; Revised December 205 Abstract We consider the problem of recovering a signal consisting of a superposition of

More information

Non-convex Robust PCA: Provable Bounds

Non-convex Robust PCA: Provable Bounds Non-convex Robust PCA: Provable Bounds Anima Anandkumar U.C. Irvine Joint work with Praneeth Netrapalli, U.N. Niranjan, Prateek Jain and Sujay Sanghavi. Learning with Big Data High Dimensional Regime Missing

More information

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization

Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization Shuyang Ling Department of Mathematics, UC Davis Oct.18th, 2016 Shuyang Ling (UC Davis) 16w5136, Oaxaca, Mexico Oct.18th, 2016

More information

Compressive Sensing and Beyond

Compressive Sensing and Beyond Compressive Sensing and Beyond Sohail Bahmani Gerorgia Tech. Signal Processing Compressed Sensing Signal Models Classics: bandlimited The Sampling Theorem Any signal with bandwidth B can be recovered

More information

Nonconvex Methods for Phase Retrieval

Nonconvex Methods for Phase Retrieval Nonconvex Methods for Phase Retrieval Vince Monardo Carnegie Mellon University March 28, 2018 Vince Monardo (CMU) 18.898 Midterm March 28, 2018 1 / 43 Paper Choice Main reference paper: Solving Random

More information

Constrained optimization

Constrained optimization Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained

More information

PhaseLift: A novel framework for phase retrieval. Vladislav Voroninski

PhaseLift: A novel framework for phase retrieval. Vladislav Voroninski PhaseLift: A novel framework for phase retrieval. by Vladislav Voroninski A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Mathematics in the

More information

Introduction to Compressed Sensing

Introduction to Compressed Sensing Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral

More information

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Convex relaxation. In example below, we have N = 6, and the cut we are considering Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution

More information

Sparse and Low-Rank Matrix Decompositions

Sparse and Low-Rank Matrix Decompositions Forty-Seventh Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 30 - October 2, 2009 Sparse and Low-Rank Matrix Decompositions Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo,

More information

Demixing Sines and Spikes: Robust Spectral Super-resolution in the Presence of Outliers

Demixing Sines and Spikes: Robust Spectral Super-resolution in the Presence of Outliers Demixing Sines and Spikes: Robust Spectral Super-resolution in the Presence of Outliers Carlos Fernandez-Granda, Gongguo Tang, Xiaodong Wang and Le Zheng September 6 Abstract We consider the problem of

More information

arxiv: v1 [cs.it] 26 Oct 2015

arxiv: v1 [cs.it] 26 Oct 2015 Phase Retrieval: An Overview of Recent Developments Kishore Jaganathan Yonina C. Eldar Babak Hassibi Department of Electrical Engineering, Caltech Department of Electrical Engineering, Technion, Israel

More information

Overview. Optimization-Based Data Analysis. Carlos Fernandez-Granda

Overview. Optimization-Based Data Analysis.   Carlos Fernandez-Granda Overview Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 1/25/2016 Sparsity Denoising Regression Inverse problems Low-rank models Matrix completion

More information

A Convex Approach for Designing Good Linear Embeddings. Chinmay Hegde

A Convex Approach for Designing Good Linear Embeddings. Chinmay Hegde A Convex Approach for Designing Good Linear Embeddings Chinmay Hegde Redundancy in Images Image Acquisition Sensor Information content

More information

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016 Lecture notes 5 February 9, 016 1 Introduction Random projections Random projections are a useful tool in the analysis and processing of high-dimensional data. We will analyze two applications that use

More information

Lecture: Examples of LP, SOCP and SDP

Lecture: Examples of LP, SOCP and SDP 1/34 Lecture: Examples of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:

More information

Optimisation Combinatoire et Convexe.

Optimisation Combinatoire et Convexe. Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix

More information

Self-Calibration and Biconvex Compressive Sensing

Self-Calibration and Biconvex Compressive Sensing Self-Calibration and Biconvex Compressive Sensing Shuyang Ling Department of Mathematics, UC Davis July 12, 2017 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 1 / 22 Acknowledgements

More information

Super-resolution of point sources via convex programming

Super-resolution of point sources via convex programming Information and Inference: A Journal of the IMA 06) 5, 5 303 doi:0.093/imaiai/iaw005 Advance Access publication on 0 April 06 Super-resolution of point sources via convex programming Carlos Fernandez-Granda

More information

Analysis of Robust PCA via Local Incoherence

Analysis of Robust PCA via Local Incoherence Analysis of Robust PCA via Local Incoherence Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 3244 hzhan23@syr.edu Yi Zhou Department of EECS Syracuse University Syracuse, NY 3244 yzhou35@syr.edu

More information

Lecture Notes 10: Matrix Factorization

Lecture Notes 10: Matrix Factorization Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To

More information

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning David Steurer Cornell Approximation Algorithms and Hardness, Banff, August 2014 for many problems (e.g., all UG-hard ones): better guarantees

More information

Semidefinite Programming Basics and Applications

Semidefinite Programming Basics and Applications Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent

More information

Sparse and Low Rank Recovery via Null Space Properties

Sparse and Low Rank Recovery via Null Space Properties Sparse and Low Rank Recovery via Null Space Properties Holger Rauhut Lehrstuhl C für Mathematik (Analysis), RWTH Aachen Convexity, probability and discrete structures, a geometric viewpoint Marne-la-Vallée,

More information

Convex sets, conic matrix factorizations and conic rank lower bounds

Convex sets, conic matrix factorizations and conic rank lower bounds Convex sets, conic matrix factorizations and conic rank lower bounds Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute

More information

A Geometric Analysis of Phase Retrieval

A Geometric Analysis of Phase Retrieval A Geometric Analysis of Phase Retrieval Ju Sun, Qing Qu, John Wright js4038, qq2105, jw2966}@columbiaedu Dept of Electrical Engineering, Columbia University, New York NY 10027, USA Abstract Given measurements

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

Reshaped Wirtinger Flow for Solving Quadratic System of Equations

Reshaped Wirtinger Flow for Solving Quadratic System of Equations Reshaped Wirtinger Flow for Solving Quadratic System of Equations Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 344 hzhan3@syr.edu Yingbin Liang Department of EECS Syracuse University

More information

Symmetric Factorization for Nonconvex Optimization

Symmetric Factorization for Nonconvex Optimization Symmetric Factorization for Nonconvex Optimization Qinqing Zheng February 24, 2017 1 Overview A growing body of recent research is shedding new light on the role of nonconvex optimization for tackling

More information

Compressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming

Compressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming Compressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming Henrik Ohlsson, Allen Y. Yang Roy Dong S. Shankar Sastry Department of Electrical Engineering and Computer Sciences,

More information

Recovery of Simultaneously Structured Models using Convex Optimization

Recovery of Simultaneously Structured Models using Convex Optimization Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)

More information

Information and Resolution

Information and Resolution Information and Resolution (Invited Paper) Albert Fannjiang Department of Mathematics UC Davis, CA 95616-8633. fannjiang@math.ucdavis.edu. Abstract The issue of resolution with complex-field measurement

More information

Conditions for Robust Principal Component Analysis

Conditions for Robust Principal Component Analysis Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and

More information

Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016

Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016 Optimization for Tensor Models Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016 1 Tensors Matrix Tensor: higher-order matrix

More information

ELE 538B: Sparsity, Structure and Inference. Super-Resolution. Yuxin Chen Princeton University, Spring 2017

ELE 538B: Sparsity, Structure and Inference. Super-Resolution. Yuxin Chen Princeton University, Spring 2017 ELE 538B: Sparsity, Structure and Inference Super-Resolution Yuxin Chen Princeton University, Spring 2017 Outline Classical methods for parameter estimation Polynomial method: Prony s method Subspace method:

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,

More information

The Iterative and Regularized Least Squares Algorithm for Phase Retrieval

The Iterative and Regularized Least Squares Algorithm for Phase Retrieval The Iterative and Regularized Least Squares Algorithm for Phase Retrieval Radu Balan Naveed Haghani Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 17, 2016

More information

Uncertainty principles and sparse approximation

Uncertainty principles and sparse approximation Uncertainty principles and sparse approximation In this lecture, we will consider the special case where the dictionary Ψ is composed of a pair of orthobases. We will see that our ability to find a sparse

More information

Lecture Notes 9: Constrained Optimization

Lecture Notes 9: Constrained Optimization Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form

More information

Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit

Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit Arvind Ganesh, John Wright, Xiaodong Li, Emmanuel J. Candès, and Yi Ma, Microsoft Research Asia, Beijing, P.R.C Dept. of Electrical

More information

Parameter Estimation for Mixture Models via Convex Optimization

Parameter Estimation for Mixture Models via Convex Optimization Parameter Estimation for Mixture Models via Convex Optimization Yuanxin Li Department of Electrical and Computer Engineering The Ohio State University Columbus Ohio 432 Email: li.3822@osu.edu Yuejie Chi

More information

arxiv: v1 [math.oc] 11 Jun 2009

arxiv: v1 [math.oc] 11 Jun 2009 RANK-SPARSITY INCOHERENCE FOR MATRIX DECOMPOSITION VENKAT CHANDRASEKARAN, SUJAY SANGHAVI, PABLO A. PARRILO, S. WILLSKY AND ALAN arxiv:0906.2220v1 [math.oc] 11 Jun 2009 Abstract. Suppose we are given a

More information

Sparse linear models

Sparse linear models Sparse linear models Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 2/22/2016 Introduction Linear transforms Frequency representation Short-time

More information

Low Rank Matrix Completion Formulation and Algorithm

Low Rank Matrix Completion Formulation and Algorithm 1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9

More information

AN ALGORITHM FOR EXACT SUPER-RESOLUTION AND PHASE RETRIEVAL

AN ALGORITHM FOR EXACT SUPER-RESOLUTION AND PHASE RETRIEVAL 2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) AN ALGORITHM FOR EXACT SUPER-RESOLUTION AND PHASE RETRIEVAL Yuxin Chen Yonina C. Eldar Andrea J. Goldsmith Department

More information

Phase Retrieval via Matrix Completion

Phase Retrieval via Matrix Completion SIAM J. IMAGING SCIENCES Vol. 6, No. 1, pp. 199 225 c 2013 Society for Industrial and Applied Mathematics Phase Retrieval via Matrix Completion Emmanuel J. Candès, Yonina C. Eldar, Thomas Strohmer, and

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial

More information

Linear dimensionality reduction for data analysis

Linear dimensionality reduction for data analysis Linear dimensionality reduction for data analysis Nicolas Gillis Joint work with Robert Luce, François Glineur, Stephen Vavasis, Robert Plemmons, Gabriella Casalino The setup Dimensionality reduction for

More information

Phase Retrieval with Random Illumination

Phase Retrieval with Random Illumination Phase Retrieval with Random Illumination Albert Fannjiang, UC Davis Collaborator: Wenjing Liao NCTU, July 212 1 Phase retrieval Problem Reconstruct the object f from the Fourier magnitude Φf. Why do we

More information

Minimizing the Difference of L 1 and L 2 Norms with Applications

Minimizing the Difference of L 1 and L 2 Norms with Applications 1/36 Minimizing the Difference of L 1 and L 2 Norms with Department of Mathematical Sciences University of Texas Dallas May 31, 2017 Partially supported by NSF DMS 1522786 2/36 Outline 1 A nonconvex approach:

More information

Rank minimization via the γ 2 norm

Rank minimization via the γ 2 norm Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises

More information

Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies

Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies July 12, 212 Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies Morteza Mardani Dept. of ECE, University of Minnesota, Minneapolis, MN 55455 Acknowledgments:

More information

INEXACT ALTERNATING OPTIMIZATION FOR PHASE RETRIEVAL WITH OUTLIERS

INEXACT ALTERNATING OPTIMIZATION FOR PHASE RETRIEVAL WITH OUTLIERS INEXACT ALTERNATING OPTIMIZATION FOR PHASE RETRIEVAL WITH OUTLIERS Cheng Qian Xiao Fu Nicholas D. Sidiropoulos Lei Huang Dept. of Electronics and Information Engineering, Harbin Institute of Technology,

More information

Fundamental Limits of Weak Recovery with Applications to Phase Retrieval

Fundamental Limits of Weak Recovery with Applications to Phase Retrieval Proceedings of Machine Learning Research vol 75:1 6, 2018 31st Annual Conference on Learning Theory Fundamental Limits of Weak Recovery with Applications to Phase Retrieval Marco Mondelli Department of

More information

Phase Retrieval using the Iterative Regularized Least-Squares Algorithm

Phase Retrieval using the Iterative Regularized Least-Squares Algorithm Phase Retrieval using the Iterative Regularized Least-Squares Algorithm Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD August 16, 2017 Workshop on Phaseless

More information

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3 MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications

More information

Phase Retrieval from Local Measurements: Deterministic Measurement Constructions and Efficient Recovery Algorithms

Phase Retrieval from Local Measurements: Deterministic Measurement Constructions and Efficient Recovery Algorithms Phase Retrieval from Local Measurements: Deterministic Measurement Constructions and Efficient Recovery Algorithms Aditya Viswanathan Department of Mathematics and Statistics adityavv@umich.edu www-personal.umich.edu/

More information

Latent Variable Graphical Model Selection Via Convex Optimization

Latent Variable Graphical Model Selection Via Convex Optimization Latent Variable Graphical Model Selection Via Convex Optimization The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published

More information

Robust Asymmetric Nonnegative Matrix Factorization

Robust Asymmetric Nonnegative Matrix Factorization Robust Asymmetric Nonnegative Matrix Factorization Hyenkyun Woo Korea Institue for Advanced Study Math @UCLA Feb. 11. 2015 Joint work with Haesun Park Georgia Institute of Technology, H. Woo (KIAS) RANMF

More information

Going off the grid. Benjamin Recht University of California, Berkeley. Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang

Going off the grid. Benjamin Recht University of California, Berkeley. Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang Going off the grid Benjamin Recht University of California, Berkeley Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang imaging astronomy seismology spectroscopy x(t) = kx j= c j e i2 f jt DOA Estimation

More information

Robust PCA via Outlier Pursuit

Robust PCA via Outlier Pursuit Robust PCA via Outlier Pursuit Huan Xu Electrical and Computer Engineering University of Texas at Austin huan.xu@mail.utexas.edu Constantine Caramanis Electrical and Computer Engineering University of

More information

PHASELIFTOFF: AN ACCURATE AND STABLE PHASE RETRIEVAL METHOD BASED ON DIFFERENCE OF TRACE AND FROBENIUS NORMS

PHASELIFTOFF: AN ACCURATE AND STABLE PHASE RETRIEVAL METHOD BASED ON DIFFERENCE OF TRACE AND FROBENIUS NORMS COMMUN. MATH. SCI. Vol. 3, No. Special-9, pp. xxx xxx c 205 International Press PHASELIFTOFF: AN ACCURATE AND STABLE PHASE RETRIEVAL METHOD BASED ON DIFFERENCE OF TRACE AND FROBENIUS NORMS PENGHANG YIN

More information

Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information

Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information 1 Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information Emmanuel Candès, California Institute of Technology International Conference on Computational Harmonic

More information

Near Optimal Signal Recovery from Random Projections

Near Optimal Signal Recovery from Random Projections 1 Near Optimal Signal Recovery from Random Projections Emmanuel Candès, California Institute of Technology Multiscale Geometric Analysis in High Dimensions: Workshop # 2 IPAM, UCLA, October 2004 Collaborators:

More information

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles Or: the equation Ax = b, revisited University of California, Los Angeles Mahler Lecture Series Acquiring signals Many types of real-world signals (e.g. sound, images, video) can be viewed as an n-dimensional

More information

Collaborative Filtering Matrix Completion Alternating Least Squares

Collaborative Filtering Matrix Completion Alternating Least Squares Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016

More information

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March

More information

Super-resolution by means of Beurling minimal extrapolation

Super-resolution by means of Beurling minimal extrapolation Super-resolution by means of Beurling minimal extrapolation Norbert Wiener Center Department of Mathematics University of Maryland, College Park http://www.norbertwiener.umd.edu Acknowledgements ARO W911NF-16-1-0008

More information

Lecture 20: November 1st

Lecture 20: November 1st 10-725: Optimization Fall 2012 Lecture 20: November 1st Lecturer: Geoff Gordon Scribes: Xiaolong Shen, Alex Beutel Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information