Three Recent Examples of The Effectiveness of Convex Programming in the Information Sciences
|
|
- Shannon Farmer
- 5 years ago
- Views:
Transcription
1 Three Recent Examples of The Effectiveness of Convex Programming in the Information Sciences Emmanuel Candès International Symposium on Information Theory, Istanbul, July 2013
2 Three stories about a theme Often have missing information: (1) Missing phase (phase retrieval) (2) Missing and/or corrupted entries in data matrix (robust PCA) (3) Missing high-frequency spectrum (super-resolution) Makes signal/data recovery difficult This lecture Convex programming usually (but not always) returns the right answer!
3 Story # 1: Phase Retrieval Collaborators: Y. Eldar, X. Li, T. Strohmer, V. Voroninski
4 X-ray crystallography Method for determining atomic structure within a crystal principle typical setup 10 Nobel Prizes in X-ray crystallography, and counting...
5 Importance principle Franklin s photograph
6 Missing phase problem Detectors only record intensities of diffracted rays magnitude measurements only! Fraunhofer diffraction intensity of electrical field 2 ˆx(f 1, f 2 ) 2 = x(t 1, t 2 )e i2π(f1t1+f2t2) dt 1 dt 2
7 Missing phase problem Detectors only record intensities of diffracted rays magnitude measurements only! Fraunhofer diffraction intensity of electrical field 2 ˆx(f 1, f 2 ) 2 = x(t 1, t 2 )e i2π(f1t1+f2t2) dt 1 dt 2 Phase retrieval problem (inversion) How can we recover the phase (or equivalently signal x(t 1, t 2 )) from ˆx(f 1, f 2 )?
8 About the importance of phase...
9 About the importance of phase... DFT DFT
10 About the importance of phase... DFT keep magnitude swap phase DFT
11 About the importance of phase... DFT keep magnitude swap phase DFT
12 X-ray imaging: now and then Röntgen (1895) Dierolf (2010)
13 Ultrashort X-ray pulses Imaging single large protein complexes
14 Discrete mathematical model Phaseless measurements about x 0 C n b k = a k, x 0 2 k {1,..., m} = [m] Phase retrieval is feasibility problem find x subject to a k, x 2 = b k k [m] Solving quadratic equations is NP hard in general
15 Discrete mathematical model Phaseless measurements about x 0 C n b k = a k, x 0 2 k {1,..., m} = [m] Phase retrieval is feasibility problem find x subject to a k, x 2 = b k k [m] Solving quadratic equations is NP hard in general Nobel Prize for Hauptman and Karle ( 85): make use of very specific prior knowledge
16 Discrete mathematical model Phaseless measurements about x 0 C n b k = a k, x 0 2 k {1,..., m} = [m] Phase retrieval is feasibility problem find x subject to a k, x 2 = b k k [m] Solving quadratic equations is NP hard in general Nobel Prize for Hauptman and Karle ( 85): make use of very specific prior knowledge Standard approach: Gerchberg Saxton (or Fienup) iterative algorithm Sometimes works well Sometimes does not
17 PhaseLift a k, x 2 = b k k [m]
18 PhaseLift a k, x 2 = b k k [m] Lifting: X = xx a k, x 2 = Tr(x a k a kx) = Tr(a k a kxx ) := Tr(A k X) a k a k = A k Turns quadratic measurements into linear measurements about xx
19 PhaseLift a k, x 2 = b k k [m] Lifting: X = xx a k, x 2 = Tr(x a k a kx) = Tr(a k a kxx ) := Tr(A k X) a k a k = A k Turns quadratic measurements into linear measurements about xx Phase retrieval: equivalent formulation find X s. t. Tr(A k X) = b k k [m] X 0, rank(x) = 1 Combinatorially hard min rank(x) s. t. Tr(A k X) = b k k [m] X 0
20 PhaseLift a k, x 2 = b k k [m] Lifting: X = xx a k, x 2 = Tr(x a k a kx) = Tr(a k a kxx ) := Tr(A k X) a k a k = A k Turns quadratic measurements into linear measurements about xx PhaseLift: tractable semidefinite relaxation minimize Tr(X) subject to Tr(A k X) = b k k [m] X 0 This is a semidefinite program (SDP) Trace is convex proxy for rank
21 Semidefinite programming (SDP) Special class of convex optimization problems Relatively natural extension of linear programming (LP) Efficient numerical solvers (interior point methods) LP (std. form): x R n SDP (std. form): X R n n minimize c, x subject to a T k x = b k k = 1,... x 0 minimize C, X subject to A k, X = b k k = 1,... X 0 Standard inner product: C, X = Tr(C X)
22 From overdetermined to highly underdetermined Quadratic equations Lift b k = a k, x 2 k [m] b = A(xx ) minimize subject to Tr(X) A(X) = b X 0 Have we made things worse? overdetermined (m > n) highly underdetermined (m n 2 )
23 This is not really new... Relaxation of quadratically constrained QP s Shor (87) [Lower bounds on nonconvex quadratic optimization problems] Goemans and Williamson (95) [MAX-CUT] Ben-Tal and Nemirovskii (01) [Monograph]... Similar approach for array imaging: Chai, Moscoso, Papanicolaou (11)
24 Quadratic equations: geometric view I
25 Quadratic equations: geometric view I
26 Quadratic equations: geometric view I
27 Quadratic equations: geometric view I
28 Quadratic equations: geometric view I
29 Quadratic equations: geometric view I
30 Quadratic equations: geometric view II
31 Exact phase retrieval via SDP Quadratic equations b k = a k, x 2 k [m] b = A(xx ) Simplest model: a k independently and uniformly sampled on unit sphere of C n if x C n (complex-valued problem) of R n if x R n (real-valued problem)
32 Exact phase retrieval via SDP Quadratic equations b k = a k, x 2 k [m] b = A(xx ) Simplest model: a k independently and uniformly sampled on unit sphere of C n if x C n (complex-valued problem) of R n if x R n (real-valued problem) Theorem (C. and Li ( 12); C., Strohmer and Voroninski ( 11)) Assume m n. With prob. 1 O(e γm ), for all x C n, only point in feasible set {X : A(X) = b and X 0} is xx
33 Exact phase retrieval via SDP Quadratic equations b k = a k, x 2 k [m] b = A(xx ) Simplest model: a k independently and uniformly sampled on unit sphere of C n if x C n (complex-valued problem) of R n if x R n (real-valued problem) Theorem (C. and Li ( 12); C., Strohmer and Voroninski ( 11)) Assume m n. With prob. 1 O(e γm ), for all x C n, only point in feasible set {X : A(X) = b and X 0} is xx Injectivity if m 4n 2 (Balan, Bodmann, Casazza, Edidin 09)
34 Numerical results: noiseless recovery (a) Smooth signal (real part) (b) Random signal (real part) Figure: Recovery (with reweighting) of n-dimensional complex signal (2n unknowns) from 4n quadratic measurements (random binary masks)
35 How is this possible? How can feasible set {X 0} {A(X) = b} have a unique point? [ ] x y Intersection of 0 with affine space y z
36 Correct representation Rank-1 matrices are on the boundary (extreme rays) of PSD cone
37 My mental representation
38 My mental representation
39 My mental representation
40 My mental representation
41 My mental representation
42 With noise b k x, a k 2 k [m] Noise aware recovery (SDP) minimize A(X) b 1 = k Tr(a ka k X) b k subject to X 0 Signal ˆx obtained by extracting first eigenvector (PC) of solution matrix
43 With noise b k x, a k 2 k [m] Noise aware recovery (SDP) minimize A(X) b 1 = k Tr(a ka k X) b k subject to X 0 Signal ˆx obtained by extracting first eigenvector (PC) of solution matrix In same setup as before and for realistic noise models, no method whatsoever can possibly yield a fundamentally smaller recovery error [C. and Li (2012)]
44 Numerical results: noisy recovery Figure: SNR versus relative MSE on a db-scale for different numbers of illuminations with binary masks
45 Numerical results: noiseless 2D images original image Gaussian masks Courtesy S. Marchesini (LBL) binary masks error with 8 binary masks
46 Story #2: Robust Principal Component Analysis Collaborators: X. Li, Y. Ma, J. Wright
47 The separation problem (Chandrasekahran et al.) M = L + S M: data matrix (observed) L: low-rank (unobserved) S: sparse (unobserved)
48 The separation problem (Chandrasekahran et al.) M = L + S M: data matrix (observed) L: low-rank (unobserved) S: sparse (unobserved) Problem: can we recover L and S accurately? Again, missing information
49 Motivation: robust principal component analysis (RPCA) PCA sensitive to outliers: breaks down with one (badly) corrupted data point x 11 x x 1n x 21 x x 2n x d1 x d2... x dn
50 Motivation: robust principal component analysis (RPCA) PCA sensitive to outliers: breaks down with one (badly) corrupted data point x 11 x x 1n x 11 x x 1n x 21 x x 2n x 21 x x 2n.... = x d1 x d2... x dn x d1 A... x dn
51 Robust PCA Data increasingly high dimensional Gross errors frequently occur in many applications Image processing Web data analysis Bioinformatics... Occlusions Malicious tampering Sensor failures... Important to make PCA robust
52 Gross errors Observe corrupted entries Users Movies A A A Y ij = L ij + S ij (i, j) Ω obs L low-rank matrix S entries that have been tampered with (impulsive noise) Problem Recover L from missing and corrupted samples
53 The L+S model (Partial) information y = A(M) about M }{{} object = L }{{} low rank + S }{{} sparse A????? A?????????? A? A????? A?? RPCA Dynamic MR data = low-dimensional structure + corruption video seq. = static background + sparse innovation Graphical modeling with hidden variables: Chandrasekaran, Sanghavi, Parrilo, Willsky ( 09, 11) marginal inverse covariance of observed variables = low-rank + sparse
54 When does separation make sense? Low-rank component cannot be sparse: L =
55 When does separation make sense? Low-rank component cannot be sparse: L = Sparse component cannot be low rank: S =
56 Low-rank component cannot be sparse L = Incoherent condition [C. and Recht ( 08)]: column and row spaces not aligned with coordinate axes (singular vectors are not sparse)
57 Low-rank component cannot be sparse A A M = Incoherent condition [C. and Recht ( 08)]: column and row spaces not aligned with coordinate axes (singular vectors are not sparse)
58 Sparse component cannot be low-rank x 1 x 2 x n 1 x n x 2 x n 1 x n x 1 x 2 x n 1 x n A x 2 x n 1 x n L = L + S = x 1 x 2 x n 1 x n A x }{{} 2 x n 1 x n 1x Sparsity pattern will be assumed (uniform) random
59 Demixing by convex programming L unknown (rank unknown) M = L + S S unknown (# of entries 0, locations, magnitudes all unknown)
60 Demixing by convex programming L unknown (rank unknown) M = L + S S unknown (# of entries 0, locations, magnitudes all unknown) Recovery via SDP minimize ˆL + λ Ŝ 1 subject to ˆL + Ŝ = M See also Chandrasekaran, Sanghavi, Parrilo, Willsky ( 09) nuclear norm: L = i σ i(l) (sum of sing. values) l 1 norm: S 1 = ij S ij (sum of abs. values)
61 Exact recovery via SDP min ˆL + λ Ŝ 1 s. t. ˆL + Ŝ = M
62 Exact recovery via SDP min ˆL + λ Ŝ 1 s. t. ˆL + Ŝ = M Theorem L is n n of rank(l) ρ r n (log n) 2 and incoherent S is n n, random sparsity pattern of cardinality at most ρ s n 2 Then with probability 1 O(n 10 ), SDP with λ = 1/ n is exact: ˆL = L, Ŝ = S Same conclusion for rectangular matrices with λ = 1/ max dim
63 Exact recovery via SDP min ˆL + λ Ŝ 1 s. t. ˆL + Ŝ = M Theorem L is n n of rank(l) ρ r n (log n) 2 and incoherent S is n n, random sparsity pattern of cardinality at most ρ s n 2 Then with probability 1 O(n 10 ), SDP with λ = 1/ n is exact: ˆL = L, Ŝ = S Same conclusion for rectangular matrices with λ = 1/ max dim No tuning parameter! Whatever the magnitudes of L and S A A A A A A A A A A A A
64 Phase transitions in probability of success (a) PCP, Random Signs (b) PCP, Coherent Signs (c) Matrix Completion L = XY T is a product of independent n r i.i.d. N (0, 1/n) matrices
65 Missing and corrupted RPCA min ˆL + λ Ŝ 1 s. t. ˆLij + Ŝij = L ij + S ij (i, j) Ω obs A????? A?????????? A? A????? A??
66 Missing and corrupted RPCA min ˆL + λ Ŝ 1 s. t. ˆLij + Ŝij = L ij + S ij (i, j) Ω obs A????? A?????????? A? A????? A?? Theorem L as before Ω obs random set of size 0.1n 2 (missing frac. is arbitrary) Each observed entry corrupted with prob. τ τ 0 Then with prob. 1 O(n 10 ), PCP with λ = 1/ 0.1n is exact: ˆL = L Same conclusion for rectangular matrices with λ = 1/ 0.1max dim
67 Background subtraction
68 With noise With Li, Ma, Wright & Zhou ( 10) Z stochastic or deterministic perturbation Y ij = L ij + S ij + Z ij (i, j) Ω minimize ˆL + λ Ŝ 1 subject to (i,j) Ω (M ij ˆL ij Ŝij) 2 δ 2 When perfect (noiseless) separation occurs = noisy variant is stable
69 Story #3: Super-resolution Collaborator: C. Fernandez-Granda
70 Limits of resolution In any optical imaging system, diffraction imposes fundamental limit on resolution The physical phenomenon called diffraction is of the utmost importance in the theory of optical imaging systems (Joseph Goodman)
71 Bandlimited imaging systems (Fourier optics) f obs (t) = (h f)(t) h : point spread function (PSF) ˆf obs (ω) = ĥ(ω) ˆf(ω) ĥ : transfer function (TF) Bandlimited system ω > Ω ĥ(ω) = 0 ˆf obs (ω) = ĥ(ω) ˆf(ω) suppresses all high-frequency components
72 Bandlimited imaging systems (Fourier optics) f obs (t) = (h f)(t) h : point spread function (PSF) ˆf obs (ω) = ĥ(ω) ˆf(ω) ĥ : transfer function (TF) Bandlimited system ω > Ω ĥ(ω) = 0 ˆf obs (ω) = ĥ(ω) ˆf(ω) suppresses all high-frequency components Example: coherent imaging ĥ(ω) = 1 P (ω) indicator of pupil element TF PSF cross-section (PSF) Pupil Airy disk
73 Rayleigh resolution limit Lord Rayleigh
74 The super-resolution problem objective data Retrieve fine scale information from low-pass data Equivalent description: extrapolate spectrum (ill posed)
75 Random vs. low-frequency sampling Random sampling (CS) Low-frequency sampling (SR) Compressive sensing: spectrum interpolation Super-resolution: spectrum extrapolation
76 Super-resolving point sources This lecture Signal of interest is superposition of point sources Celestial bodies in astronomy Line spectra in speech analysis Fluorescent molecules in single-molecule microscopy Many applications Radar Spectroscopy Medical imaging Astronomy Geophysics...
77 Single molecule imaging Microscope receives light from fluorescent molecules Problem Resolution is much coarser than size of individual molecules (low-pass data) Can we beat the diffraction limit and super-resolve those molecules? Higher molecule density faster imaging
78 Mathematical model Signal x = j a jδ τj a j C, τ j T [0, 1] Data y = F n x: n = 2f lo + 1 low-frequency coefficients (Nyquist sampling) y(k) = 1 0 e i2πkt x(dt) = j a j e i2πkτj k Z, k f lo Resolution limit: (λ lo /2 is Rayleigh distance) 1/f lo = λ lo
79 Mathematical model Signal x = j a jδ τj a j C, τ j T [0, 1] Data y = F n x: n = 2f lo + 1 low-frequency coefficients (Nyquist sampling) y(k) = 1 0 e i2πkt x(dt) = j a j e i2πkτj k Z, k f lo Resolution limit: (λ lo /2 is Rayleigh distance) 1/f lo = λ lo Question Can we resolve the signal beyond this limit? Swap time and frequency spectral estimation
80 Can you find the spikes? Low-frequency data about spike train
81 Can you find the spikes? Low-frequency data about spike train
82 Recovery by minimum total-variation Recovery by cvx prog. min x TV subject to F n x = y x TV = x(dt) is continuous analog of l 1 norm x = j a j δ τj = x TV = j a j With noise min 1 2 y F n x 2 l 2 + λ x TV
83 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo
84 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices
85 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision!
86 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes
87 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes Can recover (2λ lo ) 1 = f lo /2 = n/4 spikes from n low-freq. samples
88 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes Can recover (2λ lo ) 1 = f lo /2 = n/4 spikes from n low-freq. samples Cannot go below λ lo
89 Recovery by convex programming y(k) = 1 0 e i2πkt x(dt) k f lo Theorem (C. and Fernandez Granda (2012)) If spikes are separated by at least 2 /f lo := 2 λ lo then min TV solution is exact! For real-valued x, a min dist. of 1.87λ lo suffices Infinite precision! Whatever the amplitudes Can recover (2λ lo ) 1 = f lo /2 = n/4 spikes from n low-freq. samples Cannot go below λ lo Essentially same result in higher dimensions
90 About separation: sparsity is not enough! CS: sparse signals are away from null space of sampling operator Super-res: this is not the case Signal Spectrum
91 About separation: sparsity is not enough! CS: sparse signals are away from null space of sampling operator Super-res: this is not the case Signal Spectrum
92 Analysis via prolate spheroidal functions David Slepian If distance between spikes less than λ lo /2 (Rayleigh), problem hopelessly ill posed
93 The super-resolution factor (SRF) Have data at resolution λ lo Super-resolution factor Wish resolution λ hi SRF = λ lo λ hi
94 The super-resolution factor (SRF): frequency viewpoint Observe spectrum up to f lo Wish to extrapolate up to f hi Super-resolution factor SRF = f hi f lo
95 With noise y = F n x + noise F n x = 1 0 e i2πkt x(dt) k f lo At native resolution (ˆx x) ϕ λlo TV noise level
96 With noise y = F n x + noise F n x = 1 0 e i2πkt x(dt) k f lo At finer resolution λ hi = λ lo /SRF, convex programming achieves (ˆx x) ϕ λhi TV SRF 2 noise level
97 With noise y = F n x + noise F n x = 1 0 e i2πkt x(dt) k f lo At finer resolution λ hi = λ lo /SRF, convex programming achieves (ˆx x) ϕ λhi TV SRF 2 noise level Modulus of continuity studies for super-resolution: Donoho ( 92)
98 Formulation as a finite-dimensional problem Primal problem min x TV s. t. F n x = y Dual problem max Re y, c s. t. F nc 1 Infinite-dimensional variable x Finitely many constraints Finite-dimensional variable c Infinitely many constraints (Fn c)(t) = c k e i2πkt k f lo
99 Formulation as a finite-dimensional problem Primal problem min x TV s. t. F n x = y Dual problem max Re y, c s. t. F nc 1 Infinite-dimensional variable x Finitely many constraints Finite-dimensional variable c Infinitely many constraints (Fn c)(t) = c k e i2πkt k f lo Semidefinite representability (F n c)(t) 1 for all t [0, 1] equivalent to (1) there is Q Hermitian s. t. [ ] Q c c 0 1 (2) Tr(Q) = 1 (3) sums along superdiagonals vanish: n j i=1 Q i,i+j = 0 for 1 j n 1
100 SDP formulation Dual as an SDP maximize Re y, c subject to [ ] Q c c 0 1 n j i=1 Q i,i+j = δ j 0 j n 1 Dual solution c: coeffs. of low-pass trig. polynomial k c ke i2πkt interpolating the sign of the primal solution
101 SDP formulation Dual as an SDP maximize Re y, c subject to [ ] Q c c 0 1 n j i=1 Q i,i+j = δ j 0 j n 1 Dual solution c: coeffs. of low-pass trig. polynomial k c ke i2πkt interpolating the sign of the primal solution To recover spike locations (1) Solve dual (2) Check when polynomial takes on magnitude 1
102 Noisy example SNR: 14 db Measurements
103 Noisy example SNR: 14 db Measurements Low res signal
104 Noisy example SNR: 14 db Measurements High res signal
105 Noisy example Average localization error: High res Signal Estimate
106 Summary Three important problems with missing data Phase retrieval Matrix completion/rpca Super-resolution Three simple and model-free recovery procedures via convex programming Three near-perfect solutions
107 Apologies: things I have not talked about Algorithms Applications Avalanche of related works
108 A small sample of papers I have greatly enjoyed Phase retrieval Netrapalli, Jain, Sanghavi, Phase retrieval using alternating minimization ( 13) Waldspurger, d Aspremont, Mallat, Phase recovery, MaxCut and complex semidefinite programming ( 12) Robust PCA Gross, Recovering low-rank matrices from few coefficients in any basis ( 09) Chandrasekaran, Parrilo and Willsky, Latent variable graphical model selection via convex optimization ( 11) Hsu, Kakade and Zhang, Robust matrix decomposition with outliers ( 11) Super-resolution Kahane, Analyse et synthèse harmoniques ( 11) Slepian, Prolate spheroidal wave functions, Fourier analysis, and uncertainty. V - The discrete case ( 78)
Super-resolution via Convex Programming
Super-resolution via Convex Programming Carlos Fernandez-Granda (Joint work with Emmanuel Candès) Structure and Randomness in System Identication and Learning, IPAM 1/17/2013 1/17/2013 1 / 44 Index 1 Motivation
More informationTowards a Mathematical Theory of Super-resolution
Towards a Mathematical Theory of Super-resolution Carlos Fernandez-Granda www.stanford.edu/~cfgranda/ Information Theory Forum, Information Systems Laboratory, Stanford 10/18/2013 Acknowledgements This
More informationFast and Robust Phase Retrieval
Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National
More informationSupport Detection in Super-resolution
Support Detection in Super-resolution Carlos Fernandez-Granda (Stanford University) 7/2/2013 SampTA 2013 Support Detection in Super-resolution C. Fernandez-Granda 1 / 25 Acknowledgements This work was
More informationAgenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization
Agenda Applications of semidefinite programming 1 Control and system theory 2 Combinatorial and nonconvex optimization 3 Spectral estimation & super-resolution Control and system theory SDP in wide use
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationPhase recovery with PhaseCut and the wavelet transform case
Phase recovery with PhaseCut and the wavelet transform case Irène Waldspurger Joint work with Alexandre d Aspremont and Stéphane Mallat Introduction 2 / 35 Goal : Solve the non-linear inverse problem Reconstruct
More informationThe Phase Problem: A Mathematical Tour from Norbert Wiener to Random Matrices and Convex Optimization
The Phase Problem: A Mathematical Tour from Norbert Wiener to Random Matrices and Convex Optimization Thomas Strohmer Department of Mathematics University of California, Davis Norbert Wiener Colloquium
More informationCombining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation
UIUC CSL Mar. 24 Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation Yuejie Chi Department of ECE and BMI Ohio State University Joint work with Yuxin Chen (Stanford).
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationStatistical Issues in Searches: Photon Science Response. Rebecca Willett, Duke University
Statistical Issues in Searches: Photon Science Response Rebecca Willett, Duke University 1 / 34 Photon science seen this week Applications tomography ptychography nanochrystalography coherent diffraction
More informationThe convex algebraic geometry of rank minimization
The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming
More informationSolving Corrupted Quadratic Equations, Provably
Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin
More informationROBUST BLIND SPIKES DECONVOLUTION. Yuejie Chi. Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 43210
ROBUST BLIND SPIKES DECONVOLUTION Yuejie Chi Department of ECE and Department of BMI The Ohio State University, Columbus, Ohio 4 ABSTRACT Blind spikes deconvolution, or blind super-resolution, deals with
More informationSparse Recovery Beyond Compressed Sensing
Sparse Recovery Beyond Compressed Sensing Carlos Fernandez-Granda www.cims.nyu.edu/~cfgranda Applied Math Colloquium, MIT 4/30/2018 Acknowledgements Project funded by NSF award DMS-1616340 Separable Nonlinear
More informationNonlinear Analysis with Frames. Part III: Algorithms
Nonlinear Analysis with Frames. Part III: Algorithms Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD July 28-30, 2015 Modern Harmonic Analysis and Applications
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationOptimization-based sparse recovery: Compressed sensing vs. super-resolution
Optimization-based sparse recovery: Compressed sensing vs. super-resolution Carlos Fernandez-Granda, Google Computational Photography and Intelligent Cameras, IPAM 2/5/2014 This work was supported by a
More informationProvable Alternating Minimization Methods for Non-convex Optimization
Provable Alternating Minimization Methods for Non-convex Optimization Prateek Jain Microsoft Research, India Joint work with Praneeth Netrapalli, Sujay Sanghavi, Alekh Agarwal, Animashree Anandkumar, Rashish
More informationLow-Rank Matrix Recovery
ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery
More informationConvex relaxation. In example below, we have N = 6, and the cut we are considering
Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution
More informationGoing off the grid. Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison
Going off the grid Benjamin Recht Department of Computer Sciences University of Wisconsin-Madison Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang We live in a continuous world... But we work
More informationCoherent imaging without phases
Coherent imaging without phases Miguel Moscoso Joint work with Alexei Novikov Chrysoula Tsogka and George Papanicolaou Waves and Imaging in Random Media, September 2017 Outline 1 The phase retrieval problem
More informationStructured matrix factorizations. Example: Eigenfaces
Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix
More informationFast Angular Synchronization for Phase Retrieval via Incomplete Information
Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department
More informationTractable Upper Bounds on the Restricted Isometry Constant
Tractable Upper Bounds on the Restricted Isometry Constant Alex d Aspremont, Francis Bach, Laurent El Ghaoui Princeton University, École Normale Supérieure, U.C. Berkeley. Support from NSF, DHS and Google.
More informationarxiv: v2 [cs.it] 20 Sep 2011
Phase Retrieval via Matrix Completion Emmanuel J. Candès, Yonina C. Eldar, Thomas Strohmer and Vladislav Voroninski August 2011 Abstract arxiv:1109.0573v2 [cs.it] 20 Sep 2011 This paper develops a novel
More informationGauge optimization and duality
1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel
More informationSparse Parameter Estimation: Compressed Sensing meets Matrix Pencil
Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil Yuejie Chi Departments of ECE and BMI The Ohio State University Colorado School of Mines December 9, 24 Page Acknowledgement Joint work
More informationSparse Solutions of an Undetermined Linear System
1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research
More informationSuper-Resolution of Point Sources via Convex Programming
Super-Resolution of Point Sources via Convex Programming Carlos Fernandez-Granda July 205; Revised December 205 Abstract We consider the problem of recovering a signal consisting of a superposition of
More informationNon-convex Robust PCA: Provable Bounds
Non-convex Robust PCA: Provable Bounds Anima Anandkumar U.C. Irvine Joint work with Praneeth Netrapalli, U.N. Niranjan, Prateek Jain and Sujay Sanghavi. Learning with Big Data High Dimensional Regime Missing
More informationRapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization
Rapid, Robust, and Reliable Blind Deconvolution via Nonconvex Optimization Shuyang Ling Department of Mathematics, UC Davis Oct.18th, 2016 Shuyang Ling (UC Davis) 16w5136, Oaxaca, Mexico Oct.18th, 2016
More informationCompressive Sensing and Beyond
Compressive Sensing and Beyond Sohail Bahmani Gerorgia Tech. Signal Processing Compressed Sensing Signal Models Classics: bandlimited The Sampling Theorem Any signal with bandwidth B can be recovered
More informationNonconvex Methods for Phase Retrieval
Nonconvex Methods for Phase Retrieval Vince Monardo Carnegie Mellon University March 28, 2018 Vince Monardo (CMU) 18.898 Midterm March 28, 2018 1 / 43 Paper Choice Main reference paper: Solving Random
More informationConstrained optimization
Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained
More informationPhaseLift: A novel framework for phase retrieval. Vladislav Voroninski
PhaseLift: A novel framework for phase retrieval. by Vladislav Voroninski A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Mathematics in the
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationConvex relaxation. In example below, we have N = 6, and the cut we are considering
Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution
More informationSparse and Low-Rank Matrix Decompositions
Forty-Seventh Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 30 - October 2, 2009 Sparse and Low-Rank Matrix Decompositions Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo,
More informationDemixing Sines and Spikes: Robust Spectral Super-resolution in the Presence of Outliers
Demixing Sines and Spikes: Robust Spectral Super-resolution in the Presence of Outliers Carlos Fernandez-Granda, Gongguo Tang, Xiaodong Wang and Le Zheng September 6 Abstract We consider the problem of
More informationarxiv: v1 [cs.it] 26 Oct 2015
Phase Retrieval: An Overview of Recent Developments Kishore Jaganathan Yonina C. Eldar Babak Hassibi Department of Electrical Engineering, Caltech Department of Electrical Engineering, Technion, Israel
More informationOverview. Optimization-Based Data Analysis. Carlos Fernandez-Granda
Overview Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 1/25/2016 Sparsity Denoising Regression Inverse problems Low-rank models Matrix completion
More informationA Convex Approach for Designing Good Linear Embeddings. Chinmay Hegde
A Convex Approach for Designing Good Linear Embeddings Chinmay Hegde Redundancy in Images Image Acquisition Sensor Information content
More informationRandom projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016
Lecture notes 5 February 9, 016 1 Introduction Random projections Random projections are a useful tool in the analysis and processing of high-dimensional data. We will analyze two applications that use
More informationLecture: Examples of LP, SOCP and SDP
1/34 Lecture: Examples of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:
More informationOptimisation Combinatoire et Convexe.
Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix
More informationSelf-Calibration and Biconvex Compressive Sensing
Self-Calibration and Biconvex Compressive Sensing Shuyang Ling Department of Mathematics, UC Davis July 12, 2017 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 1 / 22 Acknowledgements
More informationSuper-resolution of point sources via convex programming
Information and Inference: A Journal of the IMA 06) 5, 5 303 doi:0.093/imaiai/iaw005 Advance Access publication on 0 April 06 Super-resolution of point sources via convex programming Carlos Fernandez-Granda
More informationAnalysis of Robust PCA via Local Incoherence
Analysis of Robust PCA via Local Incoherence Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 3244 hzhan23@syr.edu Yi Zhou Department of EECS Syracuse University Syracuse, NY 3244 yzhou35@syr.edu
More informationLecture Notes 10: Matrix Factorization
Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To
More informationSum-of-Squares Method, Tensor Decomposition, Dictionary Learning
Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning David Steurer Cornell Approximation Algorithms and Hardness, Banff, August 2014 for many problems (e.g., all UG-hard ones): better guarantees
More informationSemidefinite Programming Basics and Applications
Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent
More informationSparse and Low Rank Recovery via Null Space Properties
Sparse and Low Rank Recovery via Null Space Properties Holger Rauhut Lehrstuhl C für Mathematik (Analysis), RWTH Aachen Convexity, probability and discrete structures, a geometric viewpoint Marne-la-Vallée,
More informationConvex sets, conic matrix factorizations and conic rank lower bounds
Convex sets, conic matrix factorizations and conic rank lower bounds Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute
More informationA Geometric Analysis of Phase Retrieval
A Geometric Analysis of Phase Retrieval Ju Sun, Qing Qu, John Wright js4038, qq2105, jw2966}@columbiaedu Dept of Electrical Engineering, Columbia University, New York NY 10027, USA Abstract Given measurements
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationReshaped Wirtinger Flow for Solving Quadratic System of Equations
Reshaped Wirtinger Flow for Solving Quadratic System of Equations Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 344 hzhan3@syr.edu Yingbin Liang Department of EECS Syracuse University
More informationSymmetric Factorization for Nonconvex Optimization
Symmetric Factorization for Nonconvex Optimization Qinqing Zheng February 24, 2017 1 Overview A growing body of recent research is shedding new light on the role of nonconvex optimization for tackling
More informationCompressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming
Compressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming Henrik Ohlsson, Allen Y. Yang Roy Dong S. Shankar Sastry Department of Electrical Engineering and Computer Sciences,
More informationRecovery of Simultaneously Structured Models using Convex Optimization
Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)
More informationInformation and Resolution
Information and Resolution (Invited Paper) Albert Fannjiang Department of Mathematics UC Davis, CA 95616-8633. fannjiang@math.ucdavis.edu. Abstract The issue of resolution with complex-field measurement
More informationConditions for Robust Principal Component Analysis
Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and
More informationDonald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016
Optimization for Tensor Models Donald Goldfarb IEOR Department Columbia University UCLA Mathematics Department Distinguished Lecture Series May 17 19, 2016 1 Tensors Matrix Tensor: higher-order matrix
More informationELE 538B: Sparsity, Structure and Inference. Super-Resolution. Yuxin Chen Princeton University, Spring 2017
ELE 538B: Sparsity, Structure and Inference Super-Resolution Yuxin Chen Princeton University, Spring 2017 Outline Classical methods for parameter estimation Polynomial method: Prony s method Subspace method:
More informationAcyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs
Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,
More informationThe Iterative and Regularized Least Squares Algorithm for Phase Retrieval
The Iterative and Regularized Least Squares Algorithm for Phase Retrieval Radu Balan Naveed Haghani Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 17, 2016
More informationUncertainty principles and sparse approximation
Uncertainty principles and sparse approximation In this lecture, we will consider the special case where the dictionary Ψ is composed of a pair of orthobases. We will see that our ability to find a sparse
More informationLecture Notes 9: Constrained Optimization
Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form
More informationDense Error Correction for Low-Rank Matrices via Principal Component Pursuit
Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit Arvind Ganesh, John Wright, Xiaodong Li, Emmanuel J. Candès, and Yi Ma, Microsoft Research Asia, Beijing, P.R.C Dept. of Electrical
More informationParameter Estimation for Mixture Models via Convex Optimization
Parameter Estimation for Mixture Models via Convex Optimization Yuanxin Li Department of Electrical and Computer Engineering The Ohio State University Columbus Ohio 432 Email: li.3822@osu.edu Yuejie Chi
More informationarxiv: v1 [math.oc] 11 Jun 2009
RANK-SPARSITY INCOHERENCE FOR MATRIX DECOMPOSITION VENKAT CHANDRASEKARAN, SUJAY SANGHAVI, PABLO A. PARRILO, S. WILLSKY AND ALAN arxiv:0906.2220v1 [math.oc] 11 Jun 2009 Abstract. Suppose we are given a
More informationSparse linear models
Sparse linear models Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 2/22/2016 Introduction Linear transforms Frequency representation Short-time
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationAN ALGORITHM FOR EXACT SUPER-RESOLUTION AND PHASE RETRIEVAL
2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) AN ALGORITHM FOR EXACT SUPER-RESOLUTION AND PHASE RETRIEVAL Yuxin Chen Yonina C. Eldar Andrea J. Goldsmith Department
More informationPhase Retrieval via Matrix Completion
SIAM J. IMAGING SCIENCES Vol. 6, No. 1, pp. 199 225 c 2013 Society for Industrial and Applied Mathematics Phase Retrieval via Matrix Completion Emmanuel J. Candès, Yonina C. Eldar, Thomas Strohmer, and
More informationConvex Optimization M2
Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial
More informationLinear dimensionality reduction for data analysis
Linear dimensionality reduction for data analysis Nicolas Gillis Joint work with Robert Luce, François Glineur, Stephen Vavasis, Robert Plemmons, Gabriella Casalino The setup Dimensionality reduction for
More informationPhase Retrieval with Random Illumination
Phase Retrieval with Random Illumination Albert Fannjiang, UC Davis Collaborator: Wenjing Liao NCTU, July 212 1 Phase retrieval Problem Reconstruct the object f from the Fourier magnitude Φf. Why do we
More informationMinimizing the Difference of L 1 and L 2 Norms with Applications
1/36 Minimizing the Difference of L 1 and L 2 Norms with Department of Mathematical Sciences University of Texas Dallas May 31, 2017 Partially supported by NSF DMS 1522786 2/36 Outline 1 A nonconvex approach:
More informationRank minimization via the γ 2 norm
Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises
More informationRecovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies
July 12, 212 Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies Morteza Mardani Dept. of ECE, University of Minnesota, Minneapolis, MN 55455 Acknowledgments:
More informationINEXACT ALTERNATING OPTIMIZATION FOR PHASE RETRIEVAL WITH OUTLIERS
INEXACT ALTERNATING OPTIMIZATION FOR PHASE RETRIEVAL WITH OUTLIERS Cheng Qian Xiao Fu Nicholas D. Sidiropoulos Lei Huang Dept. of Electronics and Information Engineering, Harbin Institute of Technology,
More informationFundamental Limits of Weak Recovery with Applications to Phase Retrieval
Proceedings of Machine Learning Research vol 75:1 6, 2018 31st Annual Conference on Learning Theory Fundamental Limits of Weak Recovery with Applications to Phase Retrieval Marco Mondelli Department of
More informationPhase Retrieval using the Iterative Regularized Least-Squares Algorithm
Phase Retrieval using the Iterative Regularized Least-Squares Algorithm Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD August 16, 2017 Workshop on Phaseless
More informationMIT Algebraic techniques and semidefinite optimization February 14, Lecture 3
MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications
More informationPhase Retrieval from Local Measurements: Deterministic Measurement Constructions and Efficient Recovery Algorithms
Phase Retrieval from Local Measurements: Deterministic Measurement Constructions and Efficient Recovery Algorithms Aditya Viswanathan Department of Mathematics and Statistics adityavv@umich.edu www-personal.umich.edu/
More informationLatent Variable Graphical Model Selection Via Convex Optimization
Latent Variable Graphical Model Selection Via Convex Optimization The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published
More informationRobust Asymmetric Nonnegative Matrix Factorization
Robust Asymmetric Nonnegative Matrix Factorization Hyenkyun Woo Korea Institue for Advanced Study Math @UCLA Feb. 11. 2015 Joint work with Haesun Park Georgia Institute of Technology, H. Woo (KIAS) RANMF
More informationGoing off the grid. Benjamin Recht University of California, Berkeley. Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang
Going off the grid Benjamin Recht University of California, Berkeley Joint work with Badri Bhaskar Parikshit Shah Gonnguo Tang imaging astronomy seismology spectroscopy x(t) = kx j= c j e i2 f jt DOA Estimation
More informationRobust PCA via Outlier Pursuit
Robust PCA via Outlier Pursuit Huan Xu Electrical and Computer Engineering University of Texas at Austin huan.xu@mail.utexas.edu Constantine Caramanis Electrical and Computer Engineering University of
More informationPHASELIFTOFF: AN ACCURATE AND STABLE PHASE RETRIEVAL METHOD BASED ON DIFFERENCE OF TRACE AND FROBENIUS NORMS
COMMUN. MATH. SCI. Vol. 3, No. Special-9, pp. xxx xxx c 205 International Press PHASELIFTOFF: AN ACCURATE AND STABLE PHASE RETRIEVAL METHOD BASED ON DIFFERENCE OF TRACE AND FROBENIUS NORMS PENGHANG YIN
More informationRobust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information
1 Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information Emmanuel Candès, California Institute of Technology International Conference on Computational Harmonic
More informationNear Optimal Signal Recovery from Random Projections
1 Near Optimal Signal Recovery from Random Projections Emmanuel Candès, California Institute of Technology Multiscale Geometric Analysis in High Dimensions: Workshop # 2 IPAM, UCLA, October 2004 Collaborators:
More informationCompressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles
Or: the equation Ax = b, revisited University of California, Los Angeles Mahler Lecture Series Acquiring signals Many types of real-world signals (e.g. sound, images, video) can be viewed as an n-dimensional
More informationCollaborative Filtering Matrix Completion Alternating Least Squares
Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016
More informationLarge-Scale L1-Related Minimization in Compressive Sensing and Beyond
Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March
More informationSuper-resolution by means of Beurling minimal extrapolation
Super-resolution by means of Beurling minimal extrapolation Norbert Wiener Center Department of Mathematics University of Maryland, College Park http://www.norbertwiener.umd.edu Acknowledgements ARO W911NF-16-1-0008
More informationLecture 20: November 1st
10-725: Optimization Fall 2012 Lecture 20: November 1st Lecturer: Geoff Gordon Scribes: Xiaolong Shen, Alex Beutel Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More information