Compressed Sensing: a Subgradient Descent Method for Missing Data Problems

Similar documents
Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Sensing systems limited by constraints: physical size, time, cost, energy

Sparse Solutions of Systems of Equations and Sparse Modelling of Signals and Images

Chapter 2. Vectors and Vector Spaces

Douglas-Rachford splitting for nonconvex feasibility problems

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION

Uniqueness Conditions for A Class of l 0 -Minimization Problems

Lecture Note 5: Semidefinite Programming for Stability Analysis

Math 273a: Optimization Convex Conjugacy

Subgradient Method. Ryan Tibshirani Convex Optimization

Optimization and Optimal Control in Banach Spaces

Algorithms for Nonsmooth Optimization

STAT 200C: High-dimensional Statistics

Primal/Dual Decomposition Methods

Constrained optimization

Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France

Gauge optimization and duality

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Convex Optimization Notes

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases

Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images!

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Nonsmooth optimization: conditioning, convergence, and semi-algebraic models

A user s guide to Lojasiewicz/KL inequalities

Proximal methods. S. Villa. October 7, 2014

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction

Dual and primal-dual methods

Subdifferential representation of convex functions: refinements and applications

Convex Optimization and l 1 -minimization

Robust Asymmetric Nonnegative Matrix Factorization

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Stability and Robustness of Weak Orthogonal Matching Pursuits

15-780: LinearProgramming

7 Duality and Convex Programming

A Proximal Method for Identifying Active Manifolds

Conditional Gradient (Frank-Wolfe) Method

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina

The speed of Shor s R-algorithm

arxiv: v1 [math.oc] 15 Apr 2016

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Sparsity in Underdetermined Systems

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

M. Marques Alves Marina Geremia. November 30, 2017

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Final Presentation

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble

Chapter 1. Preliminaries

Uncertainty principles and sparse approximation

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods

On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean

Maximal Monotone Inclusions and Fitzpatrick Functions

Recent Developments in Compressed Sensing

Victoria Martín-Márquez

Lecture 7: Semidefinite programming

Optimization Algorithms for Compressed Sensing

Sparsity Matters. Robert J. Vanderbei September 20. IDA: Center for Communications Research Princeton NJ.

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

FINDING BEST APPROXIMATION PAIRS RELATIVE TO A CONVEX AND A PROX-REGULAR SET IN A HILBERT SPACE

The Frank-Wolfe Algorithm:

Sparse Optimization Lecture: Basic Sparse Optimization Models

The dual simplex method with bounds

An Introduction to Sparse Approximation

Introduction to Compressed Sensing

Lecture Notes 9: Constrained Optimization

Monotone operators and bigger conjugate functions

SCRIBERS: SOROOSH SHAFIEEZADEH-ABADEH, MICHAËL DEFFERRARD

Lecture 5 : Projections

NUMERICAL algorithms for nonconvex optimization models are often eschewed because the usual optimality

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

12. Interior-point methods

Dual Decomposition.

Provable Alternating Minimization Methods for Non-convex Optimization

BASICS OF CONVEX ANALYSIS

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

Key words. Complementarity set, Lyapunov rank, Bishop-Phelps cone, Irreducible cone

Sparse Optimization Lecture: Dual Methods, Part I

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Subgradient Method. Guest Lecturer: Fatma Kilinc-Karzan. Instructors: Pradeep Ravikumar, Aarti Singh Convex Optimization /36-725

Enhanced Compressive Sensing and More

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices

SPARSE signal representations have gained popularity in recent

SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS

Generalized Orthogonal Matching Pursuit- A Review and Some

Optimization for Compressed Sensing

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Machine Learning for Signal Processing Sparse and Overcomplete Representations

A Unified Approach to Proximal Algorithms using Bregman Distance

Constrained Optimization and Lagrangian Duality

Auxiliary-Function Methods in Optimization

Introduction to Real Analysis Alternative Chapter 1

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Reconstruction from Anisotropic Random Measurements

Transcription:

Compressed Sensing: a Subgradient Descent Method for Missing Data Problems ANZIAM, Jan 30 Feb 3, 2011 Jonathan M. Borwein Jointly with D. Russell Luke, University of Goettingen FRSC FAAAS FBAS FAA Director, CARMA, the University of Newcastle Revised: 02/02/2011

A Central Problem: l 0 minimization Given a linear map A : R n R m full-rank with 0 < m < n, solve Program (P 0 ) minimize x R n x 0 subject to Ax = b where x 0 := j sign(x j) with sign (0) := 0. x 0 = lim p 0 + j x j p is a metric but not a norm. p-balls for 1/5, 1/2, 1, 100 Combinatorial optimization problem (hard to solve).

Central Problem: l 0 minimization Solve instead Program (P 1 ) minimize x R n x 1 subject to Ax = b where x 1 is the usual l 1 norm. l 1 minimization now routine in statistics and elsewhere for missing data under-determined problems. A nonsmooth convex, actually linear, programming problem... easy to solve for small problems. Let s illustrate by trying to solve the problem for x a 512 512 image...

Application: Crystallography Given data: (autocorrelation transfer function ATF missing pixels) Desired reconstruction: (The true ATF with all pixels)

Application: Crystallography Formulate as: Solve Program minimize x R n x 1 subject to x C where C := {x R n Ax = b} for a linear A : R n R m (m < n). Could apply Douglas-Rachford iteration originated in 1956 for convex heat transfer problems (Laplacian): x n+1 := 1 ( Rf1 R f2 + I ) (x n ) 2 where R fj x := 2 prox α,fj x x for f 1 (x) := x 1 and f 2 (x) := ι C (x), and α > 0 fixed (a generalized best approximation or prox-mapping).

Application: Crystallography Great strategy for big problems, but convergence is (arbitrarily) slow and accuracy is likely to be poor... (Douglas-Rachford and Russell Luke) It seemed to me that a better approach was to think about real dynamics and see where they go. Maybe they go to the [classical] equilibrium solution and maybe they don t. Peter Diamond (2010 Economics co-nobel)

Motivation A variational/geometrical interpretation of the Candes-Tao (2004) probabilistic criterion for the solution to (P 1 ) to be unique and exactly match the true signal x. As a by-product, better practical methods for solving the underlying problem. Aim to use entropy/penalty ideas and duality and also prove some rigorous theorems. The counterpart paper (largely successful): J. M. Borwein and D. R. Luke, Entropic Regularization of the l 0 function. In Fixed-Point Algorithms for Inverse Problems in Science and Engineering, Springer Optimization and Its Applications. In press, 2011. Available at http://carma.newcastle.edu.au/jon/sensing.pdf.

Motivation A variational/geometrical interpretation of the Candes-Tao (2004) probabilistic criterion for the solution to (P 1 ) to be unique and exactly match the true signal x. As a by-product, better practical methods for solving the underlying problem. Aim to use entropy/penalty ideas and duality and also prove some rigorous theorems. The counterpart paper (largely successful): J. M. Borwein and D. R. Luke, Entropic Regularization of the l 0 function. In Fixed-Point Algorithms for Inverse Problems in Science and Engineering, Springer Optimization and Its Applications. In press, 2011. Available at http://carma.newcastle.edu.au/jon/sensing.pdf.

Motivation A variational/geometrical interpretation of the Candes-Tao (2004) probabilistic criterion for the solution to (P 1 ) to be unique and exactly match the true signal x. As a by-product, better practical methods for solving the underlying problem. Aim to use entropy/penalty ideas and duality and also prove some rigorous theorems. The counterpart paper (largely successful): J. M. Borwein and D. R. Luke, Entropic Regularization of the l 0 function. In Fixed-Point Algorithms for Inverse Problems in Science and Engineering, Springer Optimization and Its Applications. In press, 2011. Available at http://carma.newcastle.edu.au/jon/sensing.pdf.

Outline 1 Dual Convex (Entropic) Regularization 2 Subgradient Descent with Exact Line-search 3 Our Main Theorem 4 Computational Results 5 Conclusion and Questions

Fenchel duality The Fenchel-Legendre conjugate f : X ], + ] of f is f (x ) := sup{ x, x f (x)}. x X For the l 1 problem, the norm is proper, convex, lsc and b core (A dom f ) so strong Fenchel duality holds. That is: Program inf x R n{ x 1 : Ax = b} = sup { b, y (A y) 1} y R m where x 1 = ι [ 1,1] (x ) is zero on the supremum ball and is infinite otherwise.

Elementary Observations The dual to (P 1 ) is Program (D 1 ) maximize b T y y R m subject to (A y) j [ 1, 1] j = 1, 2,..., n. The solution includes a vertex of the constraint polyhedron. Uniqueness of primal solutions depends on whether dual solutions live on edges or faces of the dual polyhedron. We deduce that if a solution x to (P 1 ) is unique, then m { number of active constraints in (D 1 ) } = x 0.

Elementary observations The l 0 function is proper, lsc but not convex, so only weak Fenchel duality holds: Program inf { x x R n 0 Ax = b} sup { b, y (A y) 0}. y R m where x 0 := { 0 x = 0 + else

Elementary observations In other words, the dual to (P 0 ) is Program (D 0 ) maximize b T y y R m subject to A y = 0. primal problem is a combinatorial optimization problem. dual problem, however, is a linear program, which is finitely terminating. The solution to the dual problem is trivial: y = 0... which tells us nothing useful about the primal problem.

The Main Idea Relax and Regularize the dual and either solve this directly, or solve the corresponding regularized primal problem, or some mixture. Performance will be part art and part science and tied to the specific problem at hand. We do not expect a method which works all of the time...

The Fermi-Dirac Entropy (1926) The Fermi-Dirac entropy is ideal for [0, 1] problems: FD(x) := j x j log(x j ) + (1 x j ) log(1 x j ). Fermi-Dirac in 1-dim A Legendre barrier function with smooth finite conjugate FD (y) := j log (1 + e y j ).

Regularization/Relaxation: Fermi-Dirac Entropy For L, ε > 0, define a shifted nonnegative entropy: f ε,l(x) := n j 1 [ ( (L + xj ) ln(l + x j ) + (L x j ) ln(l x j ) ε ln(l) )] 2L ln(2) ln(2) for x [ L, L] n := + for x > L. Then f ε,l(x) = n j=1 [ ε ( ) ] ln(2) ln 4 xjl/ε + 1 x j L ε. (1) f is proper, convex and lsc, thus f = f. We set f := f.

Regularization/Relaxation: Fermi-Dirac Entropy

Regularization/Relaxation: Fermi-Dirac Entropy For L > 0 fixed, in the limit as ε 0 we have { 0 y [ L, L] lim f ε,l(y) = ε 0 + else and lim f ε,l (x) = L x. ε 0 For ε > 0 fixed we have lim f ε,l(y) = L 0 { 0 y = 0 + else and lim f ε,l (x) := 0. L 0 0 and fε0 have the same conjugate; 0 0 while fε0 = fε0 ; f ε,l and fε,l are convex and smooth on the interior of their domains for all ε, L > 0. ( ) This is in contrast to the metrics of the form j x j p which are nonconvex for p < 1.

Regularization/Relaxation: Fermi-Dirac Entropy For L > 0 fixed, in the limit as ε 0 we have { 0 y [ L, L] lim f ε,l(y) = ε 0 + else and lim f ε,l (x) = L x. ε 0 For ε > 0 fixed we have lim f ε,l(y) = L 0 { 0 y = 0 + else and lim f ε,l (x) := 0. L 0 0 and fε0 have the same conjugate; 0 0 while fε0 = fε0 ; f ε,l and fε,l are convex and smooth on the interior of their domains for all ε, L > 0. ( ) This is in contrast to the metrics of the form j x j p which are nonconvex for p < 1.

Regularization/Relaxation: FD Entropy Hence we aim to solve Program (D L,ε ) minimize y R m f L,ε(A y) b, y for appropriately updated L and ε. This is a convex optimization problem, so equivalently we solve the inclusion: 0 A f L,ε(A y) b (DI) We can also model more realistic relaxed inequality constraints such as Ax b δ (JMB-Lewis)

Regularization/Relaxation: FD Entropy Hence we aim to solve Program (D L,ε ) minimize y R m f L,ε(A y) b, y for appropriately updated L and ε. This is a convex optimization problem, so equivalently we solve the inclusion: 0 A f L,ε(A y) b (DI) We can also model more realistic relaxed inequality constraints such as Ax b δ (JMB-Lewis)

Regularization/Relaxation: FD Entropy Hence we aim to solve Program (D L,ε ) minimize y R m f L,ε(A y) b, y for appropriately updated L and ε. This is a convex optimization problem, so equivalently we solve the inclusion: 0 A f L,ε(A y) b (DI) We can also model more realistic relaxed inequality constraints such as Ax b δ (JMB-Lewis)

Outline 1 Dual Convex (Entropic) Regularization 2 Subgradient Descent with Exact Line-search 3 Our Main Theorem 4 Computational Results 5 Conclusion and Questions

ε = 0: f L,0 = ι [ L,L] n Solve 0 A f L,0(A y) b via subgradient descent: Program Given y, choose v f L,0 (A y ), λ 0 and construct y + as y + := y + λ (b Av ). Only two issues remain: Direction and Step size (a) how to choose direction v f L,0 (A y ) (b) how to choose step length λ.

ε = 0: (a) Choose v f L,0 (A y ) Recall that fl,0 = ι [ L,L] n so ι [ L,L] n(x ) = N [ L,L] (x ) (normal cone) = {v R n ±v j 0 (j J ± ), v j = 0 (j J 0 )} where J := {j N x j = L}, J + := {j N x j = L} and J 0 := {j N x j ] L, L[}. Program Choose v N [ L,L] (A y ) to be the solution to minimize 1 v R n 2 b Av 2 subject to v N [ L,L] n(a y )

ε = 0: Choose v f L,0 (A y ) That is: Program ( Pv ) minimize 1 v R n 2 b Av 2 subject to v N [ L,L] n(a y ) Defining we reformulate as: Program B := {v R n Av = b}, ( Pv ) minimize v R n β 2(1 β) dist 2 (v, B) + ι N[ L,L] n(a y )(v), for given 1 2 < β < 1.

ε = 0: Choose v f L,0 (A y ) Approximate (dynamic) averaged alternating reflections: Program We choose v (0) R n. For ν N set v (ν+1) := 1 ) (R 1 (R 2 v (ν) + ε ν + ρ ν + v (ν)), (2) 2 - where R 1 x := 2 prox β 2(1 β) dist x x (v,b)2 R 2 x := 2 prox x x ιn[ L,L] n (A y ) {ε ν } and {ρ ν } are the errors at each iteration, assumed summable. In the dynamic version we may adjust L as we go.

ε = 0: Choose v f L,0 (A y ) Luke (2005 08) shows (2) is equivalent to Inexact Relaxed Averaged Alternating Reflections: Program Choose v (0) R n and β [1/2, 1[. For ν N set v (ν+1) := β 2 ( +(1 β) (R B (R N[ L,L]n(A y )v (ν) + ε n ) + ρ n + v (ν)) P N[ L,L] n(a y )v (ν) + ε n 2 ). (3) where R B := 2P B I also for R N[ L,L] n(a y ). We can show: Lemma (Luke 2008, Combettes 2004) The sequence {v (ν) } ν=1 converges to v where P Bv solves (P v ).

ε = 0: (b) Choose λ Exact line search: choose largest λ that solves Program minimize λ R + f L,0(A y + A λ(b Av )) Note that f L,0 (A y + A λ(b Av )) = 0 = min f for all A y + A λ(b Av ) [ L, L] n. So we solve: Program (Exact line-search) (P λ ) minimize λ R + subject to λ λ(a (b Av )) j L (A y ) j λ(a (b Av )) j L (A y ) j j = 1,..., n

ε = 0: Choose λ Exact line search is practicable: Define and set λ := min J + := {j (A (b Av )) j > TOL}, J := {j (A (b Av )) j < TOL} { minj J+ {(L (A y ) j )/(A (b Av )) j }, min j J {( L (A y ) j )/(A (b Av )) j } } Relies on simulating exact arithmetic... harder for ε > 0. Full algorithm terminates when current v is such that J + = J = Ø.

ε = 0: Choose λ Exact line search is practicable: Define and set λ := min J + := {j (A (b Av )) j > TOL}, J := {j (A (b Av )) j < TOL} { minj J+ {(L (A y ) j )/(A (b Av )) j }, min j J {( L (A y ) j )/(A (b Av )) j } } Relies on simulating exact arithmetic... harder for ε > 0. Full algorithm terminates when current v is such that J + = J = Ø.

Choosing dynamically reweighted weights: details The algorithm in our paper has a nice convex-analytic criterion for choosing the reweighting parameter L k : L k is chosen so the projection of the data onto the normal cone of the rescaled problem at the rescaled iterate y k lies in the relative interior to said normal cone. This guarantees orthogonality of the search directions to the (rescaled) active constraints. What are optimal (in some sense) reweightings L k (dogmatic or otherwise)?

Choosing dynamically reweighted weights: details The algorithm in our paper has a nice convex-analytic criterion for choosing the reweighting parameter L k : L k is chosen so the projection of the data onto the normal cone of the rescaled problem at the rescaled iterate y k lies in the relative interior to said normal cone. This guarantees orthogonality of the search directions to the (rescaled) active constraints. What are optimal (in some sense) reweightings L k (dogmatic or otherwise)?

Choosing dynamically reweighted weights: details The algorithm in our paper has a nice convex-analytic criterion for choosing the reweighting parameter L k : L k is chosen so the projection of the data onto the normal cone of the rescaled problem at the rescaled iterate y k lies in the relative interior to said normal cone. This guarantees orthogonality of the search directions to the (rescaled) active constraints. What are optimal (in some sense) reweightings L k (dogmatic or otherwise)?

Choosing dynamically reweighted weights: details The algorithm in our paper has a nice convex-analytic criterion for choosing the reweighting parameter L k : L k is chosen so the projection of the data onto the normal cone of the rescaled problem at the rescaled iterate y k lies in the relative interior to said normal cone. This guarantees orthogonality of the search directions to the (rescaled) active constraints. What are optimal (in some sense) reweightings L k (dogmatic or otherwise)?

Outline 1 Dual Convex (Entropic) Regularization 2 Subgradient Descent with Exact Line-search 3 Our Main Theorem 4 Computational Results 5 Conclusion and Questions

Sufficient sparsity The mutual coherence of a matrix A is defined as µ(a) := max 1 k,j n, k j a T k a j a k a j where 0/0 := 1 and a j denotes the jth column of A. Mutual coherence measures the dependence between columns of A. The mutual coherence of unitary matrices, for instance, is zero; for matrices with columns of zeros, the mutual coherence is 1. What is a variational analytic/geometric interpretation (constraint qualification or the like) of the mutual coherence condition (4) below?

Sufficient sparsity The mutual coherence of a matrix A is defined as µ(a) := max 1 k,j n, k j a T k a j a k a j where 0/0 := 1 and a j denotes the jth column of A. Mutual coherence measures the dependence between columns of A. The mutual coherence of unitary matrices, for instance, is zero; for matrices with columns of zeros, the mutual coherence is 1. What is a variational analytic/geometric interpretation (constraint qualification or the like) of the mutual coherence condition (4) below?

Sufficient sparsity Lemma (uniqueness of sparse representations (Donoho Elad, 2003)) Let A R m n (m < n) be full rank. If there exists an element x such that Ax = b and x 0 < 1 ( 1 + 1 ), (4) 2 µ(a) then it is unique and sparsest possible (has minimal support). In the case of matrices that are not full rank and thus unitarily equivalent to matrices with columns of zeros only the trivial equation Ax = 0 has a unique sparsest possible solution.

A Precise Result Theorem (Recovery of sufficiently sparse solutions) Let A R m n (m < n) be full rank (denote jth column of A by a j ). Initialize the algorithm above with y 0 and weight L 0 such that y 0 j = 0 and L 0 j = a j for j = 1, 2,..., n. If x R n with Ax = b satisfies (4), then, with tolerance τ = 0, we converge in finitely many steps to a point y and a weight L where, argmin { Aw b 2 w N RL (y )} = x, the unique sparsest solution to Ax = b. We showed a greedy adaptive rescaling of our Algorithm is equivalent to a well-known greedy algorithm: Orthogonal Matching Pursuit (Bruckstein Donoh-Elad 09).

Greedy or What? Orthogonality of the search directions is important for guaranteeing finite termination, but it is a strong condition to impose, and is really the mathematical manifestation of what it means to be a greedy algorithm. I think greed is a misnomer because what really happens is that you forgo any RECOURSE once a decision has been made: the orthogonality condition means that once you ve made your decision, you don t have the option of throwing some candidates out of your active constraint set. I d call the strategy dogmatic, but as a late arrival to the scene I don t get naming rights. Russell Luke

Outline 1 Dual Convex (Entropic) Regularization 2 Subgradient Descent with Exact Line-search 3 Our Main Theorem 4 Computational Results 5 Conclusion and Questions

Computational Results The image of the solution to the dual y under A A y Used 2 2 7 length real vectors with 70 non-zero entries. As often, this is a qualitative solution to the primal: yielding location and sign of nonzero signal elements.

Computational Results The primal solution x as determined by the solution to Program (P y ) minimize 1 x R n 2 b Ax 2 subject to x N [ L,L] n(a y) where y solves the dual with L := 0.001 and l error of 10 12.

Computational Results Observations: Inner iterations can be shown to be arbitrarily slow: the solution sets to the subproblems are not metrically regular, and the indicator function ι N[ L,L] n is not coercive in the sense of Lions. The algorithm fails when there are too few samples relative to the sparsity of the true solution. All-in-all the method seems highly competitive (and there is still much to tune). It s generally the way with progress that it looks much greater than it really is. Ludwig Wittgenstein (1889 1951)

Outline 1 Dual Convex (Entropic) Regularization 2 Subgradient Descent with Exact Line-search 3 Our Main Theorem 4 Computational Results 5 Conclusion and Questions

Conclusion We have given a finitely terminating subgradient descent algorithm one specialization of which yields a variational interpretation and proof of a known greedy algorithm. Work in progress: 1 Characterize recoverability of true solution in terms of the regularity of the subproblem Program ( Pv ) minimize 1 v R n 2 b Av 2 subject to v N [ L,L] n(a y ) 2 Recovery of. 0 in the limit; not just its convex envelope, 0. 3 Robust code with parameters automatically adjusted (eventually) appears insensitive to L but weighted norms seem useful; also for ε > 0 and non-dogmatically.

Conclusion Thank you...