Least squares regularized or constrained by L0: relationship between their global minimizers. Mila Nikolova

Size: px
Start display at page:

Download "Least squares regularized or constrained by L0: relationship between their global minimizers. Mila Nikolova"

Transcription

1 Least squares regularized or constrained by L0: relationship between their global minimizers Mila Nikolova CMLA, CNRS, ENS Cachan, Université Paris-Saclay, France SIAM Minisymposium on Trends in the Mathematics of Signal Processing and Imaging organized by Willi Freeden and Zuhair Nashed Joint Mathematics Meetings, January 6 9, 2016, Seattle

2 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 2

3 1. Two optimization problems involving the l 0 pseudo norm A = ( A 1,, A N ) R M N (matrix) N > M d R M \ {0} (data) A vector û R N is k-sparse if u 0 := { } i : u[i] 0 k. One looks for a sparse vector û such that Aû d. Two desirable optimization problems to find a sparse û: (C k ) min u R N Au d 2 2 subject to u 0 k (constrained) (R β ) F β (u) = Au d β u 0 β > 0 (regularized) These are NP hard (combinatorial) nonconvex problems. Our goal: Clarify the relationship between the global minimizers of (R β ) and (C k ). [M. N., ACHA 2016]. 3

4 Applications: signal and image processing, sparse coding, compression, dictionary building, compressive sensing, machine learning, model selection, classification... 0 has served as a regularizer or as a penalty for a long time Markov random fields, MAP F β (u) = Au d β Du 0 Geman & Geman (1984), Besag (1986) - labeled images, stochastic algorithms Robini & Reissman (2012) - global convergence / computation speed (!) Subset selection via (R β ) - numerous algorithms - c.f. textbook Miller (2002) (C k ) natural sparse coding constraint. Also the best K-term approximation [DeVore 1998)]. Sparse-Land, M < N - strong assumptions on A (RIP, spark, etc.) / various approximations. A huge amount of papers with approximating algorithms, e.g. Haupt & Nowak (06), Blumensath & Davies (08), Tropp (10), Zhang et al (12), Beck & Eldar (14) Typical assumptions: RIP or K spark(a) plus others (e.g. bounds on A etc.) Important progress in solving problems (C k ) and (R β ). The numerical schemes common points. = Explore the relationship between their optimal sets. 4

5 The optimal values / the optimal solution setsof problems (C k ) and (R β ): (C k ) c k := inf { Au d 2 u R N and u 0 k } Ĉ k := { u R N and u 0 k Au d 2 = c k } (R β ) r β := inf { F β (u) u R N} R β := { u R N F β (u) = r β } Theorem 1 For any d R M : Ĉ k k and Rβ β > 0. H 1 Assumption : rank(a) = M < N no further reminder How to evaluate the extent of assumption dependent properties? Definition 1 A property is generic on R M if it holds on a subset of R M \ S where S is closed in R M and its Lebesgue measure in R M is null. A generic property is stronger than a property that holds only with probability one. I n := ( {1,..., n}, < ) and I 0 n := ( {0, 1,..., n}, < ) (totally strictly ordered) L := min { k I N c k = 0 } (uniquely defined) generically L = M 5

6 Main results There is a strictly decreasing sequence {β k } k J {β Jk } for J I L such that û is global minimizer of F β for β ( β Jk, β Jk 1 ) û is global minimizer of (C Jk ) Equivalently { Rβ β ( β Jk, β Jk 1 ) } = ĈJ k k J In a generic sense R βjk = ĈJ k ĈJ k+1 All β Jk s are obtained from the optimal values c k s of the problems (C k ), k I 0 L. The global minimizers of problems (C k ) and (R β ) are strict and generically uniques J is always nonempty For any n I 0 L \ J the global minimizers of (C n) are not global minimizers of (R β ) β When J = I 0 L, problems (C k) and (R β ) are quasi-completely equivalent: { Rβ β (β k, β k 1 )} = Ĉk β k = c k c k+1 k 6

7 Notation. :=. 2. { } supp (u) := i I N : u[i] 0 For any ω I 0 N A ω := ( A ω1,..., A ω ω ) R M ω, A T ω is the transposed of A ω u ω := ( u ω1,..., u ω ω ) T R ω Definition 2 Let f : R N R and S R N. Consider the problem min {f(u) u S}. û is a strict minimizer if there is a neighborhood O S, û O so that f(u)>f(û) u O\ {û}. û is an isolated (local) minimizer if û is the only minimizer in an open subset O O 7

8 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) Preliminaries On the global minimizers of problem (C k ) Necessary and sufficient conditions 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 8

9 2. Common optimality conditions for (C k ) and (R β ) Goal: Derive tests relating the optimal solutions of (C k ) and (R β ). 2.1 Preliminaries A constrained quadratic optimization problem: given d R M and ω I N (P ω ) min u R N Au d 2 subject to u[i] = 0 i I 0 N \ ω The convex problem (P ω ) always has solutions, for any ω I 0 N and for any d RM. Some useful facts on the relation of (P ω ) to (C k ) and (R β ) [M. N. SIIMS 2013] (R β ) û solves (P ω ) for some ω I 0 N û is a (local) minimizer of F β, β > 0 û solves (P ω ) for ω I 0 N with rank(a ω ) = ω û = strict (local) minimizer of F β, β (C k ) û solves (P ω ) for ω I 0 N with ω = k û is a (local) minimizer of (C k) Remark 1 For any ω I N with rank(a ω ) = ω, the minimizer û of (P ω ) is isolated. 9

10 2.2 On the optimal solution sets of problem (C k ) Lemma 1 c 0 = d 2 and {c k } k 0 is decreasing with c k = 0 k M. Lemma 2 For k I M let (C k ) have a global minimizer û obeying û 0 = k n for n 1. Then Aû = d. Furthermore û Ĉm and c m = 0 m k n. L := min { k I M c k = 0 }. Example 1 One has L M 1 if d = Au for u 0 M 1. Theorem 2 û Ĉk for k I 0 L = k L + 1 = ĈL Ĉk. û 0 = k = rank (A σ ) for σ := supp (û) so û is a strict global minimizer of (C k ). Corollary 1 Ĉk Ĉn = (k, n) ( I 0 L )2 such that k n. 10

11 Example 2 A = and d = 1 1 û = (0, 0, 1) T = Ĉ1 (strict, rank(a 3 ) = û 0 ) = c 1 = 0 = L = 1. û = (1, 1, 0) T is a strict global minimizer of (C 2 ) because rank ( ) A supp (û) = 2 and c2 = 0. A = and d = 1 0 Ĉ 1 = { (1, 0, 0, 0) T, (0, 0, 1, 0) T} (strict minimizers) = c 1 = 0 = L = 1. For k 2 all optimal solutions Ĉ1 are nonstrict and have the form û = (x, y, 1 x, y) T, x R \ {0, 1}. If y = 0 then û 0 = 2 and otherwise û 0 = 4. Remark 2 By Theorem 2 the optimal value c k of problem (C k ) for any k I 0 L obeys { } c k = min Aũ d 2 where ũ R N solves (P ω ) ω Ω k where Ω k := { ω I N ω = k = rank (A ω ) }. 11

12 2.3. Necessary and sufficient conditions Proposition 1 û R û β = Ĉk where k := û 0 I 0 L Ĉ k R β for k := û 0 I 0 L The global minimizers of F β are composed of some optimal sets Ĉk for k L. Ĉ = the collection of all optimal solutions Ĉk of problems (C k ) for all k I 0 L; R = the set of all global minimizers R β of F β for all β > 0 Ĉ := L Ĉ k and R := k=0 β>0 R β. Theorem 3 R Ĉ. When β ranges on (0, + ), F β can have at most L + 1 different sets of global minimizers which are optimal solutions of (C k ) for k {0,..., L}. Theorem 4 For any k I 0 L one has: Ĉk R β if and only if F β (u) F β (û) 0 û Ĉk u Ĉ Ĉk = R β if and only if F β (u) F β (û) > 0 û Ĉk u Ĉ \ Ĉk 12

13 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) The entire list of parameter values Conditions for agreement between their optimal sets The effective parameters values 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 13

14 3. Parameter values for equality between optimal sets Definition 3 (Critical parameter values) { ck c k+n β k := max n { ck n c βk U k := min n 3.1. The entire list of parameter values } n {1,..., L k} } n {1,..., k} k I 0 L 1 and β L = 0, k I L and β U 0 β 1 := +. We have β L = 0 < βl U and β 0 < β0 U. The cases where β k < βk U will be of particular interest. Proposition 2 S finite union of vector subspaces of dimension M 1 such that d R M \ S = β k βk U k I 0 L. β k β U k k I 0 L is a generic property. 14

15 3.2. Conditions for agreement between the optimal sets of (C k ) and (R β ) Theorem 5 k I 0 L Ĉ k R β if and only if β 0 β < β U 0 for k = 0 ; β k β β U k for k {1,..., L 1} ; β L < β β U L for k = L. Ĉ k = R β if and only if β k < β < β U k. Proof based on Theorem 4. To exploit Theorem 5 we have to clarify the links between (β k, β U k ) and c k 15

16 3.3. The effective parameters values The global minimizers of F β are always in Ĉ (Theorem 3), so we are interested in the indexes k for which there exist values of β such that Ĉk R β. Their set is obtained from Theorem 5. Definition 4 The effective index set J J E : J := { k I 0 } L β k < βk U and J E := { m I 0 L β m = β U m}. Ordering: J = {J 0, J 1,..., J p } where p := J 1 and J k 1 < J k k. Further: (J 0 = 0, J p = L) J 2 with β J 1 := β U J 0 β U 0 = + and β Jp β L = 0. The set J is always nonempty. Lemma 3 R Ĉ k = if and only if k I 0 L \ {J J E }. 16

17 Definition 3 for reminder: { ck c k+n β k := max n { βk U ck n c k := min n } n {1,..., L k} k I 0 L 1 and β L := 0, } n {1,..., k} k I L and β0 U := +. Simplification of {β k, βk U} k J J E Proposition 3 Let {β k, βk U } and J be as in Definition 3 and Definition 4, resp. Then (a) β Jk < βj U k = β Jk 1 J k J \ {J 0 } and β J U 0 β J 1 = +. (b) β Jk = c J k c Jk+1 J k+1 J k J k J \ {J p } and β Jp β L = 0. (c) { β m m J E } { β Jk J k J \ {J p } }. {β k } k J is strictly decreasing and its first entry is β 0. 17

18 Example 3 Let {c k } L k=0 for L = 7 reads as c 0 = 48 c 1 = 40 c 2 = 30 c 3 = 22 c 4 = 14 c 5 = 10 c 6 = 4 c 7 = 0. By Definition 3 the sequences {β k, β U k } 7 k=0 are given by β 0 = 9 β 1 = 10 β 2 = 8 β 3 = 8 β 4 = 5 β 5 = 6 β 6 = 4 β 7 = 0 β U 0 = + β U 1 = 8 β U 2 = 9 β U 3 = 8 β U 4 = 8 β U 5 = 4 β U 6 = 5 β U 7 = 4 From Definition 4, J = { J 0 = 0, J 1 = 2, J 2 = 4, J 3 = 6, J 4 = 7} and J E = { 3 }. One has β Jk = β U J k+1 for any J k J (Proposition 3(a)). The formula in Proposition 3(b) holds. {β 3 3 J E } β 3 =β J1 =8 {β 3 3 J E } {β Jk J k J \ {J 4 }} (Proposition 3(c)). J E β J1 := {m J E β m = β J1 } = {3 J E J 1 < 3 < J 2 }, see Lemma 4. J has the smallest indexes so that {β k } k J = {9, 8, 5, 4, 0} is the longest strictly decreasing subsequence of {β k } 7 k=0 containing β 0 see Proposition 4. One has {β k } k J = {β k } k J for J := {0, 3, 4, 6, 7}; however, J 2 > J 2. 18

19 The location of {β m m J E } is given by the (probably empty) subsets J E β Jk := {m J E β m = β Jk }. Lemma 4 The sets J E β Jk fulfill J E β Jk = for k = p and for any k p 1 J E β Jk = {m J E J k < m < J k+1 }. J and {β k } k J are characterized next Proposition 4 Let {β k } L k=0 read as in Definition 3 and J as in Definition 4. Then 0 J and J contains the smallest indexes such that {β k } k J is the longest strictly decreasing subsequence of {β k } L k=0 containing β 0. In order to find the effective J and {β k } k J we need only {β k } L k=0 in Definition 3. 19

20 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 20

21 4. Equivalence relations between the optimal sets of (C k ) and (R β ) 4.1. Partial equivalence Theorem 6 Let {β k } be as in Definition 3 and J as in Definition 4. Then: { Rβ β ( β Jk, β Jk 1 ) } = ĈJ k J k J, ( p [ βjk, β Jk 1 ] ) [ β J0, β J 1 ) = [ 0, + ). n=1 { } βj0,..., β Jp 1 in Definition 4 partition (0, + ) into J proper intervals. For any β ( ) β Jk, β Jk 1 the optimal sets of (Rβ ) and of (C n ) for n = J k coincide. If I 0 L \ J, the optimal sets (C k ) for k I 0 L \ J cannot be optimal solutions of (R β ) β > 0. = partial equivalence. {β Jk } is a finite set of isolated values hence β β Jk generically. 21

22 The optimal sets of problem (R β ) for β k, k J: Theorem 7 Let H1 hold. Let {β k } be as in Definition 3 and J as in Definition 4. Then R βjk = ĈJ k ĈJ k+1 ( m J E β Jk Ĉ m ) J k J \ {J p }, where J E β Jk = {m J E J k < m < J k+1 } and Ĉk Ĉn = (k, n) (J J E ) 2, k n. Example 4 [Ex.3, cont.] We had J={0, 2, 4, 6, 7}, (β J0 =9, β J1 =8, β J2 =5, β J3 =4, β J4 =0), and J E 2 = {3} with β J E 2 =8 and J E k = otherwise. By Theorems 6 and 7 { R β β >9}=Ĉ0 { R β β (8, 9)}=Ĉ2 { R β β (5, 8)}=Ĉ4 { R β β (4, 5)}=Ĉ6 { R β β (0, 4)}=Ĉ7 and Rβ=9 = Ĉ0 Ĉ2 Rβ=8 = Ĉ2 Ĉ3 Ĉ4 Rβ=5 = Ĉ4 Ĉ6 Rβ=4 = Ĉ6 Ĉ7. A partial equivalence between problems (C k ) and (R β ) always exists. For the J 1 isolated values { β k k J \ {L} } problem (R β ) has normally two optimal sets (Proposition 2). 22

23 4.2. Quasi-complete equivalence Lemma 5 Let J be as in Definition 4. Then the following hold: (a) If the sequence {β k } L k=0 in Definition 3 is strictly decreasing, then its entries read as β k = c k c k+1 k I 0 L 1 and β L = 0, β 1 := β0 U = +. (1) (b) If the sequence {β k } L k=0 in (1) is strictly decreasing then J = I0 L. Theorem 8 Let {β k } L k=0 in (1) be strictly decreasing. Then { } Rβ β ( β k, β k 1 ) = Ĉk k I 0 L R βk = Ĉk Ĉk+1 with Ĉ k Ĉk+1 = k I 0 L 1. {β k } L k=0 in (1) strictly decreasing means that c k 1 c k > c k c k+1, k I L 1. Let u, û and ũ be optimal solutions of problems (C k 1 ), (C k ) and (C k+1 ), resp. Denote σ := supp (u), û := supp (û) and σ := supp (ũ). The condition on c k s reds as d T (Proj(A σ ) Proj(A σ )) d > d T (Proj(A σ ) Proj(A σ )) d > 0. Mid-way scenarios... 23

24 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 24

25 5. On the optimal values of (C k ) and (R β ) { } Ω k := ω I N ω = k = rank (A ω ). E k := range (A ω ) and G k := ω Ω k E 0 = G M = R M and E M = G 0 = {0} by H1. range (A ω ). ω Ω k Proposition 5 Let L M be arbitrarily fixed. c k > 0 k L 1 d R M \ G L 1 ; d R M \ (E 2 G L 1) = c k 1 > c k k I L. E 2 and G M 1 are finite unions of vector subspaces of dimensions M 2 and M 1, respectively. Hence, d R M \ (E 2 G M 1 ) is a generic property. = {c k } M k=0 is strictly decreasing and L = M generically. 25

26 Proposition 6 k I 0 L = F β (û) = c k + β k û Ĉk. By Theorem 3 R β Ĉ = L k=0 Ĉk = the optimal value of problem (R β ) reads as r β = min { c k + β k k I 0 L }. Corollary 2 The application β r β r β = c Jk + β J k : (0, + ) R fulfills = F β (û) û ĈJ k if and only if β β r β is continuous and concave. [β J0, + ) for J 0 0 [ ] βjk, β Jk 1 for Jk J \ {0, L} ( 0, βjp 1] for J p L r βjk 1 > r βjk J k J, r βj0 = c J0 = r β β β J0 and r βj0 > r β β < β J0. β r β is affine increasing on each interval ( β Jk, β Jk 1 ) with upward kinks at βjk J k J \ {L} and bounded by c 0. for any 26

27 Example 5 [Cont. of Example 3] From Corollary 2, β r β is given by β (0, 4] r β = c 7 + 7β = 7β r β=4 = 28 β [4, 5] r β = c 6 + 6β = 4 + 6β r β=5 = 34 β [5, 8] r β = c 4 + 4β = β r β=8 = 46 β [8, 9] r β = c 2 + 2β = β r β=9 = 48 β [9, + ) r β = c 0 + 0β = 48 affine expressions 27

28 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 28

29 6. Cardinality of the optimal sets of (C k ) and of (R β ) For any β > 0 and k I 0 L the optimal sets of problems (C k ) and (R β ) are composed out of a certain finite number of isolated (hence strict) minimizers Uniqueness of the optimal solutions of (C k ) and of (R β ) k min{l, M 1} and (û, ũ) (Ĉk) 2, û ũ. Then σ := supp (û), σ := supp (ũ) are in (Ω k ) 2. c k = A σ û σ d 2 = A σ ũ σ d 2 where σ σ. Π ω the orthogonal projector onto range (A ω ) A σ û σ d 2 A σ ũ σ d 2 = d T (Π σ Π σ ) d = 0. H For K min{m 1, L} fixed, A R M N obeys Π ω Π ω (ω, ω) Ω 2 k ω ω k I K. H is a generic property of all matrices in R M N [M. N., SIIMS 2013]. K := K k=1 dim( K ) M 1, hence d R M \ K generically. (ω,ω) (Ω k ) 2 { g R M ω ω and g ker (Π ω Π ω ) } H and d R M \ K = (C k ) for k I K has a unique optimal solution. K := max {k J k K} (R β ) has a unique global minimizer β (β K, + ) \ {β k } k J 29

30 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests Monte Carlo on {β k } with 10 5 tests for (M, N) = (5, 10) Tests on (partial) equivalence with selected matrix data 8. Conclusions and future directions 30

31 7. Numerical tests Two kind of experiments using matrices A R M N for (M, N) = (5, 10), original vectors u o R N and data samples d = Au o (+noise) with two different goals: to roughly see the behaviour of the parameters β k in Definition 3 ; to verify and illustrate our theoretical findings. All results were calculated using an exhaustive combinatorial search Monte Carlo experiments on {β k } with 10 5 tests for (M, N) = (5, 10) Two experiments, each one composed of 10 5 trials with A R M N for (M, N) = (5, 10) In each trial: an original u o R N, random support on {1,..., N} with u o 0 M 1 = 4. the coefficients of A and u o supp (u o ) i.i.d. d = Au o + i.i.d. centered Gaussian noise. compute the optimal values {c k } and then compute (β k, βk U ) by Definition 3. 31

32 Two different distributions for A and u o supp (u o ) Experiment N (0, 10). A(i, j) and u o supp (u o ) N (0, 10). Support length supp (u o ) {1,..., 4}, mean = 3.8. SNR of d in [10.1, 61.1], mean = db. Experiment Uni [0, 10]. A(i, j) and u o supp (u o ) uniform on [0, 10]. supp (u o ) {1,..., 4}, mean = 3.8. SNR of d in [20, 55], mean = db. N k := { k I 0 M β k > β k 1 }. Table 1: Results on the behaviour of {β k } in Definition 3 for two experiments, each one composed of 10 5 random trials. For k 3 we have found N k = 0. β k < β k 1, k I 0 M N k = 1 N k = 2 mean(snr) mean( u o 0 ) N (0, 10) % % % Uni [0, 10] % % % Observations: L = M in each trial (Remainder: L := min { k I N c k = 0 } ); {c k } M k=0 was always strictly decreasing (see Proposition 5); β k β U k in each trial (see Proposition 3), so J E = ; For every A there were d so that {β k := c k c k 1 } L k=0 was strictly decreasing 32

33 6.2. Tests on quasi-equivalence with a selected matrix and selected data A = ( ) u o T = A(i, j) nearly normal distribution with variance 10 and rank(a) = M = 5 Problem (C M ) has Ω M = 252 optimal solutions; { none } of them is shown. Since β 0 < β0 U = +, in all cases Ĉ0 = Rβ β > β 0 by Theorem 5. We selected a couple (A, u o ) so that β k are seldom strictly decreasing compared to Tab. 1. Table 2: 10 5 trials where d = Au o + i.i.d. centered Gaussian noise. β k < β k 1, k I 0 M N k = 1 N k = 2 mean(snr) mean( u o 0 ) u o in (2) % % 0 %

34 Noise-free data d = Au o = ( ) T. û = u o is an optimal solution to problems (C k ) with c k = 0 for k {3, 4, 5} and L = 3. β 3 = 0 < β U 3 = β 1 = < β U 1 = β 0 = and β 2 = 3968 > β U 2 = J = {0, 1, 3} k c k Ĉ k = the optimal solution of (C k ), singleton Ĉ k = R β β (β 3, β 1 ) no β (β 1, β 0 ) β > β 0 34

35 Noisy data 1. Nearly normal, centered, i.i.d. noise and SNR= db: ( ) T d = β 5 = 0 < β U 5 = β 4 = < β U 4 = β 3 = < β U 3 = β 1 = < β U 1 = β 0 = 63154, while β 2 = > β U 2 = Hence, J = I 0 5 \ {2} k c k Ĉ k = the optimal solution of (C k ), singleton Ĉ k = R β β (β 4, β 3 ) β (β 3, β 1 ) no β (β 1, β 0 ) β > β 0 35

36 Noisy data 2. The noise is nearly normal, centered, i.i.d., SNR= db: ( ) T d = β 0 = β 1 = 3825 β 2 = β 3 = β 4 = β 5 = 0. {β k } is strictly decreasing and hence β k = c k c k 1 (C k ) and (R β ) are quasi-completely equivalent. k c k Ĉ k = the optimal solution of (C k ), singleton Ĉ k = R β β (β 4, β 3 ) β (β 3, β 2 ) β (β 2, β 1 ) β (β 1, β 0 ) β > β 0 36

37 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 37

38 8. Conclusions and open questions The main equivalence result in a nutshell: Ĉ Jk ĈJ k+1 Ĉ0 J k }{{} R β = Ĉ L Ĉ Jk+1 Ĉ Jk Ĉ 0 0 = β L <... < β Jk+1 < β Jk < β Jk 1 <... < β J0 < The agreement between the optimal sets of problems (C k ) and (R β ) is driven by the critical parameters {β k } L k=0 which depend only on the optimal values c k of problem (C k ). Our comparative results clarify a proper choice between models (C k ) and (R β ) in applications. If one needs solutions with a fixed number of nonzero entries, (C k ) is the best choice. If only information on the perturbations is available, (R β ) is a more flexible model. If one can solve problem (C k ) for all k, the global minimizers of problem (R β ) are immediate. Our detailed results can give rise to innovative algorithms. The degree of partial equivalence depends on the distribution of the coefficients of A and d. 38

39 By specifying a class of matrices A and assumptions on data d, one can infer statistical knowledge on the optimal values c k of problems (C k ) and thus on the critical parameters {β k }. Promising theoretical and practical results can be expected. A related open question is to know if the optimal solutions of (R β ) are able to eliminate some meaningless solutions of (C k ). Extensions to analysis type penalties Du 0, to low rank matrix recovery, etc., are important. Other important results concern algorithms that are known to converge to local minimizers. Remark 3 Problem (R β ) (for some β > 0) and problems (C k ) for k {0, 1,..., M} have the same sets of (strict) local minimizers. 39

40 Thank you for the attention. Thanks to Zuhair Nashed for the invitation 40

Relationship between the optimal solutions of least squares regularized with l 0 -norm and constrained by k-sparsity

Relationship between the optimal solutions of least squares regularized with l 0 -norm and constrained by k-sparsity Applied and Computational Harmonic Analysis, Volume 41, Issue 1, July 2016, pages 237 265 http://dx.doi.org/10.1016/j.acha.2015.10.010 Relationship between the optimal solutions of least squares regularized

More information

Description of the minimizers of least squares regularized with l_0-norm. Uniqueness of the global minimizer

Description of the minimizers of least squares regularized with l_0-norm. Uniqueness of the global minimizer Description of the minimizers of least squares regularized with l_0-norm. Uniqueness of the global minimizer Mila Nikolova To cite this version: Mila Nikolova. Description of the minimizers of least squares

More information

Constrained optimization

Constrained optimization Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained

More information

Least Sparsity of p-norm based Optimization Problems with p > 1

Least Sparsity of p-norm based Optimization Problems with p > 1 Least Sparsity of p-norm based Optimization Problems with p > Jinglai Shen and Seyedahmad Mousavi Original version: July, 07; Revision: February, 08 Abstract Motivated by l p -optimization arising from

More information

Contre-examples for Bayesian MAP restoration. Mila Nikolova

Contre-examples for Bayesian MAP restoration. Mila Nikolova Contre-examples for Bayesian MAP restoration Mila Nikolova CMLA ENS de Cachan, 61 av. du Président Wilson, 94235 Cachan cedex (nikolova@cmla.ens-cachan.fr) Obergurgl, September 26 Outline 1. MAP estimators

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

Introduction to Compressed Sensing

Introduction to Compressed Sensing Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral

More information

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk Model-Based Compressive Sensing for Signal Ensembles Marco F. Duarte Volkan Cevher Richard G. Baraniuk Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional

More information

The Analysis Cosparse Model for Signals and Images

The Analysis Cosparse Model for Signals and Images The Analysis Cosparse Model for Signals and Images Raja Giryes Computer Science Department, Technion. The research leading to these results has received funding from the European Research Council under

More information

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina INDUSTRIAL MATHEMATICS INSTITUTE 2007:08 A remark on compressed sensing B.S. Kashin and V.N. Temlyakov IMI Preprint Series Department of Mathematics University of South Carolina A remark on compressed

More information

Lecture Notes 9: Constrained Optimization

Lecture Notes 9: Constrained Optimization Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form

More information

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions Linear Systems Carlo Tomasi June, 08 Section characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix

More information

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1 The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1 Simon Foucart Department of Mathematics Vanderbilt University Nashville, TN 3784. Ming-Jun Lai Department of Mathematics,

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Uniqueness Conditions for A Class of l 0 -Minimization Problems

Uniqueness Conditions for A Class of l 0 -Minimization Problems Uniqueness Conditions for A Class of l 0 -Minimization Problems Chunlei Xu and Yun-Bin Zhao October, 03, Revised January 04 Abstract. We consider a class of l 0 -minimization problems, which is to search

More information

An Introduction to Sparse Approximation

An Introduction to Sparse Approximation An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER 2011 7255 On the Performance of Sparse Recovery Via `p-minimization (0 p 1) Meng Wang, Student Member, IEEE, Weiyu Xu, and Ao Tang, Senior

More information

ANALYTICAL BOUNDS ON THE MINIMIZERS OF (NONCONVEX) REGULARIZED LEAST-SQUARES. Mila Nikolova. (Communicated by Jackie (Jianhong) Shen)

ANALYTICAL BOUNDS ON THE MINIMIZERS OF (NONCONVEX) REGULARIZED LEAST-SQUARES. Mila Nikolova. (Communicated by Jackie (Jianhong) Shen) Inverse Problems and Imaging Volume 1, No. 4, 2007, 661 677 Web site: http://www.aimsciences.org ANALYTICAL BOUNDS ON THE MINIMIZERS OF NONCONVEX) REGULARIZED LEAST-SQUARES Mila Nikolova CMLA, ENS Cachan,

More information

Necessary and Sufficient Conditions of Solution Uniqueness in 1-Norm Minimization

Necessary and Sufficient Conditions of Solution Uniqueness in 1-Norm Minimization Noname manuscript No. (will be inserted by the editor) Necessary and Sufficient Conditions of Solution Uniqueness in 1-Norm Minimization Hui Zhang Wotao Yin Lizhi Cheng Received: / Accepted: Abstract This

More information

ACCORDING to Shannon s sampling theorem, an analog

ACCORDING to Shannon s sampling theorem, an analog 554 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL 59, NO 2, FEBRUARY 2011 Segmented Compressed Sampling for Analog-to-Information Conversion: Method and Performance Analysis Omid Taheri, Student Member,

More information

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation

Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 12, DECEMBER 2008 2009 Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation Yuanqing Li, Member, IEEE, Andrzej Cichocki,

More information

SPARSE signal representations have gained popularity in recent

SPARSE signal representations have gained popularity in recent 6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying

More information

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html Acknowledgement: this slides is based on Prof. Emmanuel Candes and Prof. Wotao Yin

More information

EUSIPCO

EUSIPCO EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

Linear Systems. Carlo Tomasi

Linear Systems. Carlo Tomasi Linear Systems Carlo Tomasi Section 1 characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix and of

More information

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

Constructing Explicit RIP Matrices and the Square-Root Bottleneck Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry

More information

Self-Calibration and Biconvex Compressive Sensing

Self-Calibration and Biconvex Compressive Sensing Self-Calibration and Biconvex Compressive Sensing Shuyang Ling Department of Mathematics, UC Davis July 12, 2017 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 1 / 22 Acknowledgements

More information

The definition of a vector space (V, +, )

The definition of a vector space (V, +, ) The definition of a vector space (V, +, ) 1. For any u and v in V, u + v is also in V. 2. For any u and v in V, u + v = v + u. 3. For any u, v, w in V, u + ( v + w) = ( u + v) + w. 4. There is an element

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo January 29, 2012 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION CHRISTIAN GÜNTHER AND CHRISTIANE TAMMER Abstract. In this paper, we consider multi-objective optimization problems involving not necessarily

More information

GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA

GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA D Pimentel-Alarcón 1, L Balzano 2, R Marcia 3, R Nowak 1, R Willett 1 1 University of Wisconsin - Madison, 2 University of Michigan - Ann Arbor, 3 University

More information

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar

More information

DS-GA 1002 Lecture notes 12 Fall Linear regression

DS-GA 1002 Lecture notes 12 Fall Linear regression DS-GA Lecture notes 1 Fall 16 1 Linear models Linear regression In statistics, regression consists of learning a function relating a certain quantity of interest y, the response or dependent variable,

More information

THEORY OF COMPRESSIVE SENSING VIA l 1 -MINIMIZATION: A NON-RIP ANALYSIS AND EXTENSIONS

THEORY OF COMPRESSIVE SENSING VIA l 1 -MINIMIZATION: A NON-RIP ANALYSIS AND EXTENSIONS THEORY OF COMPRESSIVE SENSING VIA l 1 -MINIMIZATION: A NON-RIP ANALYSIS AND EXTENSIONS YIN ZHANG Abstract. Compressive sensing (CS) is an emerging methodology in computational signal processing that has

More information

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published

More information

Conditions for Robust Principal Component Analysis

Conditions for Robust Principal Component Analysis Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and

More information

Energy minimization methods

Energy minimization methods Handbook of Mathematical Methods in Imaging, Springer 211, 1 st edition Editor: Otmar SCHERZER Chapter 5, p. 138 186 Energy minimization methods Mila NIKOLOVA CMLA, ENS Cachan, CNRS, UniverSud, 61 Av.

More information

Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France.

Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France. Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France remi.gribonval@inria.fr Overview of the course Introduction sparsity & data compression inverse problems

More information

Recovery of Low Rank and Jointly Sparse. Matrices with Two Sampling Matrices

Recovery of Low Rank and Jointly Sparse. Matrices with Two Sampling Matrices Recovery of Low Rank and Jointly Sparse 1 Matrices with Two Sampling Matrices Sampurna Biswas, Hema K. Achanta, Mathews Jacob, Soura Dasgupta, and Raghuraman Mudumbai Abstract We provide a two-step approach

More information

Decomposition of Riesz frames and wavelets into a finite union of linearly independent sets

Decomposition of Riesz frames and wavelets into a finite union of linearly independent sets Decomposition of Riesz frames and wavelets into a finite union of linearly independent sets Ole Christensen, Alexander M. Lindner Abstract We characterize Riesz frames and prove that every Riesz frame

More information

MATH 8253 ALGEBRAIC GEOMETRY WEEK 12

MATH 8253 ALGEBRAIC GEOMETRY WEEK 12 MATH 8253 ALGEBRAIC GEOMETRY WEEK 2 CİHAN BAHRAN 3.2.. Let Y be a Noetherian scheme. Show that any Y -scheme X of finite type is Noetherian. Moreover, if Y is of finite dimension, then so is X. Write f

More information

RSP-Based Analysis for Sparsest and Least l 1 -Norm Solutions to Underdetermined Linear Systems

RSP-Based Analysis for Sparsest and Least l 1 -Norm Solutions to Underdetermined Linear Systems 1 RSP-Based Analysis for Sparsest and Least l 1 -Norm Solutions to Underdetermined Linear Systems Yun-Bin Zhao IEEE member Abstract Recently, the worse-case analysis, probabilistic analysis and empirical

More information

arxiv: v1 [math.na] 26 Nov 2009

arxiv: v1 [math.na] 26 Nov 2009 Non-convexly constrained linear inverse problems arxiv:0911.5098v1 [math.na] 26 Nov 2009 Thomas Blumensath Applied Mathematics, School of Mathematics, University of Southampton, University Road, Southampton,

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Optimization for Compressed Sensing

Optimization for Compressed Sensing Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve

More information

Recovery Guarantees for Rank Aware Pursuits

Recovery Guarantees for Rank Aware Pursuits BLANCHARD AND DAVIES: RECOVERY GUARANTEES FOR RANK AWARE PURSUITS 1 Recovery Guarantees for Rank Aware Pursuits Jeffrey D. Blanchard and Mike E. Davies Abstract This paper considers sufficient conditions

More information

Strengthened Sobolev inequalities for a random subspace of functions

Strengthened Sobolev inequalities for a random subspace of functions Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)

More information

A Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems

A Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems A Continuation Approach to Estimate a Solution Path of Mixed L2-L Minimization Problems Junbo Duan, Charles Soussen, David Brie, Jérôme Idier Centre de Recherche en Automatique de Nancy Nancy-University,

More information

Sparse Optimization Lecture: Sparse Recovery Guarantees

Sparse Optimization Lecture: Sparse Recovery Guarantees Those who complete this lecture will know Sparse Optimization Lecture: Sparse Recovery Guarantees Sparse Optimization Lecture: Sparse Recovery Guarantees Instructor: Wotao Yin Department of Mathematics,

More information

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence

More information

Recent Developments in Compressed Sensing

Recent Developments in Compressed Sensing Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline

More information

Sparse solutions of underdetermined systems

Sparse solutions of underdetermined systems Sparse solutions of underdetermined systems I-Liang Chern September 22, 2016 1 / 16 Outline Sparsity and Compressibility: the concept for measuring sparsity and compressibility of data Minimum measurements

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Linear Methods for Regression. Lijun Zhang

Linear Methods for Regression. Lijun Zhang Linear Methods for Regression Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Linear Regression Models and Least Squares Subset Selection Shrinkage Methods Methods Using Derived

More information

Matrix Support Functional and its Applications

Matrix Support Functional and its Applications Matrix Support Functional and its Applications James V Burke Mathematics, University of Washington Joint work with Yuan Gao (UW) and Tim Hoheisel (McGill), CORS, Banff 2016 June 1, 2016 Connections What

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption

BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption Daniel Reem (joint work with Simeon Reich and Alvaro De Pierro) Department of Mathematics, The Technion,

More information

Problem set 5: SVD, Orthogonal projections, etc.

Problem set 5: SVD, Orthogonal projections, etc. Problem set 5: SVD, Orthogonal projections, etc. February 21, 2017 1 SVD 1. Work out again the SVD theorem done in the class: If A is a real m n matrix then here exist orthogonal matrices such that where

More information

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit Deanna Needell and Roman Vershynin Abstract We demonstrate a simple greedy algorithm that can reliably

More information

Analysis of Greedy Algorithms

Analysis of Greedy Algorithms Analysis of Greedy Algorithms Jiahui Shen Florida State University Oct.26th Outline Introduction Regularity condition Analysis on orthogonal matching pursuit Analysis on forward-backward greedy algorithm

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo September 6, 2011 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

IMA Preprint Series # 2066

IMA Preprint Series # 2066 THE CARDINALITY OF SETS OF k-independent VECTORS OVER FINITE FIELDS By S.B. Damelin G. Michalski and G.L. Mullen IMA Preprint Series # 2066 ( October 2005 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS

More information

Chapter 1. Preliminaries

Chapter 1. Preliminaries Introduction This dissertation is a reading of chapter 4 in part I of the book : Integer and Combinatorial Optimization by George L. Nemhauser & Laurence A. Wolsey. The chapter elaborates links between

More information

Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach

Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach Boaz Nadler The Weizmann Institute of Science Israel Joint works with Inbal Horev, Ronen Basri, Meirav Galun and Ery Arias-Castro

More information

A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing

A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing Emilie Chouzenoux emilie.chouzenoux@univ-mlv.fr Université Paris-Est Lab. d Informatique Gaspard

More information

Geometry of log-concave Ensembles of random matrices

Geometry of log-concave Ensembles of random matrices Geometry of log-concave Ensembles of random matrices Nicole Tomczak-Jaegermann Joint work with Radosław Adamczak, Rafał Latała, Alexander Litvak, Alain Pajor Cortona, June 2011 Nicole Tomczak-Jaegermann

More information

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES Fenghui Wang Department of Mathematics, Luoyang Normal University, Luoyang 470, P.R. China E-mail: wfenghui@63.com ABSTRACT.

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu

More information

A New Estimate of Restricted Isometry Constants for Sparse Solutions

A New Estimate of Restricted Isometry Constants for Sparse Solutions A New Estimate of Restricted Isometry Constants for Sparse Solutions Ming-Jun Lai and Louis Y. Liu January 12, 211 Abstract We show that as long as the restricted isometry constant δ 2k < 1/2, there exist

More information

Discussion of Hypothesis testing by convex optimization

Discussion of Hypothesis testing by convex optimization Electronic Journal of Statistics Vol. 9 (2015) 1 6 ISSN: 1935-7524 DOI: 10.1214/15-EJS990 Discussion of Hypothesis testing by convex optimization Fabienne Comte, Céline Duval and Valentine Genon-Catalot

More information

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the

More information

Vector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then

Vector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then 1 Def: V be a set of elements with a binary operation + is defined. F be a field. A multiplication operator between a F and v V is also defined. The V is called a vector space over the field F if: V is

More information

Sparsity Regularization

Sparsity Regularization Sparsity Regularization Bangti Jin Course Inverse Problems & Imaging 1 / 41 Outline 1 Motivation: sparsity? 2 Mathematical preliminaries 3 l 1 solvers 2 / 41 problem setup finite-dimensional formulation

More information

An algebraic perspective on integer sparse recovery

An algebraic perspective on integer sparse recovery An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:

More information

The Information-Theoretic Requirements of Subspace Clustering with Missing Data

The Information-Theoretic Requirements of Subspace Clustering with Missing Data The Information-Theoretic Requirements of Subspace Clustering with Missing Data Daniel L. Pimentel-Alarcón Robert D. Nowak University of Wisconsin - Madison, 53706 USA PIMENTELALAR@WISC.EDU NOWAK@ECE.WISC.EDU

More information

Three Generalizations of Compressed Sensing

Three Generalizations of Compressed Sensing Thomas Blumensath School of Mathematics The University of Southampton June, 2010 home prev next page Compressed Sensing and beyond y = Φx + e x R N or x C N x K is K-sparse and x x K 2 is small y R M or

More information

arxiv: v1 [math.oc] 22 Sep 2016

arxiv: v1 [math.oc] 22 Sep 2016 EUIVALENCE BETWEEN MINIMAL TIME AND MINIMAL NORM CONTROL PROBLEMS FOR THE HEAT EUATION SHULIN IN AND GENGSHENG WANG arxiv:1609.06860v1 [math.oc] 22 Sep 2016 Abstract. This paper presents the equivalence

More information

CSC 576: Variants of Sparse Learning

CSC 576: Variants of Sparse Learning CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in

More information

The Pros and Cons of Compressive Sensing

The Pros and Cons of Compressive Sensing The Pros and Cons of Compressive Sensing Mark A. Davenport Stanford University Department of Statistics Compressive Sensing Replace samples with general linear measurements measurements sampled signal

More information

Joint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces

Joint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces Aalborg Universitet Joint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces Christensen, Mads Græsbøll; Nielsen, Jesper Kjær Published in: I E E E / S P Workshop

More information

Compressed Sensing and Sparse Recovery

Compressed Sensing and Sparse Recovery ELE 538B: Sparsity, Structure and Inference Compressed Sensing and Sparse Recovery Yuxin Chen Princeton University, Spring 217 Outline Restricted isometry property (RIP) A RIPless theory Compressed sensing

More information

Rank Determination for Low-Rank Data Completion

Rank Determination for Low-Rank Data Completion Journal of Machine Learning Research 18 017) 1-9 Submitted 7/17; Revised 8/17; Published 9/17 Rank Determination for Low-Rank Data Completion Morteza Ashraphijuo Columbia University New York, NY 1007,

More information

Information-Theoretic Limits of Matrix Completion

Information-Theoretic Limits of Matrix Completion Information-Theoretic Limits of Matrix Completion Erwin Riegler, David Stotz, and Helmut Bölcskei Dept. IT & EE, ETH Zurich, Switzerland Email: {eriegler, dstotz, boelcskei}@nari.ee.ethz.ch Abstract We

More information

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical

More information

DS-GA 1002 Lecture notes 10 November 23, Linear models

DS-GA 1002 Lecture notes 10 November 23, Linear models DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.

More information

Signal Recovery, Uncertainty Relations, and Minkowski Dimension

Signal Recovery, Uncertainty Relations, and Minkowski Dimension Signal Recovery, Uncertainty Relations, and Minkowski Dimension Helmut Bőlcskei ETH Zurich December 2013 Joint work with C. Aubel, P. Kuppinger, G. Pope, E. Riegler, D. Stotz, and C. Studer Aim of this

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Restricted Strong Convexity Implies Weak Submodularity

Restricted Strong Convexity Implies Weak Submodularity Restricted Strong Convexity Implies Weak Submodularity Ethan R. Elenberg Rajiv Khanna Alexandros G. Dimakis Department of Electrical and Computer Engineering The University of Texas at Austin {elenberg,rajivak}@utexas.edu

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 3: Sparse signal recovery: A RIPless analysis of l 1 minimization Yuejie Chi The Ohio State University Page 1 Outline

More information

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible. MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

Supremum of simple stochastic processes

Supremum of simple stochastic processes Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there

More information

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad

More information