Least squares regularized or constrained by L0: relationship between their global minimizers. Mila Nikolova
|
|
- Oswin Sutton
- 5 years ago
- Views:
Transcription
1 Least squares regularized or constrained by L0: relationship between their global minimizers Mila Nikolova CMLA, CNRS, ENS Cachan, Université Paris-Saclay, France SIAM Minisymposium on Trends in the Mathematics of Signal Processing and Imaging organized by Willi Freeden and Zuhair Nashed Joint Mathematics Meetings, January 6 9, 2016, Seattle
2 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 2
3 1. Two optimization problems involving the l 0 pseudo norm A = ( A 1,, A N ) R M N (matrix) N > M d R M \ {0} (data) A vector û R N is k-sparse if u 0 := { } i : u[i] 0 k. One looks for a sparse vector û such that Aû d. Two desirable optimization problems to find a sparse û: (C k ) min u R N Au d 2 2 subject to u 0 k (constrained) (R β ) F β (u) = Au d β u 0 β > 0 (regularized) These are NP hard (combinatorial) nonconvex problems. Our goal: Clarify the relationship between the global minimizers of (R β ) and (C k ). [M. N., ACHA 2016]. 3
4 Applications: signal and image processing, sparse coding, compression, dictionary building, compressive sensing, machine learning, model selection, classification... 0 has served as a regularizer or as a penalty for a long time Markov random fields, MAP F β (u) = Au d β Du 0 Geman & Geman (1984), Besag (1986) - labeled images, stochastic algorithms Robini & Reissman (2012) - global convergence / computation speed (!) Subset selection via (R β ) - numerous algorithms - c.f. textbook Miller (2002) (C k ) natural sparse coding constraint. Also the best K-term approximation [DeVore 1998)]. Sparse-Land, M < N - strong assumptions on A (RIP, spark, etc.) / various approximations. A huge amount of papers with approximating algorithms, e.g. Haupt & Nowak (06), Blumensath & Davies (08), Tropp (10), Zhang et al (12), Beck & Eldar (14) Typical assumptions: RIP or K spark(a) plus others (e.g. bounds on A etc.) Important progress in solving problems (C k ) and (R β ). The numerical schemes common points. = Explore the relationship between their optimal sets. 4
5 The optimal values / the optimal solution setsof problems (C k ) and (R β ): (C k ) c k := inf { Au d 2 u R N and u 0 k } Ĉ k := { u R N and u 0 k Au d 2 = c k } (R β ) r β := inf { F β (u) u R N} R β := { u R N F β (u) = r β } Theorem 1 For any d R M : Ĉ k k and Rβ β > 0. H 1 Assumption : rank(a) = M < N no further reminder How to evaluate the extent of assumption dependent properties? Definition 1 A property is generic on R M if it holds on a subset of R M \ S where S is closed in R M and its Lebesgue measure in R M is null. A generic property is stronger than a property that holds only with probability one. I n := ( {1,..., n}, < ) and I 0 n := ( {0, 1,..., n}, < ) (totally strictly ordered) L := min { k I N c k = 0 } (uniquely defined) generically L = M 5
6 Main results There is a strictly decreasing sequence {β k } k J {β Jk } for J I L such that û is global minimizer of F β for β ( β Jk, β Jk 1 ) û is global minimizer of (C Jk ) Equivalently { Rβ β ( β Jk, β Jk 1 ) } = ĈJ k k J In a generic sense R βjk = ĈJ k ĈJ k+1 All β Jk s are obtained from the optimal values c k s of the problems (C k ), k I 0 L. The global minimizers of problems (C k ) and (R β ) are strict and generically uniques J is always nonempty For any n I 0 L \ J the global minimizers of (C n) are not global minimizers of (R β ) β When J = I 0 L, problems (C k) and (R β ) are quasi-completely equivalent: { Rβ β (β k, β k 1 )} = Ĉk β k = c k c k+1 k 6
7 Notation. :=. 2. { } supp (u) := i I N : u[i] 0 For any ω I 0 N A ω := ( A ω1,..., A ω ω ) R M ω, A T ω is the transposed of A ω u ω := ( u ω1,..., u ω ω ) T R ω Definition 2 Let f : R N R and S R N. Consider the problem min {f(u) u S}. û is a strict minimizer if there is a neighborhood O S, û O so that f(u)>f(û) u O\ {û}. û is an isolated (local) minimizer if û is the only minimizer in an open subset O O 7
8 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) Preliminaries On the global minimizers of problem (C k ) Necessary and sufficient conditions 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 8
9 2. Common optimality conditions for (C k ) and (R β ) Goal: Derive tests relating the optimal solutions of (C k ) and (R β ). 2.1 Preliminaries A constrained quadratic optimization problem: given d R M and ω I N (P ω ) min u R N Au d 2 subject to u[i] = 0 i I 0 N \ ω The convex problem (P ω ) always has solutions, for any ω I 0 N and for any d RM. Some useful facts on the relation of (P ω ) to (C k ) and (R β ) [M. N. SIIMS 2013] (R β ) û solves (P ω ) for some ω I 0 N û is a (local) minimizer of F β, β > 0 û solves (P ω ) for ω I 0 N with rank(a ω ) = ω û = strict (local) minimizer of F β, β (C k ) û solves (P ω ) for ω I 0 N with ω = k û is a (local) minimizer of (C k) Remark 1 For any ω I N with rank(a ω ) = ω, the minimizer û of (P ω ) is isolated. 9
10 2.2 On the optimal solution sets of problem (C k ) Lemma 1 c 0 = d 2 and {c k } k 0 is decreasing with c k = 0 k M. Lemma 2 For k I M let (C k ) have a global minimizer û obeying û 0 = k n for n 1. Then Aû = d. Furthermore û Ĉm and c m = 0 m k n. L := min { k I M c k = 0 }. Example 1 One has L M 1 if d = Au for u 0 M 1. Theorem 2 û Ĉk for k I 0 L = k L + 1 = ĈL Ĉk. û 0 = k = rank (A σ ) for σ := supp (û) so û is a strict global minimizer of (C k ). Corollary 1 Ĉk Ĉn = (k, n) ( I 0 L )2 such that k n. 10
11 Example 2 A = and d = 1 1 û = (0, 0, 1) T = Ĉ1 (strict, rank(a 3 ) = û 0 ) = c 1 = 0 = L = 1. û = (1, 1, 0) T is a strict global minimizer of (C 2 ) because rank ( ) A supp (û) = 2 and c2 = 0. A = and d = 1 0 Ĉ 1 = { (1, 0, 0, 0) T, (0, 0, 1, 0) T} (strict minimizers) = c 1 = 0 = L = 1. For k 2 all optimal solutions Ĉ1 are nonstrict and have the form û = (x, y, 1 x, y) T, x R \ {0, 1}. If y = 0 then û 0 = 2 and otherwise û 0 = 4. Remark 2 By Theorem 2 the optimal value c k of problem (C k ) for any k I 0 L obeys { } c k = min Aũ d 2 where ũ R N solves (P ω ) ω Ω k where Ω k := { ω I N ω = k = rank (A ω ) }. 11
12 2.3. Necessary and sufficient conditions Proposition 1 û R û β = Ĉk where k := û 0 I 0 L Ĉ k R β for k := û 0 I 0 L The global minimizers of F β are composed of some optimal sets Ĉk for k L. Ĉ = the collection of all optimal solutions Ĉk of problems (C k ) for all k I 0 L; R = the set of all global minimizers R β of F β for all β > 0 Ĉ := L Ĉ k and R := k=0 β>0 R β. Theorem 3 R Ĉ. When β ranges on (0, + ), F β can have at most L + 1 different sets of global minimizers which are optimal solutions of (C k ) for k {0,..., L}. Theorem 4 For any k I 0 L one has: Ĉk R β if and only if F β (u) F β (û) 0 û Ĉk u Ĉ Ĉk = R β if and only if F β (u) F β (û) > 0 û Ĉk u Ĉ \ Ĉk 12
13 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) The entire list of parameter values Conditions for agreement between their optimal sets The effective parameters values 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 13
14 3. Parameter values for equality between optimal sets Definition 3 (Critical parameter values) { ck c k+n β k := max n { ck n c βk U k := min n 3.1. The entire list of parameter values } n {1,..., L k} } n {1,..., k} k I 0 L 1 and β L = 0, k I L and β U 0 β 1 := +. We have β L = 0 < βl U and β 0 < β0 U. The cases where β k < βk U will be of particular interest. Proposition 2 S finite union of vector subspaces of dimension M 1 such that d R M \ S = β k βk U k I 0 L. β k β U k k I 0 L is a generic property. 14
15 3.2. Conditions for agreement between the optimal sets of (C k ) and (R β ) Theorem 5 k I 0 L Ĉ k R β if and only if β 0 β < β U 0 for k = 0 ; β k β β U k for k {1,..., L 1} ; β L < β β U L for k = L. Ĉ k = R β if and only if β k < β < β U k. Proof based on Theorem 4. To exploit Theorem 5 we have to clarify the links between (β k, β U k ) and c k 15
16 3.3. The effective parameters values The global minimizers of F β are always in Ĉ (Theorem 3), so we are interested in the indexes k for which there exist values of β such that Ĉk R β. Their set is obtained from Theorem 5. Definition 4 The effective index set J J E : J := { k I 0 } L β k < βk U and J E := { m I 0 L β m = β U m}. Ordering: J = {J 0, J 1,..., J p } where p := J 1 and J k 1 < J k k. Further: (J 0 = 0, J p = L) J 2 with β J 1 := β U J 0 β U 0 = + and β Jp β L = 0. The set J is always nonempty. Lemma 3 R Ĉ k = if and only if k I 0 L \ {J J E }. 16
17 Definition 3 for reminder: { ck c k+n β k := max n { βk U ck n c k := min n } n {1,..., L k} k I 0 L 1 and β L := 0, } n {1,..., k} k I L and β0 U := +. Simplification of {β k, βk U} k J J E Proposition 3 Let {β k, βk U } and J be as in Definition 3 and Definition 4, resp. Then (a) β Jk < βj U k = β Jk 1 J k J \ {J 0 } and β J U 0 β J 1 = +. (b) β Jk = c J k c Jk+1 J k+1 J k J k J \ {J p } and β Jp β L = 0. (c) { β m m J E } { β Jk J k J \ {J p } }. {β k } k J is strictly decreasing and its first entry is β 0. 17
18 Example 3 Let {c k } L k=0 for L = 7 reads as c 0 = 48 c 1 = 40 c 2 = 30 c 3 = 22 c 4 = 14 c 5 = 10 c 6 = 4 c 7 = 0. By Definition 3 the sequences {β k, β U k } 7 k=0 are given by β 0 = 9 β 1 = 10 β 2 = 8 β 3 = 8 β 4 = 5 β 5 = 6 β 6 = 4 β 7 = 0 β U 0 = + β U 1 = 8 β U 2 = 9 β U 3 = 8 β U 4 = 8 β U 5 = 4 β U 6 = 5 β U 7 = 4 From Definition 4, J = { J 0 = 0, J 1 = 2, J 2 = 4, J 3 = 6, J 4 = 7} and J E = { 3 }. One has β Jk = β U J k+1 for any J k J (Proposition 3(a)). The formula in Proposition 3(b) holds. {β 3 3 J E } β 3 =β J1 =8 {β 3 3 J E } {β Jk J k J \ {J 4 }} (Proposition 3(c)). J E β J1 := {m J E β m = β J1 } = {3 J E J 1 < 3 < J 2 }, see Lemma 4. J has the smallest indexes so that {β k } k J = {9, 8, 5, 4, 0} is the longest strictly decreasing subsequence of {β k } 7 k=0 containing β 0 see Proposition 4. One has {β k } k J = {β k } k J for J := {0, 3, 4, 6, 7}; however, J 2 > J 2. 18
19 The location of {β m m J E } is given by the (probably empty) subsets J E β Jk := {m J E β m = β Jk }. Lemma 4 The sets J E β Jk fulfill J E β Jk = for k = p and for any k p 1 J E β Jk = {m J E J k < m < J k+1 }. J and {β k } k J are characterized next Proposition 4 Let {β k } L k=0 read as in Definition 3 and J as in Definition 4. Then 0 J and J contains the smallest indexes such that {β k } k J is the longest strictly decreasing subsequence of {β k } L k=0 containing β 0. In order to find the effective J and {β k } k J we need only {β k } L k=0 in Definition 3. 19
20 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 20
21 4. Equivalence relations between the optimal sets of (C k ) and (R β ) 4.1. Partial equivalence Theorem 6 Let {β k } be as in Definition 3 and J as in Definition 4. Then: { Rβ β ( β Jk, β Jk 1 ) } = ĈJ k J k J, ( p [ βjk, β Jk 1 ] ) [ β J0, β J 1 ) = [ 0, + ). n=1 { } βj0,..., β Jp 1 in Definition 4 partition (0, + ) into J proper intervals. For any β ( ) β Jk, β Jk 1 the optimal sets of (Rβ ) and of (C n ) for n = J k coincide. If I 0 L \ J, the optimal sets (C k ) for k I 0 L \ J cannot be optimal solutions of (R β ) β > 0. = partial equivalence. {β Jk } is a finite set of isolated values hence β β Jk generically. 21
22 The optimal sets of problem (R β ) for β k, k J: Theorem 7 Let H1 hold. Let {β k } be as in Definition 3 and J as in Definition 4. Then R βjk = ĈJ k ĈJ k+1 ( m J E β Jk Ĉ m ) J k J \ {J p }, where J E β Jk = {m J E J k < m < J k+1 } and Ĉk Ĉn = (k, n) (J J E ) 2, k n. Example 4 [Ex.3, cont.] We had J={0, 2, 4, 6, 7}, (β J0 =9, β J1 =8, β J2 =5, β J3 =4, β J4 =0), and J E 2 = {3} with β J E 2 =8 and J E k = otherwise. By Theorems 6 and 7 { R β β >9}=Ĉ0 { R β β (8, 9)}=Ĉ2 { R β β (5, 8)}=Ĉ4 { R β β (4, 5)}=Ĉ6 { R β β (0, 4)}=Ĉ7 and Rβ=9 = Ĉ0 Ĉ2 Rβ=8 = Ĉ2 Ĉ3 Ĉ4 Rβ=5 = Ĉ4 Ĉ6 Rβ=4 = Ĉ6 Ĉ7. A partial equivalence between problems (C k ) and (R β ) always exists. For the J 1 isolated values { β k k J \ {L} } problem (R β ) has normally two optimal sets (Proposition 2). 22
23 4.2. Quasi-complete equivalence Lemma 5 Let J be as in Definition 4. Then the following hold: (a) If the sequence {β k } L k=0 in Definition 3 is strictly decreasing, then its entries read as β k = c k c k+1 k I 0 L 1 and β L = 0, β 1 := β0 U = +. (1) (b) If the sequence {β k } L k=0 in (1) is strictly decreasing then J = I0 L. Theorem 8 Let {β k } L k=0 in (1) be strictly decreasing. Then { } Rβ β ( β k, β k 1 ) = Ĉk k I 0 L R βk = Ĉk Ĉk+1 with Ĉ k Ĉk+1 = k I 0 L 1. {β k } L k=0 in (1) strictly decreasing means that c k 1 c k > c k c k+1, k I L 1. Let u, û and ũ be optimal solutions of problems (C k 1 ), (C k ) and (C k+1 ), resp. Denote σ := supp (u), û := supp (û) and σ := supp (ũ). The condition on c k s reds as d T (Proj(A σ ) Proj(A σ )) d > d T (Proj(A σ ) Proj(A σ )) d > 0. Mid-way scenarios... 23
24 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 24
25 5. On the optimal values of (C k ) and (R β ) { } Ω k := ω I N ω = k = rank (A ω ). E k := range (A ω ) and G k := ω Ω k E 0 = G M = R M and E M = G 0 = {0} by H1. range (A ω ). ω Ω k Proposition 5 Let L M be arbitrarily fixed. c k > 0 k L 1 d R M \ G L 1 ; d R M \ (E 2 G L 1) = c k 1 > c k k I L. E 2 and G M 1 are finite unions of vector subspaces of dimensions M 2 and M 1, respectively. Hence, d R M \ (E 2 G M 1 ) is a generic property. = {c k } M k=0 is strictly decreasing and L = M generically. 25
26 Proposition 6 k I 0 L = F β (û) = c k + β k û Ĉk. By Theorem 3 R β Ĉ = L k=0 Ĉk = the optimal value of problem (R β ) reads as r β = min { c k + β k k I 0 L }. Corollary 2 The application β r β r β = c Jk + β J k : (0, + ) R fulfills = F β (û) û ĈJ k if and only if β β r β is continuous and concave. [β J0, + ) for J 0 0 [ ] βjk, β Jk 1 for Jk J \ {0, L} ( 0, βjp 1] for J p L r βjk 1 > r βjk J k J, r βj0 = c J0 = r β β β J0 and r βj0 > r β β < β J0. β r β is affine increasing on each interval ( β Jk, β Jk 1 ) with upward kinks at βjk J k J \ {L} and bounded by c 0. for any 26
27 Example 5 [Cont. of Example 3] From Corollary 2, β r β is given by β (0, 4] r β = c 7 + 7β = 7β r β=4 = 28 β [4, 5] r β = c 6 + 6β = 4 + 6β r β=5 = 34 β [5, 8] r β = c 4 + 4β = β r β=8 = 46 β [8, 9] r β = c 2 + 2β = β r β=9 = 48 β [9, + ) r β = c 0 + 0β = 48 affine expressions 27
28 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 28
29 6. Cardinality of the optimal sets of (C k ) and of (R β ) For any β > 0 and k I 0 L the optimal sets of problems (C k ) and (R β ) are composed out of a certain finite number of isolated (hence strict) minimizers Uniqueness of the optimal solutions of (C k ) and of (R β ) k min{l, M 1} and (û, ũ) (Ĉk) 2, û ũ. Then σ := supp (û), σ := supp (ũ) are in (Ω k ) 2. c k = A σ û σ d 2 = A σ ũ σ d 2 where σ σ. Π ω the orthogonal projector onto range (A ω ) A σ û σ d 2 A σ ũ σ d 2 = d T (Π σ Π σ ) d = 0. H For K min{m 1, L} fixed, A R M N obeys Π ω Π ω (ω, ω) Ω 2 k ω ω k I K. H is a generic property of all matrices in R M N [M. N., SIIMS 2013]. K := K k=1 dim( K ) M 1, hence d R M \ K generically. (ω,ω) (Ω k ) 2 { g R M ω ω and g ker (Π ω Π ω ) } H and d R M \ K = (C k ) for k I K has a unique optimal solution. K := max {k J k K} (R β ) has a unique global minimizer β (β K, + ) \ {β k } k J 29
30 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests Monte Carlo on {β k } with 10 5 tests for (M, N) = (5, 10) Tests on (partial) equivalence with selected matrix data 8. Conclusions and future directions 30
31 7. Numerical tests Two kind of experiments using matrices A R M N for (M, N) = (5, 10), original vectors u o R N and data samples d = Au o (+noise) with two different goals: to roughly see the behaviour of the parameters β k in Definition 3 ; to verify and illustrate our theoretical findings. All results were calculated using an exhaustive combinatorial search Monte Carlo experiments on {β k } with 10 5 tests for (M, N) = (5, 10) Two experiments, each one composed of 10 5 trials with A R M N for (M, N) = (5, 10) In each trial: an original u o R N, random support on {1,..., N} with u o 0 M 1 = 4. the coefficients of A and u o supp (u o ) i.i.d. d = Au o + i.i.d. centered Gaussian noise. compute the optimal values {c k } and then compute (β k, βk U ) by Definition 3. 31
32 Two different distributions for A and u o supp (u o ) Experiment N (0, 10). A(i, j) and u o supp (u o ) N (0, 10). Support length supp (u o ) {1,..., 4}, mean = 3.8. SNR of d in [10.1, 61.1], mean = db. Experiment Uni [0, 10]. A(i, j) and u o supp (u o ) uniform on [0, 10]. supp (u o ) {1,..., 4}, mean = 3.8. SNR of d in [20, 55], mean = db. N k := { k I 0 M β k > β k 1 }. Table 1: Results on the behaviour of {β k } in Definition 3 for two experiments, each one composed of 10 5 random trials. For k 3 we have found N k = 0. β k < β k 1, k I 0 M N k = 1 N k = 2 mean(snr) mean( u o 0 ) N (0, 10) % % % Uni [0, 10] % % % Observations: L = M in each trial (Remainder: L := min { k I N c k = 0 } ); {c k } M k=0 was always strictly decreasing (see Proposition 5); β k β U k in each trial (see Proposition 3), so J E = ; For every A there were d so that {β k := c k c k 1 } L k=0 was strictly decreasing 32
33 6.2. Tests on quasi-equivalence with a selected matrix and selected data A = ( ) u o T = A(i, j) nearly normal distribution with variance 10 and rank(a) = M = 5 Problem (C M ) has Ω M = 252 optimal solutions; { none } of them is shown. Since β 0 < β0 U = +, in all cases Ĉ0 = Rβ β > β 0 by Theorem 5. We selected a couple (A, u o ) so that β k are seldom strictly decreasing compared to Tab. 1. Table 2: 10 5 trials where d = Au o + i.i.d. centered Gaussian noise. β k < β k 1, k I 0 M N k = 1 N k = 2 mean(snr) mean( u o 0 ) u o in (2) % % 0 %
34 Noise-free data d = Au o = ( ) T. û = u o is an optimal solution to problems (C k ) with c k = 0 for k {3, 4, 5} and L = 3. β 3 = 0 < β U 3 = β 1 = < β U 1 = β 0 = and β 2 = 3968 > β U 2 = J = {0, 1, 3} k c k Ĉ k = the optimal solution of (C k ), singleton Ĉ k = R β β (β 3, β 1 ) no β (β 1, β 0 ) β > β 0 34
35 Noisy data 1. Nearly normal, centered, i.i.d. noise and SNR= db: ( ) T d = β 5 = 0 < β U 5 = β 4 = < β U 4 = β 3 = < β U 3 = β 1 = < β U 1 = β 0 = 63154, while β 2 = > β U 2 = Hence, J = I 0 5 \ {2} k c k Ĉ k = the optimal solution of (C k ), singleton Ĉ k = R β β (β 4, β 3 ) β (β 3, β 1 ) no β (β 1, β 0 ) β > β 0 35
36 Noisy data 2. The noise is nearly normal, centered, i.i.d., SNR= db: ( ) T d = β 0 = β 1 = 3825 β 2 = β 3 = β 4 = β 5 = 0. {β k } is strictly decreasing and hence β k = c k c k 1 (C k ) and (R β ) are quasi-completely equivalent. k c k Ĉ k = the optimal solution of (C k ), singleton Ĉ k = R β β (β 4, β 3 ) β (β 3, β 2 ) β (β 2, β 1 ) β (β 1, β 0 ) β > β 0 36
37 Outline 1. Two optimization problems involving the l 0 pseudo norm 2. Joint optimality conditions for (C k ) and (R β ) 3. Parameter values for equality between optimal sets of (C k ) and (R β ) 4. Equivalences between the global minimizers of (C k ) and (R β ) Partial equivalence Quasi-complete equivalence 5. On the optimal values of (C k ) and (R β ) 6. Cardinality of the optimal sets of (C k ) and of (R β ) Uniqueness of the optimal solutions of (C k ) and of (R β ) 7. Numerical tests 8. Conclusions and future directions 37
38 8. Conclusions and open questions The main equivalence result in a nutshell: Ĉ Jk ĈJ k+1 Ĉ0 J k }{{} R β = Ĉ L Ĉ Jk+1 Ĉ Jk Ĉ 0 0 = β L <... < β Jk+1 < β Jk < β Jk 1 <... < β J0 < The agreement between the optimal sets of problems (C k ) and (R β ) is driven by the critical parameters {β k } L k=0 which depend only on the optimal values c k of problem (C k ). Our comparative results clarify a proper choice between models (C k ) and (R β ) in applications. If one needs solutions with a fixed number of nonzero entries, (C k ) is the best choice. If only information on the perturbations is available, (R β ) is a more flexible model. If one can solve problem (C k ) for all k, the global minimizers of problem (R β ) are immediate. Our detailed results can give rise to innovative algorithms. The degree of partial equivalence depends on the distribution of the coefficients of A and d. 38
39 By specifying a class of matrices A and assumptions on data d, one can infer statistical knowledge on the optimal values c k of problems (C k ) and thus on the critical parameters {β k }. Promising theoretical and practical results can be expected. A related open question is to know if the optimal solutions of (R β ) are able to eliminate some meaningless solutions of (C k ). Extensions to analysis type penalties Du 0, to low rank matrix recovery, etc., are important. Other important results concern algorithms that are known to converge to local minimizers. Remark 3 Problem (R β ) (for some β > 0) and problems (C k ) for k {0, 1,..., M} have the same sets of (strict) local minimizers. 39
40 Thank you for the attention. Thanks to Zuhair Nashed for the invitation 40
Relationship between the optimal solutions of least squares regularized with l 0 -norm and constrained by k-sparsity
Applied and Computational Harmonic Analysis, Volume 41, Issue 1, July 2016, pages 237 265 http://dx.doi.org/10.1016/j.acha.2015.10.010 Relationship between the optimal solutions of least squares regularized
More informationDescription of the minimizers of least squares regularized with l_0-norm. Uniqueness of the global minimizer
Description of the minimizers of least squares regularized with l_0-norm. Uniqueness of the global minimizer Mila Nikolova To cite this version: Mila Nikolova. Description of the minimizers of least squares
More informationConstrained optimization
Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained
More informationLeast Sparsity of p-norm based Optimization Problems with p > 1
Least Sparsity of p-norm based Optimization Problems with p > Jinglai Shen and Seyedahmad Mousavi Original version: July, 07; Revision: February, 08 Abstract Motivated by l p -optimization arising from
More informationContre-examples for Bayesian MAP restoration. Mila Nikolova
Contre-examples for Bayesian MAP restoration Mila Nikolova CMLA ENS de Cachan, 61 av. du Président Wilson, 94235 Cachan cedex (nikolova@cmla.ens-cachan.fr) Obergurgl, September 26 Outline 1. MAP estimators
More informationVector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.
Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationModel-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk
Model-Based Compressive Sensing for Signal Ensembles Marco F. Duarte Volkan Cevher Richard G. Baraniuk Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional
More informationThe Analysis Cosparse Model for Signals and Images
The Analysis Cosparse Model for Signals and Images Raja Giryes Computer Science Department, Technion. The research leading to these results has received funding from the European Research Council under
More informationINDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina
INDUSTRIAL MATHEMATICS INSTITUTE 2007:08 A remark on compressed sensing B.S. Kashin and V.N. Temlyakov IMI Preprint Series Department of Mathematics University of South Carolina A remark on compressed
More informationLecture Notes 9: Constrained Optimization
Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form
More informationLinear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions
Linear Systems Carlo Tomasi June, 08 Section characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix
More informationThe Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1
The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1 Simon Foucart Department of Mathematics Vanderbilt University Nashville, TN 3784. Ming-Jun Lai Department of Mathematics,
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationUniqueness Conditions for A Class of l 0 -Minimization Problems
Uniqueness Conditions for A Class of l 0 -Minimization Problems Chunlei Xu and Yun-Bin Zhao October, 03, Revised January 04 Abstract. We consider a class of l 0 -minimization problems, which is to search
More informationAn Introduction to Sparse Approximation
An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,
More informationIEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER 2011 7255 On the Performance of Sparse Recovery Via `p-minimization (0 p 1) Meng Wang, Student Member, IEEE, Weiyu Xu, and Ao Tang, Senior
More informationANALYTICAL BOUNDS ON THE MINIMIZERS OF (NONCONVEX) REGULARIZED LEAST-SQUARES. Mila Nikolova. (Communicated by Jackie (Jianhong) Shen)
Inverse Problems and Imaging Volume 1, No. 4, 2007, 661 677 Web site: http://www.aimsciences.org ANALYTICAL BOUNDS ON THE MINIMIZERS OF NONCONVEX) REGULARIZED LEAST-SQUARES Mila Nikolova CMLA, ENS Cachan,
More informationNecessary and Sufficient Conditions of Solution Uniqueness in 1-Norm Minimization
Noname manuscript No. (will be inserted by the editor) Necessary and Sufficient Conditions of Solution Uniqueness in 1-Norm Minimization Hui Zhang Wotao Yin Lizhi Cheng Received: / Accepted: Abstract This
More informationACCORDING to Shannon s sampling theorem, an analog
554 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL 59, NO 2, FEBRUARY 2011 Segmented Compressed Sampling for Analog-to-Information Conversion: Method and Performance Analysis Omid Taheri, Student Member,
More informationEquivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 19, NO. 12, DECEMBER 2008 2009 Equivalence Probability and Sparsity of Two Sparse Solutions in Sparse Representation Yuanqing Li, Member, IEEE, Andrzej Cichocki,
More informationSPARSE signal representations have gained popularity in recent
6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying
More informationLecture: Introduction to Compressed Sensing Sparse Recovery Guarantees
Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html Acknowledgement: this slides is based on Prof. Emmanuel Candes and Prof. Wotao Yin
More informationEUSIPCO
EUSIPCO 013 1569746769 SUBSET PURSUIT FOR ANALYSIS DICTIONARY LEARNING Ye Zhang 1,, Haolong Wang 1, Tenglong Yu 1, Wenwu Wang 1 Department of Electronic and Information Engineering, Nanchang University,
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course
More informationLinear Systems. Carlo Tomasi
Linear Systems Carlo Tomasi Section 1 characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix and of
More informationConstructing Explicit RIP Matrices and the Square-Root Bottleneck
Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry
More informationSelf-Calibration and Biconvex Compressive Sensing
Self-Calibration and Biconvex Compressive Sensing Shuyang Ling Department of Mathematics, UC Davis July 12, 2017 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 1 / 22 Acknowledgements
More informationThe definition of a vector space (V, +, )
The definition of a vector space (V, +, ) 1. For any u and v in V, u + v is also in V. 2. For any u and v in V, u + v = v + u. 3. For any u, v, w in V, u + ( v + w) = ( u + v) + w. 4. There is an element
More informationNear-Potential Games: Geometry and Dynamics
Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo January 29, 2012 Abstract Potential games are a special class of games for which many adaptive user dynamics
More informationON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION
ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION CHRISTIAN GÜNTHER AND CHRISTIANE TAMMER Abstract. In this paper, we consider multi-objective optimization problems involving not necessarily
More informationGROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA
GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA D Pimentel-Alarcón 1, L Balzano 2, R Marcia 3, R Nowak 1, R Willett 1 1 University of Wisconsin - Madison, 2 University of Michigan - Ann Arbor, 3 University
More informationNear Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing
Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar
More informationDS-GA 1002 Lecture notes 12 Fall Linear regression
DS-GA Lecture notes 1 Fall 16 1 Linear models Linear regression In statistics, regression consists of learning a function relating a certain quantity of interest y, the response or dependent variable,
More informationTHEORY OF COMPRESSIVE SENSING VIA l 1 -MINIMIZATION: A NON-RIP ANALYSIS AND EXTENSIONS
THEORY OF COMPRESSIVE SENSING VIA l 1 -MINIMIZATION: A NON-RIP ANALYSIS AND EXTENSIONS YIN ZHANG Abstract. Compressive sensing (CS) is an emerging methodology in computational signal processing that has
More informationOrthogonal Matching Pursuit for Sparse Signal Recovery With Noise
Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published
More informationConditions for Robust Principal Component Analysis
Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and
More informationEnergy minimization methods
Handbook of Mathematical Methods in Imaging, Springer 211, 1 st edition Editor: Otmar SCHERZER Chapter 5, p. 138 186 Energy minimization methods Mila NIKOLOVA CMLA, ENS Cachan, CNRS, UniverSud, 61 Av.
More informationInverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France.
Inverse problems and sparse models (6/6) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France remi.gribonval@inria.fr Overview of the course Introduction sparsity & data compression inverse problems
More informationRecovery of Low Rank and Jointly Sparse. Matrices with Two Sampling Matrices
Recovery of Low Rank and Jointly Sparse 1 Matrices with Two Sampling Matrices Sampurna Biswas, Hema K. Achanta, Mathews Jacob, Soura Dasgupta, and Raghuraman Mudumbai Abstract We provide a two-step approach
More informationDecomposition of Riesz frames and wavelets into a finite union of linearly independent sets
Decomposition of Riesz frames and wavelets into a finite union of linearly independent sets Ole Christensen, Alexander M. Lindner Abstract We characterize Riesz frames and prove that every Riesz frame
More informationMATH 8253 ALGEBRAIC GEOMETRY WEEK 12
MATH 8253 ALGEBRAIC GEOMETRY WEEK 2 CİHAN BAHRAN 3.2.. Let Y be a Noetherian scheme. Show that any Y -scheme X of finite type is Noetherian. Moreover, if Y is of finite dimension, then so is X. Write f
More informationRSP-Based Analysis for Sparsest and Least l 1 -Norm Solutions to Underdetermined Linear Systems
1 RSP-Based Analysis for Sparsest and Least l 1 -Norm Solutions to Underdetermined Linear Systems Yun-Bin Zhao IEEE member Abstract Recently, the worse-case analysis, probabilistic analysis and empirical
More informationarxiv: v1 [math.na] 26 Nov 2009
Non-convexly constrained linear inverse problems arxiv:0911.5098v1 [math.na] 26 Nov 2009 Thomas Blumensath Applied Mathematics, School of Mathematics, University of Southampton, University Road, Southampton,
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationOptimization for Compressed Sensing
Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve
More informationRecovery Guarantees for Rank Aware Pursuits
BLANCHARD AND DAVIES: RECOVERY GUARANTEES FOR RANK AWARE PURSUITS 1 Recovery Guarantees for Rank Aware Pursuits Jeffrey D. Blanchard and Mike E. Davies Abstract This paper considers sufficient conditions
More informationStrengthened Sobolev inequalities for a random subspace of functions
Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)
More informationA Continuation Approach to Estimate a Solution Path of Mixed L2-L0 Minimization Problems
A Continuation Approach to Estimate a Solution Path of Mixed L2-L Minimization Problems Junbo Duan, Charles Soussen, David Brie, Jérôme Idier Centre de Recherche en Automatique de Nancy Nancy-University,
More informationSparse Optimization Lecture: Sparse Recovery Guarantees
Those who complete this lecture will know Sparse Optimization Lecture: Sparse Recovery Guarantees Sparse Optimization Lecture: Sparse Recovery Guarantees Instructor: Wotao Yin Department of Mathematics,
More informationNew Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit
New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence
More informationRecent Developments in Compressed Sensing
Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline
More informationSparse solutions of underdetermined systems
Sparse solutions of underdetermined systems I-Liang Chern September 22, 2016 1 / 16 Outline Sparsity and Compressibility: the concept for measuring sparsity and compressibility of data Minimum measurements
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57
More informationLinear Methods for Regression. Lijun Zhang
Linear Methods for Regression Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Linear Regression Models and Least Squares Subset Selection Shrinkage Methods Methods Using Derived
More informationMatrix Support Functional and its Applications
Matrix Support Functional and its Applications James V Burke Mathematics, University of Washington Joint work with Yuan Gao (UW) and Tim Hoheisel (McGill), CORS, Banff 2016 June 1, 2016 Connections What
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationBISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption
BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption Daniel Reem (joint work with Simeon Reich and Alvaro De Pierro) Department of Mathematics, The Technion,
More informationProblem set 5: SVD, Orthogonal projections, etc.
Problem set 5: SVD, Orthogonal projections, etc. February 21, 2017 1 SVD 1. Work out again the SVD theorem done in the class: If A is a real m n matrix then here exist orthogonal matrices such that where
More informationSignal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit
Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit Deanna Needell and Roman Vershynin Abstract We demonstrate a simple greedy algorithm that can reliably
More informationAnalysis of Greedy Algorithms
Analysis of Greedy Algorithms Jiahui Shen Florida State University Oct.26th Outline Introduction Regularity condition Analysis on orthogonal matching pursuit Analysis on forward-backward greedy algorithm
More informationNear-Potential Games: Geometry and Dynamics
Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo September 6, 2011 Abstract Potential games are a special class of games for which many adaptive user dynamics
More informationIMA Preprint Series # 2066
THE CARDINALITY OF SETS OF k-independent VECTORS OVER FINITE FIELDS By S.B. Damelin G. Michalski and G.L. Mullen IMA Preprint Series # 2066 ( October 2005 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS
More informationChapter 1. Preliminaries
Introduction This dissertation is a reading of chapter 4 in part I of the book : Integer and Combinatorial Optimization by George L. Nemhauser & Laurence A. Wolsey. The chapter elaborates links between
More informationDetecting Sparse Structures in Data in Sub-Linear Time: A group testing approach
Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach Boaz Nadler The Weizmann Institute of Science Israel Joint works with Inbal Horev, Ronen Basri, Meirav Galun and Ery Arias-Castro
More informationA Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing
A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing Emilie Chouzenoux emilie.chouzenoux@univ-mlv.fr Université Paris-Est Lab. d Informatique Gaspard
More informationGeometry of log-concave Ensembles of random matrices
Geometry of log-concave Ensembles of random matrices Nicole Tomczak-Jaegermann Joint work with Radosław Adamczak, Rafał Latała, Alexander Litvak, Alain Pajor Cortona, June 2011 Nicole Tomczak-Jaegermann
More informationA NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang
A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES Fenghui Wang Department of Mathematics, Luoyang Normal University, Luoyang 470, P.R. China E-mail: wfenghui@63.com ABSTRACT.
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu
More informationA New Estimate of Restricted Isometry Constants for Sparse Solutions
A New Estimate of Restricted Isometry Constants for Sparse Solutions Ming-Jun Lai and Louis Y. Liu January 12, 211 Abstract We show that as long as the restricted isometry constant δ 2k < 1/2, there exist
More informationDiscussion of Hypothesis testing by convex optimization
Electronic Journal of Statistics Vol. 9 (2015) 1 6 ISSN: 1935-7524 DOI: 10.1214/15-EJS990 Discussion of Hypothesis testing by convex optimization Fabienne Comte, Céline Duval and Valentine Genon-Catalot
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationVector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then
1 Def: V be a set of elements with a binary operation + is defined. F be a field. A multiplication operator between a F and v V is also defined. The V is called a vector space over the field F if: V is
More informationSparsity Regularization
Sparsity Regularization Bangti Jin Course Inverse Problems & Imaging 1 / 41 Outline 1 Motivation: sparsity? 2 Mathematical preliminaries 3 l 1 solvers 2 / 41 problem setup finite-dimensional formulation
More informationAn algebraic perspective on integer sparse recovery
An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:
More informationThe Information-Theoretic Requirements of Subspace Clustering with Missing Data
The Information-Theoretic Requirements of Subspace Clustering with Missing Data Daniel L. Pimentel-Alarcón Robert D. Nowak University of Wisconsin - Madison, 53706 USA PIMENTELALAR@WISC.EDU NOWAK@ECE.WISC.EDU
More informationThree Generalizations of Compressed Sensing
Thomas Blumensath School of Mathematics The University of Southampton June, 2010 home prev next page Compressed Sensing and beyond y = Φx + e x R N or x C N x K is K-sparse and x x K 2 is small y R M or
More informationarxiv: v1 [math.oc] 22 Sep 2016
EUIVALENCE BETWEEN MINIMAL TIME AND MINIMAL NORM CONTROL PROBLEMS FOR THE HEAT EUATION SHULIN IN AND GENGSHENG WANG arxiv:1609.06860v1 [math.oc] 22 Sep 2016 Abstract. This paper presents the equivalence
More informationCSC 576: Variants of Sparse Learning
CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in
More informationThe Pros and Cons of Compressive Sensing
The Pros and Cons of Compressive Sensing Mark A. Davenport Stanford University Department of Statistics Compressive Sensing Replace samples with general linear measurements measurements sampled signal
More informationJoint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces
Aalborg Universitet Joint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces Christensen, Mads Græsbøll; Nielsen, Jesper Kjær Published in: I E E E / S P Workshop
More informationCompressed Sensing and Sparse Recovery
ELE 538B: Sparsity, Structure and Inference Compressed Sensing and Sparse Recovery Yuxin Chen Princeton University, Spring 217 Outline Restricted isometry property (RIP) A RIPless theory Compressed sensing
More informationRank Determination for Low-Rank Data Completion
Journal of Machine Learning Research 18 017) 1-9 Submitted 7/17; Revised 8/17; Published 9/17 Rank Determination for Low-Rank Data Completion Morteza Ashraphijuo Columbia University New York, NY 1007,
More informationInformation-Theoretic Limits of Matrix Completion
Information-Theoretic Limits of Matrix Completion Erwin Riegler, David Stotz, and Helmut Bölcskei Dept. IT & EE, ETH Zurich, Switzerland Email: {eriegler, dstotz, boelcskei}@nari.ee.ethz.ch Abstract We
More informationPCA with random noise. Van Ha Vu. Department of Mathematics Yale University
PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical
More informationDS-GA 1002 Lecture notes 10 November 23, Linear models
DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.
More informationSignal Recovery, Uncertainty Relations, and Minkowski Dimension
Signal Recovery, Uncertainty Relations, and Minkowski Dimension Helmut Bőlcskei ETH Zurich December 2013 Joint work with C. Aubel, P. Kuppinger, G. Pope, E. Riegler, D. Stotz, and C. Studer Aim of this
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationRestricted Strong Convexity Implies Weak Submodularity
Restricted Strong Convexity Implies Weak Submodularity Ethan R. Elenberg Rajiv Khanna Alexandros G. Dimakis Department of Electrical and Computer Engineering The University of Texas at Austin {elenberg,rajivak}@utexas.edu
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 3: Sparse signal recovery: A RIPless analysis of l 1 minimization Yuejie Chi The Ohio State University Page 1 Outline
More informationMATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.
MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationSupremum of simple stochastic processes
Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there
More informationA memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration
A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationCompressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery
Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad
More information