University of Luxembourg. Master in Mathematics. Student project. Compressed sensing. Supervisor: Prof. I. Nourdin. Author: Lucien May

Similar documents
DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

INNER PRODUCT SPACE. Definition 1

Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula

Conditions for Robust Principal Component Analysis

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

Methods for sparse analysis of high-dimensional data, II

There are two things that are particularly nice about the first basis

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Linear Algebra Massoud Malek

LINEAR ALGEBRA W W L CHEN

COMPRESSED Sensing (CS) is a method to recover a

1. General Vector Spaces

MTH 2032 SemesterII

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

Recall the convention that, for us, all vectors are column vectors.

Chapter 6: Orthogonality

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Methods for sparse analysis of high-dimensional data, II

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

Linear Algebra, Summer 2011, pt. 3

Chapter 3 Transformations

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

Mathematics Department Stanford University Math 61CM/DM Inner products

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

(II.B) Basis and dimension

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Math 113 Winter 2013 Prof. Church Midterm Solutions

Strengthened Sobolev inequalities for a random subspace of functions

The Multivariate Gaussian Distribution

Linear Algebra. Workbook

Supremum of simple stochastic processes

Optimization Theory. Linear Operators and Adjoints

6 Compressed Sensing and Sparse Recovery

Typical Problem: Compute.

Least Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo

Chapter 2: Linear Independence and Bases

Invertibility of random matrices

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

2. Matrix Algebra and Random Vectors

Math Linear Algebra II. 1. Inner Products and Norms

The Pros and Cons of Compressive Sensing

Z Algorithmic Superpower Randomization October 15th, Lecture 12

Linear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.

Math 61CM - Solutions to homework 6

Linear Models Review

Linear Algebra. Session 12

x 1 x 2. x 1, x 2,..., x n R. x n

THE SMALLEST SINGULAR VALUE OF A RANDOM RECTANGULAR MATRIX

DS-GA 1002 Lecture notes 10 November 23, Linear models

The following definition is fundamental.

2. Review of Linear Algebra

Intrinsic Volumes of Convex Cones Theory and Applications

Corrupted Sensing: Novel Guarantees for Separating Structured Signals

Math 407: Linear Optimization

Linear Algebra. Min Yan

Diagonalizing Matrices

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Recovery of Simultaneously Structured Models using Convex Optimization

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Compressed Sensing and Sparse Recovery

IEOR 265 Lecture 3 Sparse Linear Regression

Sparse and Low Rank Recovery via Null Space Properties

An algebraic perspective on integer sparse recovery

Exercise Solutions to Functional Analysis

PART II : Least-Squares Approximation

4.2. ORTHOGONALITY 161

Elementary linear algebra

A strongly polynomial algorithm for linear systems having a binary solution

Lecture Notes 1: Vector spaces

October 25, 2013 INNER PRODUCT SPACES

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Algebra II. Paulius Drungilas and Jonas Jankauskas

Assignment 1: From the Definition of Convexity to Helley Theorem

MATRICES. a m,1 a m,n A =

LINEAR ALGEBRA MICHAEL PENKAVA

j=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).

1 Regression with High Dimensional Data

CHEBYSHEV INEQUALITIES AND SELF-DUAL CONES

Signal Recovery from Permuted Observations

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

Conceptual Questions for Review

Systems of Linear Equations

Geometry of log-concave Ensembles of random matrices

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Lecture 1: Review of linear algebra

Vector Spaces. Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Overview. Motivation for the inner product. Question. Definition

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Uniqueness of the Solutions of Some Completion Problems

NORMS ON SPACE OF MATRICES

Sparse solutions of underdetermined systems

Least Sparsity of p-norm based Optimization Problems with p > 1

NEW FUNCTIONAL INEQUALITIES

Chapter 2. Vectors and Vector Spaces

B553 Lecture 5: Matrix Algebra Review

Transcription:

University of Luxembourg Master in Mathematics Student project Compressed sensing Author: Lucien May Supervisor: Prof. I. Nourdin Winter semester 2014

1 Introduction Let us consider an s-sparse vector x 0 in R n, that is a vector in R n with at most s non-zero entries. Suppose we have m n linear measurements of our vector x 0, i.e. we have a vector y in R m such that y = Ax 0, (1) where A Mat(R, m, n). An important question in signal processing is, whether we can reconstruct the vector x 0 out of the knowledge of y. If m < n, this linear system is under-determined and we cannot hope to solve this system uniquely. We must specify additional constraints on x 0 in order to reduce the number of interesting solutions to 1. We thus ask whether we can find an unique solution to the following optimization problem minimize max {k 1 i 1 <... < i k n : x ij 0} =: x 0 subject to Ax = Ax 0. The 0 pseudo-norm 1 controls the sparsity of the vector x. Though it is possible under suitable conditions on A to solve this problem, it is computationally untractable if n is large [2]. To circumvent this problem, it turns out that replacing the function 0 by 1, the l 1 -norm, leads to a computationally efficient algorithm. It is remarkable that the l 1 -norm is adequate to control the sparsity of x. In this project we consider that the matrix A is random, more precisely we suppose that the n m entries of A are independent and identically distributed normal random variable N (0, 1/m). The problem we shall discuss is thus the following minimize x 1 (2) subject to Ax = Ax 0. In this work we present in detail that for m 2βs log n + s where β > 1 is arbitrary but fixed, we can recover x in problem (2) with high probability. Problems of this kind fall in the domain of compressed sensing, which is currently a rapidly developing field. This presentation is based on the paper [1]. Then we illustrate the problem by computational experiments, which confirm the theoretical results. 1 It is not a norm because the triangle inequality is not satisfied. 1

2 The sparse vector recovery problem The aim will be to present the following result. Theorem 1. Let x 0 be an s-sparse vector in R n and β > 1. For a m n matrix A with normal N (0, 1 ) random variables as entries and m 2βs log n + s the m recovery of x 0 in problem (2) is exact with probability at least 1 2n f(β,s), where ( 2 β f(β, s) = + β 1 β 2s 2s). Note that f(β, s) is positive for all β > 1 and s > 0 The proof is based on special properties of the l 1 norm and on concentration of measure inequalities. Recall that the subdifferential f(x 0 ) of a convex function f : R n R at x 0 R n is defined as the set f(x 0 ) = {g R n x R n : f(x) f(x 0 ) + g, x x 0 }, where, is the usual inner product on R n. It is clear from the triangle inequality that the l 1 norm is a convex function on R n. The subdifferentials of 1 have a special structure. For x R n having at least one zero entry we have that there exists a subspace T R n and e T such that x 1 = {z R n : P T (z) = e and P T 1}, where P V designates the projection operator on V. We say in this case that the norm 1 is decomposable at x. Moreover we have that for w T w 1 = sup v, w. (3) v T, v 1 Let us illustrate this in the case of R 2. For x = (x 1, x 2 ) R we easily find the following subdifferentials. 2 sign(x 1 ) sign(x 2 ) (x 1, x 2 ) 1 T e 0 +1 {(x, 1) R 2 x 1} R(0, 1) (0, 1) 0-1 {(x, 1) R 2 x 1} R(0, 1) (0, 1) +1 0 {(1, x) R 2 x 1} R(1, 0) (1, 0) -1 0 {( 1, x) R 2 x 1} R( 1, 0) ( 1, 0) 0 0 {(x, y) R 2 x 1, y 1} {(0, 0)} (0, 0) 2 sign(x)=0 means that x = 0. 2

Formula (3) is easily seen to be valid. Similar types of computations can easily be done also in higher dimensions. If x 0 R n is an s-sparse vector, one sees that dim(t )= s, and e=sign(x 0 ) where sign( ) return the signs of the components. This shows that e 2 = s. (4) The following proposition gives a condition under which a unique solution to (2) exists. Proposition 1. Let x 0 R be a point at which the l 1 -norm is decomposable and T, e the corresponding subspace, respectively vector as previously described. Let A be a random matrix as in theorem 2.1. Suppose that A is one-to-one on the subspace T and that there exists y ima R n such that 1. P T (y) = e 2. P T (y) < 1. Then x 0 is the unique solution to the linear inverse problem (2). Proof. One has to check that if h R n is such that Ah = 0, then x 0 + h 1 x 0 1. Using formula (3) we see that there exists v T such that v 1 and v, P T (h) = P T (h) 1. Note that P T (e + v) = e and P T (e + v) = P T (v) 1, so that e + v x 0 1. Thus x 0 + h 1 x 0 1 + e + v, h = x 0 1 + e + v y, h = x 0 1 + v P T (y), h = x 0 1 + v P T (y), P T (h) x 0 1 + (1 P T (y) ) P T (h) 1. The first equality comes from the fact that y = A w, so that y, h = A w, h = w, Ah = 0. The second equality simply uses that P T (y) = e. The third equality comes from the fact that v P T (y) T. The last inequality comes from the fact that P T (y), P T (h) P T (h) 1 P T (y), where we have used (3). Since P T (y) < 1, we get that x 0 + h 1 > x 0 1 unless that P T (h) = 0. In the latter case we then have that h T. Since Ah = 0 and A is injective on T, we conclude that h = 0. This shows that x 0 is a minimizer of (2). Now observe that the previous arguments show that the minimizer x 0 is unique. 3

If the entries of A are independent and identically distributed centered gaussian random variables with variance 1/m, then it is clear that provided dimt =: d T m, A restricted to T is one-to-one. In order to find the vector y in proposition 1 we need a priori to solve the equation P T (y) = e, which is of course not uniquely solvable. The plan is to find the least square solution of the equation P T (A q) = e, (5) and then show that for y := A q we have P T (y) < 1 with high probability. This means that with high probability the problem (2) has a unique solution. Note that the map P T A is a linear map from R m R n, where m n. Let A T and A T denote the restriction of A to T and T,respectively. We then have the following proposition. Proposition 2. If A T is one-to-one, the least-square solution of (5) is given by q = A T (A T A T ) 1 e. (6) Moreover we have that P T (y) = A T q. (7) Proof. Note that by assumption A T A T is invertible, so that the MoorePenrose pseudoinverse of A T is given by A T (A T A T ) 1. The first formula and uniqueness then follows from the relation of the MoorePenrose pseudoinverse and the least square method, see for instance [4]. The following proposition will be crucial for the sequel. Proposition 3. Let A, A T and q as before. Then, for any T, A T and q are independent. Proof. By Cochran s theorem (see for example [5] theorem 5.17), if X N (0, I d ) and E 1, E 2 are two orthogonal spaces in R d, then the projections Π E1 X, Π E2 X are independent. Now note that q depends of A T, and the claim then follows. In order to see that the second condition in proposition 1 is verified with high probability for the y previously constructed, we need to know the distribution of P T (y). It is given by the following proposition. Proposition 4. Let A and q be as in proposition 2. Then, conditioned on q, P T (y) is distributed as ι T g,where ι T is an isometry from R n d T onto T and g N (0, q 2 2 m 4

Proof. Note that we have P T (y) = A T (q). Then the claim follows immediately from the fact that if X 1, X 2 are two independent normal random variables with distribution N (0, 1), then ax 1 + bx 2, a, b R, has distribution N (0, a 2 + b 2 ). In the following we will need to bound probabilities of the following type. P[ P T (y) 1] = P[ P T (y) 1 q 2 τ] P[ q 2 τ] +P[ P T (y) 1 q 2 τ] P[ q 2 τ] P[ P T (y) 1 q 2 τ] + P[ q 2 τ] (8) Here τ is some positive parameter to be specified later. Each of the probabilities in (8) will now be bounded. To bound the second probability on the right hand side of (8) we will use the following proposition. Proposition 5. We have that q 2 2 = e, (A T A T ) 1 e, and q 2 2 is distributed as e 2 2mB 11, where B 11 is the first entry in the first column of an inverse Wishart matrix with m degrees of freedom and covariance I d T. Proof. We first recall the distribution of a Wishart matrix. Suppose X is an n p matrix, each row of which is drawn independently from a p-variate random normal vector with distribution N (0, V ). Then the W ishart distribution with n degrees of freedom and covariance V is the probability distribution of the p p random matrix X T X. Note that (6) implies by the definition of the adjoint that q 2 2 = e, (A T A T ) 1 e. By what has been said before we have that (A T A T ) 1 is a d T d T inverse Wishart matrix with m degrees of freedom and covariance 1 m I d T. Now let P denote an orthogonal matrix of size n n such that P e = e 2 e 1, where e 1 = (1, 0,..., 0). We have then that the law of q 2 2 is equal in law to e, (A T A T ) 1 e = e, (P A T A T P ) 1 e = e, P (A T A T ) 1 P 1 e = P e, P (A T A T ) 1 P e = e 2 2 e1, (A T A T ) 1 e 1 = e 2 2mB 11, where we have used in the first equality that the Gaussian distribution is isotropic. 5

One will now show that B 11 is distributed as an inverse chi-squared random variable with m d T + 1 degrees of freedom. Proposition 6. Let m n. The first entry in the first column of an inverse Wishart matrix with m degrees of freedom and covariance I n is distributed as an inverse chi-squared random variable with m n + 1 degrees of freedom. Proof. We first show that the squared distance of a vector v R m to a subspace L R m spanned by the linearly independent vectors v 1,..., v n is given by where d 2 (v, L) = G(v 1,..., v n, v) G(v 1,..., v n ), x 1, x 1... x 1, x k G(x 1,..., x k ) = det......... x k, x 1... x k, x k is a Gram determinant. Note that (x 1,..., x k ) is a linear dependent family if and only if G(x 1,..., x k ) = 0. Indeed, if (x 1,..., x k ) is a linear dependent family, then there is (λ 1,..., λ k ) 0 such that k i=1 λ ix i = 0. Then k i=1 λ il i = 0, where L 1,..., L k are the lines of Gram matrix of x 1,..., x k. Conversely if G(x 1,..., x k ) = 0, then there is (λ 1,..., λ k ) 0 such that k i=1 λ il i = 0, where L 1,..., L k are the lines of the Gram matrix of x 1,..., x k as previously. Then we can see that k i=1 λ ix i = 0 is orthogonal to all x i, and is thus 0. Linear dependence follows. Now let v = u + k, where u L and k L. By using the multi-linearity of the determinant, we find that G(v 1,..., v n, v) = G(v 1,..., v n, u) + G(v 1,..., v n, k) v 1, v 1... v 1, v n 0 = G(v 1,..., v n, u) + det............ v n, v 1... v n, v n 0 k 2 = G(v 1,..., v n, u) + G(v 1,..., v n ) k 2 Now G(v 1,..., v n, u) = 0, because (v 1,..., v n, u) is a linear dependent family. So we get that u 2 = d 2 (v, L) = G(v 1,..., v n, v) G(v 1,..., v n ). 6

Let Φ be an m n matrix with independent and identically distributed standard normal random variables. Denote the columns of Φ by C 1 to C n. Since m n it is clear that the family (C 1,..., C n ) is linearly independent. Then it is easy to see that (Φ Φ) ij = C i, C j. Let us denote the elements of the inverse Wishart matrix (Φ Φ) 1 =: B by B ij (i, j = 1,..., n). Using the basic formulas for computing the inverse of matrices we find that C 2, C 2... C 2, C n det......... C p, C 2... C p, C n B 11 = C 1, C 1... C 1, C n det......... C n, C 1... C n, C n = G(C 2,..., C n ) G(C 1,..., C n ) By what we have seen previously we see that B 11 = 1 d(c 1, L) 2, where L is the vector space spanned by C 2,..., C n. Now d(c 1, L) 2 = Π L (C 1 ) 2, where Π L (C 1 ) denotes the orthogonal projection of C 1 to L, a subspace of R n of dimension m n 1 (because L is of dimension n 1). By Cochran s theorem (see for example theorem 5.17 in [5]) we have that Π L (C 1 ) 2 χ 2 (m n + 1). Then we can use the following concentration result. Proposition 7. Let (Y 1,..., Y D ) be independent and identically distributed standard normal random variables and let α 1,..., α D be nonnegative. We set α = sup i=1,...,d a i, α 2 2 = α 2 1 +... + α 2 D. and Z = D i=1 α i (Y 2 i 1). Then the following inequalities hold for all positive x: P[Z 2 α 2 x] e x. (9) 7

For a proof we refer to ([3]), lemma 1, section 4. Now let U be a random variable with a chi-squared distribution with D degrees of freedom. Then it follows readily from (9) that for any positive x P[D U 2 Dx] e x. (10) In our case for D = m d T + 1 and τ = m e m d T +1 t 2 3 in 8 we can deduce from (10) the following large deviation result for q 2. For each t > 0, P [ ] m q 2 m d T + 1 t e 2 P[z m d T + 1 t] ( t 2 exp 4(m d T + 1) ), (11) where z is a chi-squared random variable with m d T + 1 degrees of freedom. We now need to bound the first probability on the right hand side of (8). Using the same reasoning as in proposition 4, it follows directly from (7) that conditioned on q, the components of P T (y) in T are independent and identically distributed with distribution N (0, q 2 2/m). Thus for any τ > 0, we have P[ P T (y) 1 q 2 τ] (n s)p[ v m/τ] ( n exp m ), (12) 2τ 2 where in the first inequality we have used the basic union bound, the fact that dimt = n s and v N (0, 1). In the second inequality we have used the fact that P[ v t] e t2 /2 for positive t. Using the same τ as above with (4), i.e. τ = ms, we get that P[ P T (y) 1 q 2 τ] n exp m s+1 t ( m s + 1 t ), (13) 2s Let D = m s+1. As a consequence of (8), (11) and (13) we have thus proven until now that P[ P T (y) < 1] = 1 P[ P T (y) 1] ) ( 1 exp ( t2 n exp D t ). (14) 4D 2s 3 This is natural in view of what follows. 8

f(β, s) = β 2s + β 1 β 2s We now want to choose t on the right hand side ( of (14) such that ) both exponential terms are equal. For this set t = 2β log(n) 1 + 2s(β 1) 1, and choose β m = 2βs log(n) + s 1 with β > 1. Observe that t = 2β log(n) ( 1 + 2s(β 1) β = 2 log(n) 2sβ f(β, s), ( 2 β where f(β, s) = + β 1 β 2s 2s) as in theorem 2.1. We thus get that t 2 4(m s + 1) 1 = f(β, s) log(n), and using m = 2βs log(n) + s 1 we find that the first exponential term on the right hand side of 14 equals to Now observe that e t 2 4(m s+1) = n f(β,s). ) ( ) 2 f(β, β s) + = β 2s 2s + β 1 f(β, s) + 2 β f(β, s) 2s + β 2s = β 2s + β 1 2β f(β, s) = β 1 f(β, s) s From this we get that t = 2s log(n)(β 1 f(β, s)). Using once more that m = 2βs log(n) + s 1 we get that m s + 1 t = 2s(1 + f(β, s)) log(n) m s + 1 t = log(n 1 f(β,s) ) 2s 9

so that ne m s+1 t 2s = n f(β,s). Taking the previous computations all together, we find that We get thus theorem 2.1. P[ P T (y) < 1] 1 2n f(β,s). 3 Numerical simulations We use Matlab to show how a phase transition in the convex optimization problem (2) takes place. We proceed as described in appendix A of [2]. 1. Construct a vector x 0 R d with s nonzero entries. The locations of the nonzero entries are random, and the nonzero entries are equal to +1 or 1 with equal probability. 2. Draw a random standard normal matrix A R m d. 3. Solve 2 to find ˆx using the cvx package. 4. Declare success if ˆx x 0 2 10 5. This procedure is repeated for a given m, s and d times, where d will be specified. 3.1 Experimental results The following diagrams indicate the empirical probability of success to solve the convex optimization problem 2. The color bar indicates the probability of success. s is the number of non-zero elements in the vector x 0 R d to be recovered and m is the number of random linear measurements done to recover x 0. Note the sharp transition between certain failure (blue) and certain correct recovery (red). 10

Ambient dimension d=25 0.9 0.8 25 0.7 0.6 20 0.5 15 0.4 1 0.5 0 0 5 10 S 15 20 25 0 5 10 m 0.3 0.2 0.1 Figure 1: The dimension of the ambient space is d = 25. For each point the convex optimization problem has been repeated 10 times and the arithmetic mean is then used as an estimate for the probability of success. 0.9 0.8 0.7 0.6 1 30 0.5 0.5 25 0.4 0 20 0.3 0 s 5 10 15 20 25 30 0 5 10 15 m 0.2 0.1 Figure 2: The dimension of the ambient space is d = 30. For each point the convex optimization problem has been repeated 20 times and the arithmetic mean is then used as an estimate for the probability of success. 11

Figure 3: The dimension of the ambient space is d = 50. For each point the convex optimization problem has been repeated 10 times and the arithmetic mean is then used as an estimate for the probability of success. 0 5 10 15 20 25 30 35 40 45 50 50 45 40 0.9 0.8 35 m 30 0.7 0.6 25 0.5 20 0.4 15 0.3 10 0.2 5 0.1 s 0 100 90 80 70 60 0.9 0.8 0.7 0.6 50 m 40 0.5 0.4 30 0.3 20 0.2 s 10 0.1 0 10 20 30 40 50 60 70 80 90 100 s 0 Figure 4: The dimension of the ambient space is d = 100. For each point the convex optimization problem has been repeated 10 times and the arithmetic mean is then used as an estimate for the probability of success. 12

3.2 Matlab code 1 f u n c t i o n [ y]=randomvector (d, s ) 2 %d i s the s i z e o f the random v e c t o r 3 %s the number o f non zero elements. 4 5 a=z e r o s (d, 1 ) ; 6 l =1; 7 while l<s +1, 8 9 x = f l o o r (1 + ( d ) rand ( 1, 1 ) ) ; 10 %This a random element between 1 and d. 11 i f a ( x )==0, 12 a ( x )=2 binornd ( 1,1/2) 1; 13 %This a 1/1 random element 14 l=l +1; 15 end, 16 end 17 y=a ; 1 2 f u n c t i o n [Y]=RandomSystem (m, n) 3 % C r e une matrice m f o i s n avec de v. a. nomales standard. 4 Y=randn (m, n ) ; 1 f u n c t i o n w=s u c c e s s e s (N,m, d, s ) 2 %N i s the number o f experiments to do, f. ex. 50 3 %The s i z e o f the gaussian matrix i s m times d. s i s the s p a r s i t y 4 %o f the v e c t o r to be recovered. 5 %S u c c e s s e s estimated the p r o b a b i l i t y o f s u c c e s s given the data. 6 7 l =0; 8 %Counting the number s u c c e s s e s. 9 t = [ ] ; 10 y = [ ] ; 11 r =[] 12 z = [ ] ; 13

13 f o r i =1:N, 14 t=randomvector (d, s ) ; 15 y=minimizationproblem ( RandomSystem (m, d), t ) ; 16 r=y t ; 17 z=norm ( r ) ; 18 i f z <10ˆ( 5), 19 %This t h r e s h o l d can be modified 20 l=l +1; 21 end 22 end 23 w=l /N; 1 f u n c t i o n s=minimizationproblem (A, y ) 2 %Solves the convex l i n e a r o p t i m i z a t i o n problem 3 %using the cvx package 4 b=a y ; 5 n=s i z e ( y, 1 ) ; 6 cvx begin 7 v a r i a b l e x ( n ) ; 8 minimize ( norm ( x, 1 ) ) ; 9 A x == b ; 10 cvx end 11 s=x ; 1 f u n c t i o n PlottingFunction (N, d) 2 %N i s the number o f experiments to do, f. ex. 50 3 %d i s the dimension o f the ambient space 4 %This p l o t s the p r o b a b i l i t i e s o f s u c c e s s as a f u n c t i o n 5 %o f m( number o f measurements ) and s ( s p a r s i t y ) 6 7 A= [ ] ; 8 %This matrix w i l l contain de data 9 10 f o r s =1:(d 1), 11 f o r m=1:(d 1), 12 A(max( ( s 1) d 1,0)+m)=s u c c e s s e s (N,m, d, s ) ; 13 end 14 end 15 16 17 S = [ ] ; 14

18 M= [ ] ; 19 f o r i =1:(d 1), 20 f o r j =1:(d 1), 21 S (max( ( i 1) d 1,0)+j )=i ; 22 end 23 end 24 25 f o r i =1:(d 1), 26 f o r j =1:(d 1), 27 M(max( ( i 1) d 1,0)+j )=j ; 28 end 29 end 30 31 t r i = delaunay (S,M) ; 32 t r i s u r f ( t r i, S,M,A) ; References [1] Candès, E.; Recht, B.: Simple Bound for Recovering Low-complexity Models. arxiv:1106.1474v2. [2] Amelunxen, D.; Lotz, N.; Mccoy, M.; A. Tropp, Joel: Living on the edge: phase transitions in convex programs with random data. arxiv:1303.6672v2 [3] Laurent, B.; Massart, P.: Adaptive estimation of quadratic functional by model selection. Ann. Statist.,28(5):1302-1338,2000. [4] Roman, S.: Advanced linear algebra, Third edition, Springer. [5] Rivoirard, V; Stoltz, V.: Statistique mathématique en action, 2e édition, Vuibert 15