Linear Convergence of Adaptively Iterative Thresholding Algorithms for Compressed Sensing

Similar documents
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery

Introduction to Compressed Sensing

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Elaine T. Hale, Wotao Yin, Yin Zhang

Robust Sparse Recovery via Non-Convex Optimization

SPARSE signal representations have gained popularity in recent

The Pros and Cons of Compressive Sensing

Least Sparsity of p-norm based Optimization Problems with p > 1

Generalized greedy algorithms.

Optimization for Compressed Sensing

Stability and Robustness of Weak Orthogonal Matching Pursuits

Iterative Reweighted Minimization Methods for l p Regularized Unconstrained Nonlinear Programming

Sparsity Regularization

Generalized Orthogonal Matching Pursuit- A Review and Some

A New Estimate of Restricted Isometry Constants for Sparse Solutions

Complementary Matching Pursuit Algorithms for Sparse Approximation

STAT 200C: High-dimensional Statistics

The Pros and Cons of Compressive Sensing

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Conditions for Robust Principal Component Analysis

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

The Analysis Cosparse Model for Signals and Images

Recent Developments in Compressed Sensing

Does l p -minimization outperform l 1 -minimization?

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit

Greedy Signal Recovery and Uniform Uncertainty Principles

Supremum of simple stochastic processes

Compressed Sensing and Sparse Recovery

Multipath Matching Pursuit

Optimization methods

Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France

ACCORDING to Shannon s sampling theorem, an analog

Sparse Optimization Lecture: Sparse Recovery Guarantees

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm

An Introduction to Sparse Approximation

Robust multichannel sparse recovery

Exact Reconstruction Conditions and Error Bounds for Regularized Modified Basis Pursuit (Reg-Modified-BP)

Sparse Optimization Lecture: Dual Certificate in l 1 Minimization

SCRIBERS: SOROOSH SHAFIEEZADEH-ABADEH, MICHAËL DEFFERRARD

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina

Sparse Solutions of an Undetermined Linear System

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 6, JUNE

A Review of Sparsity-Based Methods for Analysing Radar Returns from Helicopter Rotor Blades

Sparse and Low Rank Recovery via Null Space Properties

Analysis of Greedy Algorithms

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1

GREEDY SIGNAL RECOVERY REVIEW

Greedy Sparsity-Constrained Optimization

Uniqueness Conditions for A Class of l 0 -Minimization Problems

MINIMIZATION OF TRANSFORMED L 1 PENALTY: THEORY, DIFFERENCE OF CONVEX FUNCTION ALGORITHM, AND ROBUST APPLICATION IN COMPRESSED SENSING

5742 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 12, DECEMBER /$ IEEE

Exact penalty decomposition method for zero-norm minimization based on MPEC formulation 1

Bias-free Sparse Regression with Guaranteed Consistency

Douglas-Rachford splitting for nonconvex feasibility problems

Three Generalizations of Compressed Sensing

Compressive Sensing and Beyond

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

Minimizing the Difference of L 1 and L 2 Norms with Applications

Pre-weighted Matching Pursuit Algorithms for Sparse Recovery

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit

Hard Thresholding Pursuit Algorithms: Number of Iterations

Bayesian Methods for Sparse Signal Recovery

Message Passing Algorithms for Compressed Sensing: II. Analysis and Validation

An algebraic perspective on integer sparse recovery

Stochastic geometry and random matrix theory in CS

List of Errata for the Book A Mathematical Introduction to Compressive Sensing

of Orthogonal Matching Pursuit

Machine Learning for Signal Processing Sparse and Overcomplete Representations

COMPARATIVE ANALYSIS OF ORTHOGONAL MATCHING PURSUIT AND LEAST ANGLE REGRESSION

RSP-Based Analysis for Sparsest and Least l 1 -Norm Solutions to Underdetermined Linear Systems

arxiv: v2 [cs.it] 7 Jan 2017

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Sparse Error Correction from Nonlinear Measurements with Applications in Bad Data Detection for Power Networks

Rui ZHANG Song LI. Department of Mathematics, Zhejiang University, Hangzhou , P. R. China

Self-Calibration and Biconvex Compressive Sensing

Reconstruction from Anisotropic Random Measurements

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Sparsity and Compressed Sensing

EUSIPCO

Binary Compressive Sensing via Analog. Fountain Coding

A Polynomial-Time Algorithm for Pliable Index Coding

A Note on the Complexity of L p Minimization

A new method on deterministic construction of the measurement matrix in compressed sensing

Invariancy of Sparse Recovery Algorithms

Enhanced Compressive Sensing and More

Constrained optimization

1 Regression with High Dimensional Data

A Novel Iterative Convex Approximation Method

Lecture Notes 9: Constrained Optimization

Machine Learning for Signal Processing Sparse and Overcomplete Representations. Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013

Sparsity in Underdetermined Systems

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

Subspace Pursuit for Compressive Sensing: Closing the Gap Between Performance and Complexity

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

Structured signal recovery from non-linear and heavy-tailed measurements

Transcription:

Linear Convergence of Adaptively Iterative Thresholding Algorithms for Compressed Sensing Yu Wang, Jinshan Zeng, Zhimin Peng, Xiangyu Chang, and Zongben Xu arxiv:48.689v3 [math.oc] 6 Dec 5 Abstract This paper studies the convergence of the adaptively iterative thresholding (AIT) algorithm for compressed sensing. We first introduce a generalized restricted isometry property (grip). Then we prove that the AIT algorithm converges to the original sparse solution at a linear rate under a certain grip condition in the noise free case. While in the noisy case, its convergence rate is also linear until attaining a certain error bound. Moreover, as by-products, we also provide some sufficient conditions for the convergence of the AIT algorithm based on the two well-known properties, i.e., the coherence property and the restricted isometry property (RIP), respectively. It should be pointed out that such two properties are special cases of grip. The solid improvements on the theoretical results are demonstrated and compared with the known results. Finally, we provide a series of simulations to verify the correctness of the theoretical assertions as well as the effectiveness of the AIT algorithm. Index Terms restricted isometric property, coherence, iterative hard thresholding,, compressed sensing, sparse optimization I. INTRODUCTION Let A R m n, b R m and x R n. Compressed sensing [], [] solves the following constrained l -minimization problem min x R n x s.t. b = Ax+ǫ, ǫ σ () where ǫ R m is the measurement noise, σ R is the noise variance and x denotes the number of the nonzero components of x. Due to the NP-hardness of problem () [3], approximate methods including the greedy method and relaxed method are introduced. The greedy method approaches the sparse solution by successively alternating one or more components that yield the greatest improvement in quality [3]. These algorithms include iterative hard thresholding (IHT) [4], accelerated hard thresholding (AHT) [5], ALPS [6], hard thresholding pursuit (HTP) [7], CLASH [8], OMP [], [], StOMP [], ROMP [3], CoSaMP [4] and SP [5]. The greedy algorithms can be quite efficient and fast in many applications, especially when the signal is very sparse. Y. Wang, J.S. Zeng and Z.B. Xu are with the Institute for Information and System Sciences, School of Mathematics and Statistics, Xi an Jiaotong University, Xi an 749 (email: shifwang@gmail.com, jsh.zeng@gmail.com,zbxu@mail.xjtu.edu.cn). Z.M. Peng is with the Department of Mathematics, University of California, Los Angeles (UCLA), Los Angeles, CA 995, United States (email: zhimin.peng@math.ucla.edu). X.Y. Chang is with the School of Management, Xi an Jiaotong University, Xi an 749. (email: xiangyuchang@gmail.com). Corresponding author: Jinshan Zeng (jsh.zeng@gmail.com). The relaxed method converts the combinatorial l - minimization into a more tractable model through replacing the l norm with a nonnegative and continuous function P( ), that is, min x R np(x) s.t. b = Ax+ǫ, ǫ σ. () One of the most important cases is the l -minimization problem (also known as basis pursuit (BP)) [6] in the noise free case and basis pursuit denoising in the noisy case) with P(x) = x, where x = n i= x i is called the l norm. The l -minimization problem is a convex optimization problem that can be efficiently solved. Nevertheless, the l norm may not induce further sparsity when applied to certain applications [7], [8], [9], []. Therefore, many nonconvex functions were proposed as substitutions of the l norm. Some typical nonconvex examples include the l q ( < q < ) norm [7], [8], [9], smoothly clipped absolute deviation () [] and minimax concave penalty (MCP) []. Compared with the l -minimization model, the nonconvex relaxed models can often induce better sparsity and reduce the bias, while they are generally more difficult to solve. The iterative reweighted method and regularization method are two main classes of algorithms to solve () when P(x) is nonconvex. The iterative reweighted method includes the iterative reweighted least squares minimization (IRLS) [3], [4], and the iterative reweighted l -minimization (IRL) algorithms []. Specifically, the IRLS algorithm solves a sequence of weighted least squares problems, which can be viewed as some approximations to the original optimization problem. Similarly, the IRL algorithm solves a sequence of non-smooth weighted l -minimization problems, and hence it is the non-smooth counterpart to the IRLS algorithm. However, the iterative reweighted algorithms are slow if the nonconvex penalty cannot be well approximated by the quadratic function or the weighted l norm function. The regularization method transforms problem () into the following unconstrained optimization problem min x R n{ Ax b +λp(x)}, (3) where λ > is a regularization parameter. For some special penalties P(x) such as the l q norms (q =,/,/3,), and MCP, an optimal solution of the model (3) is a fixed point of the following equation x = H(x sa T (Ax b)), where H( ) is a componentwise thresholding operator which will be defined in detail in the next section and s > is

a step size parameter. This yields the corresponding iterative thresholding algorithm ([9], [5], [6], [7], [8], [9]) x (t+) = H(x (t) sa T (Ax (t) b)). Compared to greedy methods and iterative reweighted algorithms, iterative thresholding algorithms have relatively lower computational complexities [3], [3], [3]. So far, most of theoretical guarantees of the iterative thresholding algorithms were developed for the regularization model (3) with fixed λ. However, it is in general difficult to determine an appropriate regularization parameter λ. Some adaptive strategies for setting the regularization parameters were proposed. One strategy is to set the regularization parameter adaptively so that x (t) remains the same at each iteration. This strategy was first applied to the iterative hard thresholding algorithm (called algorithm for short henceforth) in [33], and later the iterative soft thresholding algorithm[34] (called algorithm for short henceforth) and the iterative half thresholding algorithm [9] (called algorithm for short henceforth). The convergence of algorithm was justified whenasatisfies the restricted isometry property (RIP) with δ 3k < 3 [33], where k is the number of the nonzero components of the truely sparse signal. Later, Maleki [34] investigated the convergence of both and algorithms in terms of the coherence. Recently, Zeng et al. [35] generalized Maleki s results to a wide class of iterative thresholding algorithms. However, most of guarantees in [35] are coherence-based and focus on the noise free case with the step size equal to. While it has been observed that in practice, the AIT algorithm can have remarkable performances for noisy cases with a variety of step sizes. In this paper, we develop the theoretical guarantees of the AIT algorithm with different step sizes in both noise free and noisy cases. A. Main Contributions The main contributions of this paper are the following. i) Based on the introduced grip, we give a new uniqueness theorem for the sparse signal (see Theorem ), and then show that the AIT algorithm can converge to the original sparse signal at a linear rate (See Theorem ). Specifically, in the noise free case, the AIT algorithm converges to the original sparse signal at a linear rate. While in the noisy case, it also converges to the original sparse signal at a linear rate until reaching an error bound. ii) The tightness of our analyses is further discussed in two specific cases. The coherence based condition for algorithm is the same as those required for both OMP and BP. Moreover, the RIP based condition for algorithm is δ 3k + < 5 8, which is better than the results in [7] and [9]. The rest of this paper is organized as follows. In section II, we describe the adaptively iterative thresholding (AIT) algorithm. In section III, we introduce the generalized restricted isometry property, and then provide a new uniqueness theorem. In section IV, we prove the convergence of the AIT algorithm. In section V, we compare the obtained theoretical results with some other known results. In section VI, we implement a series of simulations to verify the correctness of the theoretical results as well as the efficiency of the AIT algorithm. In section VII, we discuss many practical issues on the implementation of the AIT algorithm, and then conclude this paper in section VIII. All the proofs are presented in the Appendices. Notations. We denote N and R as the natural number set and one-dimensional real space, respectively. For any vector x R n, x i is the i-th component of x for i =,...,n. For any matrix A R m n, A i denotes the i-th column of A. x T and A T represent the transpose of vector x and matrix A respectively. For any index set S {,...,n}, S represents its cardinality. S c is the complementary set, i.e., S c = {,...,n} \ S. For any vector x R n, x S represents the subvector of x with the components restricted to S. Similarly, A S represents the submatrix of A with the columns restricted to S. We denote x as the original sparse signal with x = k, and I = {i : x i = } is the support set of x. I r R r r is the r-dimensional identity matrix. sgn( ) represents the signum function. II. ADAPTIVELY ITERATIVE THRESHOLDING ALGORITHM The AIT algorithm for (3) is the following z (t+) = x (t) sa T (Ax (t) b), (4) where s > is a step size and x (t+) = H τ (t+)(z (t+) ), (5) H τ (t+)(x) = (h τ (t+)(x ),...,h τ (t+)(x n )) T (6) is a componentwise thresholding operator. The thresholding function h τ (u) is defined as { fτ (u), u > τ h τ (u) =, otherwise, (7) where f τ (u) is the defining function. In the following, we give some basic assumptions of the defining function, which were firstly introduced in [35]. Assumption. Assume that f τ satisfies ) Odevity. f τ (u) is an odd function of u. ) Monotonicity. f τ (u) < f τ (v) for any τ u < v. 3) Boundedness. There exist two constants c c such that u c τ f τ (u) u c τ for u τ. Note that most of the commonly used thresholding functions satisfy Assumption. In Fig., we show some typical thresholding functions including hard [7], soft [5] and half [9] thresholding functions for l,l,l / norms respectively, as well as the thresholding functions for l /3 norm [6] and penalty []. The corresponding boundedness parameters are shown in Table I. TABLE I BOUNDEDNESS PARAMETERS c FOR DIFFERENT THERSHOLDING FUNCTIONS f τ, f τ, f τ,/ f τ,/3 f τ, f τ, c 3 c

3 4 3 h τ (u) Definition. For any matrix A R m n, and a constant pair (p,q) where p [, ),q [, ] and p + q =, then the (k,p,q)-generalized restricted isometry constant (gric)β k,p,q of A is defined as u β k,p,q = sup S {,...,n}, S k (I S A T S sup A S)x q. x R S \{} x p (8) 3 /3ThFun (a=3.7) 4 4 3 3 4 Fig.. Typical thresholding functions h τ(u) with τ =. This paper considers a heuristic way for setting the threshold τ (t), specifically, we let τ (t) = z (t) [k+], where z (t) [k+] is the (k + )-th largest component of z (t) in magnitude and k is the specified sparsity level, [k+] denotes the index of this component. We formalise the AIT algorithm as in Algorithm. Algorithm : Adaptively Iterative Thresholding Algorithm Initialization: Normalize A such that A j = for j =,...,n. Given a sparsity level k, a step size s > and an initial point x (). Let t := ; Step : Calculate z (t+) = x (t) sa T (Ax (t) b); Step : Set τ (t+) = z (t+) [k+] and It+ as the index set of the largest k components of z (t+) in magnitude; Step 3: Update: if i I t+, x (t+) i = f τ (t+)(z (t+) i ), otherwise x (t+) i = ; Step 4: t = t+ and repeat Steps -3 until convergence. Remark. At the (t + )-th iteration, the AIT algorithm yields a sparse vector x (t+) with k nonzero components. The sparsity level k is a crucial parameter for the performance of the AIT algorithm. When k k, the results will get better as k decreases. Once k < k, the AIT algorithm fails to find the original sparse solution. Thus, k should be specified as an upper bound estimate of k. Remark. In Algorithm, the columns of matrix A are required to be normalized. Such operation is only for a clearer definition of the following introduced generalized restricted isometry property (grip) and more importantly, better theoretical analyses. However, as shown in Section VII B, this requirement is generally not necessary for the use of the AIT algorithm in the perspective of the recovery performance. We will conduct a series of experiments in Section VII B for a detailed explanation. III. GENERALIZED RESTRICTED ISOMETRY PROPERTY This section introduces the generalized restricted isometry property (grip) and then gives the uniqueness theorem. We will show that the introduced grip satisfies the following proposition. Proposition. For any positive constant pair (p,q) with p + q =, the generalized restricted isometric constant β k,p,q associated with A and k must satisfy 3 β k,p,q sup z R n \{}, z k z T (A T A I n )z z p β k,p,q. (9) The proof of this proposition is presented in Appendix A. It can be noted that the grip closely relates to the coherence property and restricted isometry property (RIP), whose definitions are listed in the following. Definition. For any matrix A R m n, the coherence of A is defined as A i,a j µ = max, () i j A i A j where A i denotes the i-th column of A for i =,...,n. Definition 3. For any matrix A R m n, given k n, the restricted isometry constant (RIC) of A with respect to k, δ k, is defined to be the smallest constant δ such that ( δ) z Az (+δ) z, () for all k-sparse vector, i.e., z k. By Definition 3, RIC can also be written as: z T (A T A I n )z δ k = sup z R n \{}, z k z, () which is very similar to the middle part of (9). In fact, Proposition shows that coherence and RIP are two special cases of grip. Proposition. For any column-normalized matrix A R m n, that is, A j = for j =,...,n, it holds (i) β k,, = µ, for k n. (ii) β k,, = δ k, for k n. The proof of this proposition is shown in Appendix B. A. Uniqueness Theorem Characterized via grip We first give a lemma to show the relation between two different norms for a k-sparse vector space. Lemma. For any vector x R n with x = k n, and for any q p, then x p x q k q p x p. (3)

4 This lemma is trivial based on the well-known norm equivalence theorem so the proof is omitted. Note that Lemma is equivalent to x p k max{ p q,} x q, p,q [, ]. (4) With Lemma, the following theorem shows that a k-sparse solution of the equation Ax = b will be the unique sparsest solution if A satisfies a certain grip condition. Theorem. Let x be a k-sparse solution of Ax = b. If A satisfies (k, p, q)-grip with < β k,p,q < (k) min{ q p,}, then x is the unique sparsest solution. The proof of Theorem is given in Appendix C. According to Proposition and Theorem, we can obtain the following uniqueness results characterized via coherence and RIP, respectively. Corollary. Let x be a k-sparse solution of the equation Ax = b. If µ satisfies < µ < k, then x is the unique sparsest solution. It was shown in [36] that when µ < k, the k-sparse solution should be unique. In another perspective, it can be noted that the condition µ < k is equivalent to k < µ while µ < k is equivalent to k < µ +. Since k should be an integer, these two conditions are almost the same. Corollary. Let x be a k-sparse solution of the equation Ax = b. If δ k satisfies < δ k <, then x is the unique sparsest solution. According to [37], the RIP condition obtained in Corollary is the same as the state-of-the-art result and more importantly, is tight in the sense that once the condition is violated, then we can construct two different signals with the same sparsity. IV. CONVERGENCE ANALYSIS In this section, we will study the convergence of the AIT algorithm based on the introduced grip. A. Characterization via grip To describe the convergence of the AIT algorithm, we first define and L = p (k ) max{ p q,} +( p (c ) p +)k, L = p (k ) max{ p q,} + p (c ) p k, L = min{ p L, p L }, where p [, ),q [, ] and c,c are the corresponding boundedness parameters. Theorem. Let {x (t) } be a sequence generated by the AIT algorithm. Assume that A satisfies (3k +,p,q)-grip with the constant β 3k +,p,q < L, and let (i) k = k ; (ii) s < s < s, where Then and (k ) max{ q p,} L s =, (k ) max{ q p,} β 3k +,p,q (k ) max{ q p,} + L s =. (k ) max{ q p,} +β 3k +,p,q x (t) x p (ρ s ) t x x () p + sl ρ s A T ǫ q, where ρ s = γ s L < with γ s = s (k ) max{ q p,} +sβ 3k +,p,q. Particularly, when ǫ =, it holds x (t) x p (ρ s ) t x x () p. The proof of this Theorem is presented in Appendix D. Under the conditions of this theorem, we can verify that < ρ s <. We first note that β 3k +,p,q < L < (k ) max{ q p,}, then it holds s < < s. The definition of γ s gives γ s = { ( s)(k ) max{ q p,} +sβ 3k +,p,q, if s < s (s )(k ) max{ q p,} +sβ 3k +,p,q, if < s < s. If s < s, it holds γ s < ( s)(k ) max{ q p,} +sβ 3k +,p,q = L. Similarly, if < s s γ s < (s )(k ) max{ q p,} +sβ 3k +,p,q = L. Therefore, we have γ s < L and thus, ρ s = γ s L <. Theorem demonstrates that in the noise free case, the AIT algorithm converges to the original sparse signal at a linear rate, while in the noisy case, it also converges at a linear rate until reaching an error bound. Moreover, it can be noted that the constant ρ s depends on the step size s. Since β 3k +,p,q < L < (k ) max{ q p,}, ρ s reaches its minimum at s =. The trend of ρ s with respect to s is shown in Fig.. The optimal convergence rate is obtained when s =. This observation is consistent with the conclusion drawn in [6]. By Proposition, it shows that the coherence and RIP are two special cases of grip, thus we can easily obtain some recovery guarantees based on coherence and RIP respectively in the next two subsections. Remark 3. From Theorem, we can see that the step size should lie in an appropriate interval, which depends on the grip constant, which is generally NP-hard to verify. However, we would like to emphasize that the theoretical result obtained in Thoeorem is of importance in theory and it can give some insights and theoretical guarantees of the implementation of

5 TABLE II COHERENCE BASED CONDITIONS FOR DIFFERENT AIT ALGORITHMS AIT c µ 3k 3k k 3k Fig.. The trend of ρ s with respect to s. the AIT algorithm, though it seems stringent. Empirically, we find that a small interval of the step size, i.e., [.9,] is generally sufficient for the convergence of the AIT algorithm. This is also supported by the numerical experiments conducted in section VI. In [8], it demonstrates that many algorithms perform well with either constant or adaptive step sizes. In section VII C, we will discuss and compare different stepsize schemes including the constant and an adaptive step-size strategies on the performance of AIT algorithms. B. Characterization via Coherence Let p =,q =. In this case, L = (3 c )k, L = (4+c )k, and L = (3 c )k. According to Theorem and Proposition, assume that µ < (3 c )k, then the AIT algorithm converges linearly with the convergence rate constant ρ s = γ s L = ( s +sµ)l < if we take k = k and L µ < s < + L +µ. In the following, we show that the constant γ s and thus ρ s can be further improved when p = and q =. Theorem 3. Let {x (t) } be a sequence generated by the AIT algorithm for b = Ax+ǫ. Assume that A satisfies < µ < (3 c )k, and if we take (i) k = k ; (ii) L < s < min{ Lµ,+ L }, then it holds x (t) x (ρ s ) t x x () + sl A T ǫ, ρ s where ρ s = γ s L < with γ s = max{ s,sµ}. Particularly, when ǫ =, it holds x (t) x (ρ s ) t x x (). The proof of this Theorem is given in Appendix E. As shown in Theorem 3, the constant γ s can be improved from s +sµ to max{ s,sµ}, and also the feasible ( range ) L of the step size parameter s gets larger from µ, + L ( ) +µ to L,min{ Lµ,+ L }. We list the coherence-based convergence conditions of several typical AIT algorithms in Table II. As shown in Table II, it can be observed that the recovery condition for algorithm is the same as those of OMP [38] and BP [39]. C. Characterization via RIP Let p =,q =. In this case, L = +(3 c )k,l = 4+c k, and thus L = min{ 4+c k, +(3 c )k }. According to Theorem, and by Proposition, we can directly claim the following corollary. Corollary 3. Let {x (t) } be a sequence generated by the AIT algorithm for b = Ax+ǫ. Assume thatasatisfiesδ 3k + < L, and if we take (i) k = k ; (ii) s s s, where s = L Then δ 3k + and s = + L. +δ 3k + x (t) x (ρ s ) t x x () + sl ρ s A T ǫ, where ρ s = γ s L < with γ s = s +sδ 3k +. Particularly, when ǫ =, it holds x (t) x (ρ s ) t x x (). According to Corollary 3, the RIP based sufficient conditions for some typical AIT algorithms are listed in Table III. TABLE III RIP BASED CONDITIONS FOR DIFFERENT AIT ALGORITHMS AIT c /3 δ 3k + 3 36+k +k 4+k Moreover, we note that the condition in Corollary 3 for algorithm can be further improved via using the specific expression of the hard thresholding operator. This can be shown as the following theorem. Theorem 4. Let {x (t) } be a sequence generated by algorithm for b = Ax+ǫ. Assume that A satisfies δ 3k + < 5, and if we take k = k and s =, then 5+ x (t) x ρ t x x () + ρ AT ǫ, where ρ = 5+ δ 3k + <. Particularly, when ǫ =, it holds x (t) x ρ t x x (). The proof of Theorem 4 is presented in Appendix F.

6 V. COMPARISON WITH PREVIOUS WORKS This section discusses some related works of the AIT algorithm, and then compares its computational complexity and sufficient conditions for convergence with other algorithms. ) On related works of the AIT algorithm: In [34], Maleki provided some similar results for two special AIT algorithms, i.e., and algorithms with k = k and s = for the noiseless case. The sufficient conditions for convergence are µ < 3.k and µ < 4.k for and algorithms, respectively. In [35], Zeng et al. improved and extended Maleki s results to a wide class of the AIT algorithm with step size s =. The sufficient condition based on coherence was improved to µ < (3+c )k, where the boudedness parameter c can be found in Table I. Compared with these two tightly related works, several significant improvements are made in this paper. (i) Weaker convergence conditions. The conditions obtained in this paper is weaker than those in both [34] and [35]. More specifically, we give a unified convergence condition based on the introduced grip. Particularly, as shown in Theorem 3, the coherence based conditions for convergence are µ < (3 c )k, which is much better than the condition µ < (3+c )k obtained in [35]. Moreover, except algorithm, we firstly show the convergence of the other AIT algorithms based on RIP. (ii) Better convergence rate. The asymptotic linear convergence rate was justified in both [34] and [35]. However, in this paper, we show the global linear convergence rate of the AIT algorithm, which means it converges at a linear rate from the first iteration. (iii) More general model. In this paper, besides the noiseless model b = Ax, we also consider the performance of the AIT algorithm for the noisy model b = Ax+ǫ, which is very crucial since the noise is almost inevitable in practice. (iv) More general algorithmic framework. In both [34] and [35], the AIT algorithm was only considered with unit step size (s = ). While in this paper, we show that the AIT algorithm converges when s is in an appropriate range. Among these AIT algorithms, algorithm has been widely studied. In [36], it was demonstrated that if A has unit-norm columns and coherence µ, then A has the (r,δ r )- RIP with δ r (r )µ. (5) In terms of RIP, Blumensath and Davies [33] justified the performance of algorithm when applied to signal recovery problem. It was shown that if A satisfies a certain RIP with δ 3k < 3, then algorithm has global convergence guarantee. Later, Foucart improved this condition to δ 3k < or δ k < 4 [4] and further improved it to δ 3k < 3.5773 (Theorem 6.8, [9]). Now we can improve this condition to δ 3k + < 5 8 as shown by Theorem 4. ) On comparison with other algorithms: For better comparison, we list the state-of-the-art results on sufficient conditions of some typical algorithms including BP, OMP, CoSaMP,,, and general AIT algorithms in Table IV. TABLE IV SUFFICIENT CONDITIONS FOR DIFFERENT ALGORITHMS Algorithm µ (r,δ r) BP OMP CoSaMP General AIT ([39]) k ([38]) k.384 4k (Thm. 3) 3k (Thm. 3) k (Thm. 3) 3k (Thm. 3) (3 c )k (k,.77) ([4]) (3k, )(Thm. 6.5,[9]) 6 (4k,.384) ([4]) (3k +,8) (Thm. 4) (3k +, (3k +, (3k +, +k )(Coro. 3) 3 36+k )(Coro. 3) 3) 4+c k )(Coro. : a coherence based sufficient condition for CoSaMP derived by the fact that δ 4k <.384 and δ r (r )µ. From Table IV, in the perspective of coherence, the sufficient conditions of AIT algorithms are slightly stricter than those of BP and OMP algorithms except algorithm. However, AIT algorithms are generally faster than both BP and OMP algorithms with lower computational complexities, especially for large scale applications due to their linear convergence rates. As shown in the next section, the number of iterations required for the convergence of the AIT algorithm is empirically of the same order of the original sparsity level k, that is, O(k ). At each iteration of the AIT algorithm, only some simple matrix-vector multiplications and a projection on the vector need to be done, and thus the computational complexity per iteration is O(mn). Therefore, the total computational complexity of the AIT algorithm iso(k mn). While the total computational complexities of BP and OMP algorithms are generallyo(m n) and max{o(k mn),o( (k ) (k +) 4 )}, respectively. It should be pointed out that the computational complexity of OMP algorithm is related to the commonly used halting rule of OMP algorithm, that is, the number of maximal iterations is set to be the true sparsity level k. Another important greedy algorithm, CoSaMP algorithm, identifies multicomponents (commonly k ) at each iteration. From Table IV, the RIP based sufficient condition of CoSaMP is δ 4k <.384 and a deduced coherence based sufficient condition is µ <.384 4k. In the perspective of coherence, our conditions for AIT algorithms are better than CoSaMP, though this comparison is not very reasonable. On the other hand, our conditions for AIT algorithms except algorithm are generally worse than that of CoSaMP in the perspective of RIP. However, when the true signal is very sparse, the conditions of AIT algorithms may be better than that of CoSaMP. At each iteration of CoSaMP algorithm, some simple matrixvector multiplications and a least squares problem should be considered. Thus, the computational complexity per iteration of CoSaMP algorithm is generally max{o(mn),o((3k ) 3 )}, which is higher than those of AIT algorithms, especially when k is relatively large. Besides BP and greedy algorithms, another class of tightly related algorithms is the reweighted techniques that have

7 been also widely used for solutions to l q regularization with q (, ). Two well-known examples of such reweighted techniques are the iterative reweighted least squares (IRLS) method [3] and the reweighted l minimization (IRL) method []. The convergence analysis conducted in [4] shows that the IRLS method converges with an asymptotically superlinear convergence rate under the assumptions that A possesses a certain null-space property (NSP). However, from Theorem, the rates of convergence of AIT algorithms are globally linear. Furthermore, Lai et al. [4] applied the IRLS method to the unconstrainedl q minimization problem and also extended the corresponding convergence results to the matrix case. It was shown also in [43] that the IRL algorithm can converge to a stationary point and the asymptotic convergence speed is approximately linear when applied to the unconstrained l q minimization problem. Both in [4] and [43], the authors focused on the unconstrainedl q minimization problem with a fixed regularization parameter λ, while in this paper, we focus on a different model with an adaptive regularization parameter. VI. DISCUSSION In this section, we numerically discuss some practical issues on the implementation of AIT algorithms, especially, the effects of several algorithmic factors including the estimated sparsity level parameter, the column-normalization operation, different step-size strategies as well as the formats of different thresholding operators on the performance of AIT algorithms. Moreover, we will further demonstrate the performance of several typical AIT algorihtms including, and via comparing with many state-of-the-art algorithms such as CGIHT [5], CoSaMP [4], -ALPS(4) [6] in the perspective of the 5% phase transition curves [47], [49]. A. Robustness of the estimated sparsity level In the preceding proposed algorithms, the specified sparsity level parameter k is taken exactly as the true sparsity level k, which is generally unknown in practice. Instead, we can often obtain a rough estimate of the true sparsity level. Therefore, in this experiment, we will explore the performance of the AIT algorithm with a variety of specified sparsity levels. We varied k from to 5 while kept k = 5. The experiment setup is the same with Section VI. A. From Fig. 3, we can observe that these AIT algorithms are efficient for a wide range of k. Interestingly, the point k = k is a break point of the performance of all these AIT algorithms. When k < k, all AIT algorithms fail to recover the original sparse signal, while when k k, a wide interval of k is allowed for small recovery errors, as shown in Fig. 3 (b) and (d). In the noise free case, if x (t) x <, the feasible intervals of the specified sparsity level k are [5,9] for and, [5,8] for and [5,65] for, respectively. This observation is very important for real applications of AIT algorithms becausek is usually unknown. In the noisy case, if x (t) x <, the feasible intervals of sparsity level k are [5,5] for, [5,4] for, [5,37] for and [5,6] for, respectively. Recovery Error Recovery Error 5 5 5 Estimated Sparsity Parameter k (a) Robust (Noiseless) 4 6 8 4 Estimated Sparsity Parameter k (c) Robust (Noisy) Recovery Error Recovery Error 4 6 8 Estimated Sparsity Parameter k (b) Detail (Noiseless) 5 5 3 Estimated Sparsity Parameter k (d) Detail (Noisy) Fig. 3. On robustness of the specified sparsity level. (a) The trends of the recovery precision with different estimated sparsity levels in noiseless case. (b) The detailed trends of the recovery precision with different estimated sparsity levels in noiseless case. (c) The trends of the recovery precision with different estimated sparsity levels in noiseless case. (d) The detailed trends of the recovery precision with different estimated sparsity levels in noisy case. TABLE V THE RECOVERY PRECISION OF DIFFERENT AIT ALGORITHMS WITH OR WITHOUT COLUMN-NORMALIZATION (NOISELESS CASE) Algorithm no normalization 5.79e-6.45e-8 5.6e-6 9.33e-9 normalization 5.73e-5.437e-8 5.935e-5 8.55e-9 B. With vs Without Normalization As shown in Algorithm, the column-normalization on the measurement matrixais required in consideration of a clearer definition of the introduced grip and more importantly, better theoretical analyses. However, in this subsection, we will conduct a series of simulations to show that such requirement is generally not necessary in practice. The experiment setup is similar to Section VI.A. More specifically, we set m = 5, n = 4 and k = 5. The nonzero components of x were generated randomly according to the standard Gaussian distribution. The matrix A was generated from i.i.d Gaussian distribution N(, /5) without normalization. In order to adopt Algorithm, we let Λ be the corresponding columnnormalized factor matrix of A (i.e., Λ is a diagonal matrix and its diagonal element is the l -norm of the corresponding column of A), and  = AΛ be the corresponding columnnormalized measurement matrix. Assume that ˆx is a recovery via Algorithm corresponding to Â, then x = Λ ˆx is the corresponding recovery of x. For each algorithm, we conducted times experiments independently in both noiseless and noise (signal-to-noise ratio (SNR): 6dB) cases, and recorded the average recovery precision. The recovery precision is defined as x x x, where x and x represent the recovery and original signal, respectively. The experiment results are shown in Table V and VI.

8 TABLE VI THE RECOVERY PRECISION OF DIFFERENT AIT ALGORITHMS WITH OR WITHOUT COLUMN-NORMALIZATION (WITH 6DB NOISE) Algorithm no normalization.7e-3 5.739e-3.6e-3.8e-3 normalization.4e-3 5.498e-3.5e-3.64e-3 From Table V and VI, we can see that the columnnormalization operator has almost no effect on the performance of the AIT algorithm in both noiseless and noise cases. Therefore, in the following experiments, we will adopt the more practical AIT algorithm without the columnnormalization for better comparison with the other algorithms. C. Constant vs Adaptive Step Size From Algorithm, we only consider the constant stepsize. However, according to many previous and empirical studies [6], [46], we have known that certain adaptive step-size strategies may improve the performance of AIT algorithms. In this subsection, we will compare the performance of two different step-size schemes, i.e., the constant step-size strategy and an adaptive step-size strategy introduced in [46] via the socalled 5% phase transition curve [49]. More specifically, the adaptive step-size scheme can be described as follows. Assume that x (t) is the t-th iteration, then at (t + )-th iteration, the step size s (t+) is set as s (t+) = (AT (b Ax (t) )) I t (AA T (b Ax (t) )) I t, (6) where I t is the support set of x (t), A is the measurement matrix and b is the measurement vector. Similar to [46], we will call the AIT algorithm with such adaptive step-size strategy the normalised AIT (NAIT) algorithm, and correspondingly, several typical AIT algorithms such as,, and algorithms with such adaptive step-size strategy N, N, N and N for short, respectively. Note that N algorithm studied here is actually the same with the normalised iterative hard thresholding (NIHT) algorithm proposed in [46]. 5% phase transition curve was first introduced in [48] and has been widely used to compare the recovery ability for different algorithms in compressed sensing [47], [49]. For a fixed n, any given problem setting (k,m,n) can depict a point in the space (m/n,k/m) (,]. For any algorithm, its 5% phase transition curve is actually a function f on the (k/m, m/n) space. More specifically, if the point(m/n, k/m) lies below the curve of the algorithm, i.e. k/m < f(m/n), then it means the algorithm could recover the sparse signal from the given (k, m, n)-problem with high probability, otherwise the successful recovery probability is very low [48]. Moreover, the 5% phase transition curve usually depends on the prior distribution of x as depicted in many researches [7], [47], [49]. In these experiments, we consider two common distributions of x, the first one is the standard Gaussian distribution, and the second one is a binary distribution, which takes or with an equal probability. For any given (k,m,n), the measurement matrix A R m n is generated from the Gaussian distribution N(, m ), and the nonzero components of the original k-sparse signal x are generated independently and identically distribution (i.i.d.) according to the Gaussian or binary distributions. For any experiment, we consider it as a successful recovery if x x x 3, where x is the original sparse signal and x is the corresponding recovery signal. We set n = 5, m = 5,,...,5. To determine f(m/n), we exploit a bisection search scheme as the same as the experiment setting in [47]. We compare the 5% phase transition curves of,, and algorithms with their adaptive step-size versions, i.e., N, N, N, N in Fig. 4. From Fig. 4 (a) and (c), we can see that the performances of all AIT algorithms except algorithm adopting the adaptive step-size strategy (6) are significantly better than those of the corresponding AIT algorithms with a constant step size in the Guassian case. In this case, N has the best performance, then N and N, while N is the worst. The performance of N is slightly better than those of N and N, and much better than N. While for the binary case, as shown in Fig. 4 (b) and (d), N breaks down with the curve fluctuating around. while N and N still perform well. In the binary case, as well as N perform the worst. In addition, we can see that the performances of and N are almost the same in all cases, which means that such adaptive step-size strategy (6) may not bring the improvement on the performance of algorithm. Moreover, some interesting phenomena can also be observed in Fig. 4, that is, the performance of the AIT algorithm depends to some extent on the choice of the thresholding operator, and for different prior distributions of the original sparse signal, the AIT algorithm may perform very different. For these phenomena, we will study in the future work. D. Comparison with the State-of-the-art Algorithms We also compare the performance of several AIT algorithms including N, N and N with some typical stateof-the-art algorithms such as conjugate gradient iterative hard thresholding (CGIHT) [5], CoSaMP [4], -ALPS(4) [6] in terms of their 5% phase transition curves. For more other algorithms like MP [3], HTP [7], OMP [], CSMPSP [5], CompMP [5], OLS [53] etc., their 5% phase transition curves can be found in [49], and we omit them here. For all considered algorithms, the estimated sparsity level parameters are set to be the true sparsity level of x. The result is shown in Fig 5. From Fig. 5, we can see that almost all algorithms have better performances for the Gaussian distribution case than for the binary distribution case, especially N algorithm. More specifically, as shown in Fig. 5(a), for the Gaussian distribution, N has the best performance among all these

9 k/m k/m.8.7.5.3.. N N N N Gaussian..8 m/n (a) NAIT for Gaussian case.8.7.5.3.. Gaussian..8 m/n (c) AIT for Gaussian case k/m k/m.8.7.5.3...8.7.5.3.. Binary N N N N..8 m/n (b) NAIT for Binary case Binary..8 m/n (d) AIT for Binary case Fig. 4. 5% phase transition curves of different AIT algorithms with two different step-size schemes. (a) AIT algorithms with an adaptive step size for Gaussian case. (a) AIT algorithms with an adaptive step size for Binary case. (c) AIT algorithms with a constant step size for Gaussian case. (d) AIT algorithms with a constant step size for Binary case. k/m.8.7.5.3.. Gaussian ALPS(4) CoSaMP N CGIHT N N..8 m/n (a) Gaussian Distribution k/m.8.7.5.3.. ALPS(4) CoSaMP N CGIHT N N Binary..8 m/n (b) Binary Distribution Fig. 5. 5% phase transition curves of different algorithms. (a) Gaussian distribution case. (b) Binary distribution case. algorithms, and N is slightly worse than N and better than the other algorithms. While in the binary case, it can be seen from Fig. 5(b), all AIT algorithms perform worse than the other algorithms like CGIHT, CoSaMP, - ALPS(4), especially, N algorithm is much worse than the other algorithms. These experiments demonstrate that AIT algorithms are more appropriate for the recovery problems that the original sparse signals obey the Gaussian distribution. VII. CONCLUSION We have conducted a study of a wide class of AIT algorithms for compressed sensing. It should be pointed out that almost all of the existing iterative thresholding algorithms like,, and are included in such class of algorithms. The main contribution of this paper is the establishment of the convergence analyses of the AIT algorithm. In summary, we have shown when the measurement matrix satisfies a certain grip condition, the AIT algorithm can converge to the original sparse signal at a linear rate in the noiseless case, and approach to the original sparse signal at a linear rate until achieving an error bound in the noisy case. As two special cases of grip, the coherence and RIP based conditions can be directly derived for the AIT algorithm. Moreover, the tightness of our analyses can be demonstrated by two specific cases, that is, the coherence-based condition for algorithm is the same as those of OMP and BP, and the RIP based condition for algorithm is better than the recent result δ 3k < 3.5773 obtained in Theorem 6.8 in [9]. Furthermore, the efficiency of the algorithm and the correctness of the theoretical results are also verified via a series of numerical experiments. In section VII, we have numerically discussed many practical issues on the implementation of AIT algorithms, including the specified sparsity level parameter k, the columnnormalization requirement as well as different step-size setting schemes. We can observe the following several interesting phenomena: (i) The AIT algorithm is robust to the specified sparsity level parameter k, that is, the parameter k can be specified in a large range to guarantee the well performance of the AIT algorithm. (ii) The column-normalization of the measurement matrix A is not necessary for the use of AIT algorithms in the perspective of the recovery performance. (iii) Some adaptive step-size strategies may significantly improve the performance of AIT algorithms. (iv) The performance of AIT algorithm depends to some extent on the prior distribution of the original sparse signal. Compared with the binary distribution, AIT algorithms are more appropriate for the recovery of the sparse signal generated by the Gaussian distribution. (v) The performance of the AIT algorithm depends on the specific thresholding operator. All of these phenomena are of interest, and we will study them in our future work. ACKNOWLEDGEMENTS The authors would like to thank Prof. Wotao Yin, Department of Mathematics, University of California, Los Angeles (UCLA), Los Angeles, CA, United States, for a careful discussion of the manuscript, which led to a significant improvement of this paper. Moreover, we thank the three anonymous reviewers, the associate editor and the potential reviewers for their constructive and helpful comments. This work was partially supported by the National 973 Programs (Grant No. 3CB3944), the Key Program of National Natural Science Foundation of China (Grants No. 36), the National Natural Science Foundations of China (Grants No. 7, 77, 446), NSF Grants NSF DMS-349855 and DMS-376. APPENDIX A: PROOF OF PROPOSITION Proof: For any index set S {,...,n} with S k and a vector x R S, since p + q =, then l p and l q norms are dual to each other, which implies that (I S A T SA S )x q = sup y R S \{} y T (I S A T S A S)x. y p (7)

By Definition, then β k,p,q = sup S k It is obvious that β k,p,q sup S k sup x,y R S \{} sup x R S \{} = sup z R n \{}, z k y T (I S A T S A S)x x p y p. (8) x T (I S A T S A S)x x p z T (I n A T A)z z, p which implies the right-hand side of (9). On the other hand, by (8), we can also observe that β k,p,q = sup S k and for any x,y R S, sup y T x y T A T SA S x, (9) x p, y p y T x y T A T S A Sx = ( x + y x y ) ( A Sx + A Sy A Sx A S y ) x A S x + y A S y + x y A S(x y). () Furthermore, it can be noted that sup u v A S(u v) u,v R S, u p, v p 4 sup w A Sw, () w R S, w p since u v p for u p and v p. Plugging () and () into (9), it yields βk,p,q A 3 sup sup x A S x S k x p = 3 sup z Az z k, z p z T (A T A I n )z = 3 sup z R n \{}, z k z, p which implies the left-hand side of (9). Therefore, the proof of this proposition is completed. APPENDIX B: PROOF OF PROPOSITION Proof: (i) The definition of grip inducesβ k,, β,, for all k. Therefore, if we can claim the following two facts: (a) β,, µ, and (b) β k,, µ for all k, then Proposition (i) follows. We first justify the fact (a). Suppose the maximal element of I n A T A in magnitude appears at the i -th row and the j -th column. Because for any j, the j-th diagonal elements of I n A T A equals to A j =, we knowi j. Without loss of generality, we assume that i < j. Let A i and A j be the i -th and j -th column vector of A, respectively, then Definition gives µ = A T i A j. Let S = {i,j } and e = (,) T. Then β,, (I A T SA S )e = e A T S A j = µ. () Then we prove the fact (b). For any vector x R k and a subset S {,,...,n} with S = k, let B = I k A T S A S and z = Bx. Then k k z i = B ij x j B ij x j µ x, j= j= for any i =,...,k. It implies that Bx µ x. By the definition of β k,,, it implies β k,, µ. (3) According to () and (3), for all k n, it holds β k,, = µ. (ii) From the inequality (9) and the equality (), we know To prove note that equality (9) leads to and further β k,p,q δ k β k,p,q. (4) δ k β k,p,q, (5) sup I S A T SA S, (6) S {,...,n}, S k sup I S A T SA S (7) S {,...,n}, S k x T (I S A T S = sup sup A S)x S k x R S \{} x z T (I n A T A)z = sup z R n \{}, z k z = δ k, where the last equality holds by the equivalent definition of RIP (this can be also referred to Definition in [4]). From (4)-(7), we can conclude that δ k = β k,p,q. APPENDIX C: PROOF OF THEOREM Proof: We prove this theorem by contradiction. Assume x satisfies Ax = b and x k. Then which implies A(x x ) =, (I n A T A)(x x ) = x x.

Let x = x x, S be the support of x and x S be a subvector of x with the components restricted to S. It follows and further (I S A S T A S )x S = x S, (I S A S T A S )x S q = x S q, (8) for any q [, ]. Since x k and x k, then S k. For any p [, ), and by the definition of grip, we have (I S A S T A S )x S q β k,p,q x S p. By Lemma, there holds x S p (k) max{ p q,} x S q. By the assumption of this theorem, then (I S A S T A S )x S q β k,p,q (k) max{ p q,} x S q < x S q, which contradicts with (8). Therefore, x is the unique sparsest solution. APPENDIX D: PROOF OF THEOREM Before justifying the convergence of the AIT algorithm based on grip, we first introduce two lemmas. Lemma. For any x,y R n, and p [, ), then x+y p p p ( x p p + y p p ). (9) Moreover, if x i y i for i =,...,n, then x+y p p x p p + y p p. (3) The proof of Lemma is obvious since f(z) = z p is convex for p and any z, and x p x for any x R n. We will omit it due to the limitation of the length of the paper. Lemma 3. For any t and q [, ], if k k, the following inequality holds for the AIT algorithm: τ (t) ( z (t) i x i q ) /q = z (t) x I+ t I+ t q, (3) i I t + where I+ t is the index set of the largest k + components of z (t) in magnitude. Proof: When q =, we need to show τ (t) max z (t) i I+ t i x i, (3) then Lemma shows that (3) holds for all q [, ]. Let I t be the index set of the largest k components of z (t) in magnitude, then I t + = It {[k + ]}, where [k + ] represents the index of the (k + )-th largest component of z (t) in magnitude. We will prove (3) in the following two cases. Case (i). If I = I t, then τ (t) = z (t) [k+] = z(t) [k+] x [k+] max z (t) i I+ t i x i. (33) Case (ii). If I I t, then there exists i I t such that i / I. Otherwise I t I and I t I which contradicts with I t k and I = k. Thus, x i = and τ (t) = z (t) [k+] z(t) i = z (t) i x i max z (t) i I+ t i x i. (34) Combining (33) and (34) gives (3). Proof of Theorem : In order to prove this theorem, we only need to justify the following two inequalities, i.e., for any t N, z (t+) S t x S t q γ s x (t) x p +s A T ǫ q, (35) and for any t, x (t) x p L z (t) S t x S t q. (36) Then combining (35) and (36), it holds x (t+) x p L z (t+) S x t S t q ρ s x (t) x p +sl A T ǫ q. Since < ρ s < under the assumption of this theorem, then by induction for any t, we have x (t) x p (ρ s ) t x x () p + sl ρ s A T ǫ q. First, we turn to prove the inequality (35). By the Step of Algorithm, for any t N, z (t+) = x (t) sa T (Ax (t) b), and we note that b = Ax +ǫ, then z (t+) x = (I n sa T A)(x (t) x )+sa T ǫ = ( s)(x (t) x )+s(i n A T A)(x (t) x )+sa T ǫ. For any t N and q [, ], let S t = I+ t+ I t I. Noting that I t,i S t, it follows Then we have Therefore, z (t+) S t A(x (t) x ) = A S t(x (t) S t x S t). z (t+) S t x S t = ( s)(x (t) S t x S t ) +s(i St A T S t A S t)(x (t) S t x S t )+sa T S t ǫ. x S t q s x (t) S t x S t q +s (I S t A T S ta S t)(x(t) S t x S t) q +s A T S tǫ q. (37) Since x (t) k = k and x = k then I t k, I t+ + k +, I = k,

and hence S t 3k +. For any p [, ), by (4) and the definition of grip (8), it holds and x (t) x q (k ) max{ q p,} x (t) x p, (38) (I St A T S t A S t)(x (t) S t x S t ) q β 3k +,p,q x (t) S t x S t p = β 3k +,p,q x (t) x p.(39) Plugging (38) and (39) into (37), then z (t+) S t x S t q ( s (k ) max{ q p,} +sβ 3k +,p,q +s A T S tǫ q γ s x (t) x p +s A T ǫ q. ) x (t) x p Thus, we have obtained the inequality (35). Then we turn to the proof of (36). We will prove it in two steps. Step a): For any p [, ), By Lemma, x (t) x p p = x(t) I x I p p + x(t) I t \I p p. (4) x (t) I x I p p = x (t) I z(t) I +z(t) I x I p p p z (t) I x(t) I p p +p z (t) I x I p p. (4) Moreover, by the Step 3 of Algorithm and Assumption, for any i I t : sgn(x (t) i ) = sgn(z (t) i ) and x (t) i z (t) i. Thus, for any i I t \I, it holds With (4) and by Lemma, we have x (t) i (z (t) i x (t) i ). (4) z (t) I t \I p p = x (t) I t \I +(z (t) I t \I x (t) I t \I ) p p. x (t) I t \I p p + z(t) I t \I x (t) I t \I p p. (43) Plugging (4) and (43) into (4), it becomes x (t) x p p p ( z (t) I x(t) I p p + z (t) I x I p p) + z (t) I t \I p p z(t) I t \I x (t) I t \I p p. (44) Furthermore, by the Step of Algorithm, Assumption and Lemma 3, for any t, we have: (a) if i I t, c τ (t) z (t) i x (t) (b) if i I t, z (t) i x (t) i = z (t) (c) τ (t) z (t) I t + x I t + q. By the above facts (a)-(c), it holds and i c τ (t) τ (t) ; i τ (t) ; z (t) I x(t) I p p k max i I z(t) i x (t) i p k τ (t) p, (45) z (t) I t \I x (t) I t \I p p It \I min i I t \I z(t) i x (t) i p I t \I (c ) p τ (t) p, (46) where I t \I represents the cardinality of the index set I t \I. Plugging (45), (46) into (44), it follows x (t) x p p p z (t) I x I p p + z (t) I t \I p p Furthermore, we note that +( p k (c ) p I t \I ) τ (t) p. (47) z (t) I t \I p p = z (t) I t \I x I t \I p p I t \I max i I t \I z(t) i x i p = I t \I z (t) I t \I x I t \I p I t \I z (t) I t \I x I t \I p q, where the first equality holds because x I t \I =, and the second inequality holds because of Lemma. Therefore, (47) becomes x (t) x p p p z (t) I x I p p + It \I z (t) I t \I x I t \I p q + ( p k (c ) p I t \I ) τ (t) p p z (t) I x I p p +( p k (c ) p I t \I + I t \I ) z (t) I t + x I t + p q p (k ) max{ p q,} z (t) I x I p q +( p (c ) p +)k z (t) I t + x I t + p q L z (t) S t x S t p q, (48) where the second inequality holds by the fact (c), i.e., τ (t) z (t) x I+ t I+ t q, the third inequality holds by Lemma and I t \I k and the last inequality holds because S t = I+ t It I. Thus, it implies x (t) x p p L z (t) S t x S t q. (49) Step b): By Lemma, x (t) x p p = x(t) I x t I t p p + x I \I t p p p z (t) I x t I t p p +p z (t) I x (t) t I p t p + p z (t) I \I x t I \I t p p +p z (t) I \I p t p = p z (t) I t I x I t I p p + p ( z (t) I x (t) t I p t p + z(t) I \I p t p ). (5) Moreover, by Lemma, it holds z (t) I t I x I t I p p ( I t I ) max{ p q,} z (t) I t I x I t I p q (k ) max{ p q,} z (t) I t I x I t I p q, (5) where the last inequality holds for I t I k. We also have z (t) I x (t) t I p t p k max i I z(t) t i x (t) i p k (c τ (t) ) p k (c ) p z (t) I t + x I t + p q. (5)