Accelerated Path-following Iterative Shrinkage Thresholding Algorithm with Application to Semiparametric Graph Estimation

Similar documents
arxiv: v7 [stat.ml] 9 Feb 2017

The picasso Package for Nonconvex Regularized M-estimation in High Dimensions in R

BAGUS: Bayesian Regularization for Graphical Models with Unequal Shrinkage

An efficient ADMM algorithm for high dimensional precision matrix estimation via penalized quadratic loss

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization /

Efficient Quasi-Newton Proximal Method for Large Scale Sparse Optimization

Optimization methods

Properties of optimizations used in penalized Gaussian likelihood inverse covariance matrix estimation

arxiv: v5 [cs.lg] 23 Dec 2017 Abstract

The Nonparanormal skeptic

Regularized Parameter Estimation in High-Dimensional Gaussian Mixture Models

Estimators based on non-convex programs: Statistical and computational guarantees

Sparse inverse covariance estimation with the lasso

Sparse Gaussian conditional random fields

Analysis of Greedy Algorithms

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

Approximation. Inderjit S. Dhillon Dept of Computer Science UT Austin. SAMSI Massive Datasets Opening Workshop Raleigh, North Carolina.

Sparse Learning and Distributed PCA. Jianqing Fan

Lecture 25: November 27

Sparse Covariance Matrix Estimation with Eigenvalue Constraints

A direct formulation for sparse PCA using semidefinite programming

Adaptive estimation of the copula correlation matrix for semiparametric elliptical copulas

Supplement to A Generalized Least Squares Matrix Decomposition. 1 GPMF & Smoothness: Ω-norm Penalty & Functional Data

Sparsity Regularization

High-dimensional covariance estimation based on Gaussian graphical models

Learning discrete graphical models via generalized inverse covariance matrices

Generalized Power Method for Sparse Principal Component Analysis

STAT 200C: High-dimensional Statistics

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Nonconvex Low Rank Matrix Factorization via Inexact First Order Oracle

Accelerated Stochastic Block Coordinate Gradient Descent for Sparsity Constrained Nonconvex Optimization

Permutation-invariant regularization of large covariance matrices. Liza Levina

Robust Inverse Covariance Estimation under Noisy Measurements

Causal Inference: Discussion

Nonconvex Low Rank Matrix Factorization via Inexact First Order Oracle

A direct formulation for sparse PCA using semidefinite programming

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

High Dimensional Inverse Covariate Matrix Estimation via Linear Programming

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

arxiv: v3 [stat.me] 8 Jun 2018

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Big Data Analytics: Optimization and Randomization

Tighten after Relax: Minimax-Optimal Sparse PCA in Polynomial Time

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

Lasso: Algorithms and Extensions

SVRG++ with Non-uniform Sampling

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models

Coordinate Descent and Ascent Methods

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

SOLVING NON-CONVEX LASSO TYPE PROBLEMS WITH DC PROGRAMMING. Gilles Gasso, Alain Rakotomamonjy and Stéphane Canu

Randomized Block Coordinate Non-Monotone Gradient Method for a Class of Nonlinear Programming

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Proximal Gradient Algorithm

OWL to the rescue of LASSO

WE consider an undirected, connected network of n

An Efficient Proximal Gradient Method for General Structured Sparse Learning

Optimal Linear Estimation under Unknown Nonlinear Transform

Optimal prediction for sparse linear models? Lower bounds for coordinate-separable M-estimators

Comparisons of penalized least squares. methods by simulations

Sparse PCA with applications in finance

Distributed Inexact Newton-type Pursuit for Non-convex Sparse Learning

Stochastic Quasi-Newton Methods

r=1 r=1 argmin Q Jt (20) After computing the descent direction d Jt 2 dt H t d + P (x + d) d i = 0, i / J

Robust and sparse Gaussian graphical modelling under cell-wise contamination

ADMM and Fast Gradient Methods for Distributed Optimization

Robust Principal Component Analysis

An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems

arxiv: v1 [math.st] 13 Feb 2012

Fast Nonnegative Matrix Factorization with Rank-one ADMM

Lasso Screening Rules via Dual Polytope Projection

arxiv: v1 [cs.lg] 4 Apr 2017

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method

Variable Selection for Highly Correlated Predictors

DISCUSSION OF INFLUENTIAL FEATURE PCA FOR HIGH DIMENSIONAL CLUSTERING. By T. Tony Cai and Linjun Zhang University of Pennsylvania

arxiv: v1 [stat.me] 13 Jun 2016

sparse and low-rank tensor recovery Cubic-Sketching

Lecture 1: Supervised Learning

Shrinkage Tuning Parameter Selection in Precision Matrices Estimation

Machine Learning and Computational Statistics, Spring 2017 Homework 2: Lasso Regression

GLOBAL SOLUTIONS TO FOLDED CONCAVE PENALIZED NONCONVEX LEARNING. BY HONGCHENG LIU 1,TAO YAO 1 AND RUNZE LI 2 Pennsylvania State University

l 1 and l 2 Regularization

Variable Selection for Highly Correlated Predictors

Regularization Path Algorithms for Detecting Gene Interactions

Fast Regularization Paths via Coordinate Descent

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1

Journal of Multivariate Analysis. Consistency of sparse PCA in High Dimension, Low Sample Size contexts

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty

A Fast Algorithm for Computing High-dimensional Risk Parity Portfolios

Nonconvex penalties: Signal-to-noise ratio and algorithms

Nonconvex Sparse Logistic Regression with Weakly Convex Regularization

A Distributed Newton Method for Network Utility Maximization, I: Algorithm

Exact Hybrid Covariance Thresholding for Joint Graphical Lasso

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables

Randomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity

Transcription:

Accelerated Path-following Iterative Shrinage Thresholding Algorithm with Application to Semiparametric Graph Estimation Tuo Zhao Han Liu Abstract We propose an accelerated path-following iterative shrinage thresholding algorithm (APISTA) for solving high dimensional sparse nonconvex learning problems. The main difference between APISTA and the path-following iterative shrinage thresholding algorithm (PISTA) is that APISTA exploits an additional coordinate descent subroutine to boost the computational performance. Such a modification, though simple, has profound impact: APISTA not only enjoys the same theoretical guarantee as that of PISTA, i.e., APISTA attains a linear rate of convergence to a unique sparse local optimum with good statistical properties, but also significantly outperforms PISTA in empirical benchmars. As an application, we apply APISTA to solve a family of nonconvex optimization problems motivated by estimating sparse semiparametric graphical models. APISTA allows us to obtain new statistical recovery results which do not exist in the existing literature. Thorough numerical results are provided to bac up our theory. 1 Introduction High dimensional data challenge both statistics and computation. In the statistics community, researchers have proposed a large family of regularized M-estimators, including Lasso, Group Lasso, Fused Lasso, Graphical Lasso, Sparse Inverse Column Operator, Sparse Multivariate Regression, Sparse Linear Discriminant Analysis (Tibshirani, 1996; Zou and Hastie, 005; Yuan and Lin, 005, 007; Banerjee et al., 008; Tibshirani et al., 005; Jacob et al., 009; Fan et al., 01; Liu and Luo, 015; Han et al., 01; Liu et al., 015). Theoretical analysis of these methods usually Department of Computer Science, Johns Hopins University, Baltimore, MD 118, USA; e-mail: tour@cs.jhu.edu. Department of Operations Research Research and Financial Engineering, Princeton University, Princeton, NJ 08544, USA; e-mail: hanliu@princeton.edu. Research supported by NSF IIS1116730, NSF IIS133109, NSF IIS1408910, NSF IIS154648-BIGDATA, NSF DMS1454377-CAREER, NIH R01GM083084, NIH R01HG06841, NIH R01MH10339, and FDA HHSF30100007C. 1

rely on the sparsity of the parameter space and requires the resulting optimization problems to be strongly convex over a restricted parameter space. More details can be found in Meinshausen and Bühlmann (006); Zhao and Yu (006); Zou (006); Rothman et al. (008); Zhang and Huang (008); Van de Geer (008); Zhang (009); Meinshausen and Yu (009); Wainwright (009); Fan et al. (009); Zhang (010a); Raviumar et al. (011); Liu et al. (01a); Negahban et al. (01); Han et al. (01); Kim and Kwon (01); Shen et al. (01). In the optimization community, researchers have proposed a large variety of computational algorithms including the proximal gradient methods (Nesterov, 1988, 005; NESTEROV, 013; Bec and Teboulle, 009b,a; Zhao and Liu, 01; Liu et al., 015) and coordinate descent methods (Fu, 1998; Friedman et al., 007; Wu and Lange, 008; Friedman et al., 008; Meier et al., 008; Liu et al., 009; Friedman et al., 010; Qin et al., 010; Mazumder et al., 011; Breheny and Huang, 011; Shalev-Shwartz and Tewari, 011; Zhao et al., 014c). Recently, Wang et al. (014) propose the path-following iterative soft shrinage thresholding algorithm (PISTA), which combines the proximal gradient algorithm with path-following optimization scheme. By exploiting the solution sparsity and restricted strong convexity, they show that PISTA attains a linear rate of convergence to a unique sparse local optimum with good statistical properties for solving a large class of sparse nonconvex learning problems. However, though the PISTA has superior theoretical properties, it is empirical performance is in general not as good as some heuristic competing methods such as the path-following coordinate descent algorithm (PCDA) (Tseng and Yun, 009b,a; Lu and Xiao, 013; Friedman et al., 010; Mazumder et al., 011; Zhao et al., 01, 014a). To address this concern, we propose a new computational algorithm named APISTA (Accelerated Path-following Iterative Shrinage Thresholding Algorithm). More specifically, we exploit an additional coordinate descent subroutine to assist PISTA to efficiently decrease the objective value in each iteration. This maes APISTA significantly outperform PISTA in practice. Meanwhile, the coordinate descent subroutine preserves the solution sparsity and restricted strong convexity, therefore APISTA enjoys the same theoretical guarantee as those of PISTA, i.e., APISTA attains a linear rate of convergence to a unique sparse local optimum with good statistical properties. As an application, we apply APISTA to a family of nonconvex optimization problems motivated by estimating semiparametric graphical models (Liu et al., 01b; Zhao and Liu, 014). PISTA allows us to obtain new sparse recovery results on graph estimation consistency which has not been established before. Thorough numerical results are presented to bac up our theory. NOTATIONS: Let v = (v 1,...,v d ) T R d, we define v 1 = j v j, v = j v j, and v = max j v j. We denote the number of nonzero entries in v as v 0 = j 1(v j 0). We define the soft-thresholding operator as S λ (v) = [sign(v j ) ( v j λ)] d j=1 for any λ 0. Given a matrix A Rd d, we use A j = (A 1j,...,A dj ) T to denote the j th column of A, and A = (A 1,...,A d ) T to denote the th row of A. Let Λ max (A) and Λ min (A) denote the largest and smallest eigenvalues of A. Let ψ 1 (A),...,ψ d (A) be the singular values of A, we define the following matrix norms: A F = j A j, A max = max j A j, A 1 = max j A j 1, A = max j ψ j (A), A = max A 1. We denote v \j = (v 1,...,v j 1,v j+1,...,v d ) T

R d 1 as the subvector of v with the j th entry removed. We denote A \i\j as the submatrix of A with the i th row and the j th column removed. We denote A i\j to be the i th row of A with its j th entry removed. Let A {1,...,d}, we use v A to denote a subvector of v by extracting all entries of v with indices in A, and A AA to denote a submatrix of A by extracting all entries of A with both row and column indices in A. Bacground and Problem Setup Let θ = (θ 1,...,θ d )T be a parameter vector to be estimated. We are interested in solving a class of regularized optimization problems in a generic form: min θ R d L(θ) + R λ (θ), (.1) } {{ } F λ (θ) where L(θ) is a smooth loss function and R λ (θ) is a nonsmooth regularization function with a regularization parameter λ..1 Sparsity-inducing Nonconvex Regularization Functions For high dimensional problems, we exploit sparsity-inducing regularization functions, which are usually continuous and decomposable with respect to each coordinate, i.e., R λ (θ) = d j=1 r λ (θ j ). For example, the widely used l 1 norm regularization decomposes as λ θ 1 = d j=1 λ θ j. One drawbac of the l 1 norm is that it incurs large estimation bias when θj is large. This motivates the usage of nonconvex regularizers. Examples include the SCAD (Fan and Li, 001) regularization r λ (θ j ) = λ θ j 1( θ j λ) θ j λβ θ j + λ 1(λ < θ (β 1) j λβ) and MCP (Zhang, 010a) regularization + (β + 1)λ 1( θ j > λβ) for β >, ( r λ (θ j ) = λ θ j θ ) j 1( θ λβ j < λβ) + λ β 1( θ j λβ) for β > 1. Both SCAD and MCP can be written as the sum of an l 1 norm and a concave function H λ (θ), i.e., R λ (θ) = λ θ 1 +H λ (θ). It is easy to see that H λ (θ) = d j=1 h λ (θ j ) is also decomposable with respect to each coordinate. More specifically, the SCAD regularization has h λ (θ j ) = λ θ j θ j λ (β 1) h λ (θ j) = λsign(θ j) θ j β 1 1(λ < θ j λβ) + (β + 1)λ λ θ j 1(λ < θ j λβ) λsign(θ j ) 1( θ j > λβ), 1( θ j > λβ), 3

and the MCP regularization has h λ (θ j ) = θ j β 1( θ j < λβ) + λ β λ θ j 1( θ j λβ), h λ (θ j) = θ j β 1( θ j λβ) λsign(θ j ) 1( θ j > λβ). In general, the concave function h λ ( ) is smooth and symmetric about zero with h λ (0) = 0 and h λ (0) = 0. Its gradient h λ ( ) is monotone decreasing and Lipschitz continuous, i.e., for any θ j > θ j, there exists a constant α 0 such that α(θ j θ j ) h λ (θ j) h λ (θ j ) 0. (.) Moreover, we require h λ (θ j) = λsign(θ j ) if θ j λβ, and h λ (θ j) ( λ,0) if θ j λβ. It is easy to verify that both SCAD and MCP satisfy the above properties. In particular, the SCAD regularization has α = 1/(β 1), and the MCP regularization has α = 1/β. These nonconvex regularization functions have been shown to achieve better asymptotic behavior than the convex l 1 regularization. More technical details can be found in Fan and Li (001); Zhang (010a,b); Zhang and Zhang (01); Fan et al. (014); Xue et al. (01); Wang et al. (014, 013); Liu et al. (014). We present several illustrative examples of the nonconvex regularizers in Figure.1. r ( j) 0.0 0.5 1.0 1.5.0.5 3.0 `1 SCAD MCP h ( j).5.0 1.5 1.0 0.5 0.0 `1 SCAD MCP h 0 ( j) 1.0 0.5 0.0 0.5 1.0 `1 SCAD MCP 3 1 0 1 3 j 3 1 0 1 3 j 3 1 0 1 3 j (a) r λ (θ j ) (b) h λ (θ j ) (c) h λ (θ j) Figure.1: Two illustrative examples of the nonconvex regularization functions: SCAD and MCP. Here we choose λ = 1 and β =.01 for both SCAD and MCP.. Nonconvex Loss Function A motivating application of the method proposed in this paper is sparse transelliptical graphical model estimation (Liu et al., 01b). The transelliptical graphical model is a semiparametric graphical modeling tool for exploring the relationships between a large number of variables. We start with a brief review the transelliptical distribution defined below. 4

Definition.1 (Transelliptical Distribution). Let {f j } d j=1 be a set of strictly increasing univariate functions. Given a positive semidefinite matrix Σ R d d with ran(σ ) = r d and Σ jj = 1 for j = 1,...,d, we say that a d-dimensional random vector X = (X 1,...,X d ) T follows a transelliptical distribution, denoted as X T E d (Σ,ξ,f 1,...,f d ), if X has a stochastic representation (f 1 (X 1 ),...,f d (X d )) T d = ξau, where Σ = AA T, U S r 1 is uniformly distributed on the unit sphere in R r, and ξ 0 is a continuous random variable independent of U. Note that Σ in Definition.1 is not necessarily the correlation matrix of X. To interpret Σ, Liu et al. (01b) provide a latent Gaussian representation for the transelliptical distribution, which implies that the sparsity pattern of Θ = (Σ ) 1 encodes the graph structure of some underlying Gaussian distribution. Since Σ needs to be invertible, we have r = d. To estimate Θ, Liu et al. (01b) suggest to directly plug in the following transformed Kendall s tau estimator into existing gaussian graphical model estimation procedures. Definition. (Transformed Kendall s tau Estimator). Let x 1,...,x n R d be n independent observations of X = (X 1,...,X d ) T, where x i = (x i1,...,x id ) T. The transformed Kendall s tau estimator Ŝ R d d is defined as Ŝ = [Ŝ j ] = [ sin ( π τ j )], where τj is the empirical Kendall s tau statistic between X and X j defined as τ j = ( ) sign (x n(n 1) ij x i j)(x i x i ) i<i if j, 1 otherwise. We then adopt the sparse column inverse operator to estimate the j th column of Θ. In particular, we solve the following regularized quadratic optimization problem (Liu and Luo, 015), 1 min Θ j R d ΘT jŝθ j I T j Θ j + R λ (Θ j ) for j = 1,...,d. (.3) For notational simplicity, we omit the column index j in (.3), and denote Θ j and I j by θ and e respectively. Throughout the rest of this paper, if not specified, we study the following optimization problem for the transelliptical graph estimation 1 min θ R d θt Ŝθ e T θ + R λ (θ). (.4) The quadratic loss function used in (.4) is twice differentiable with L(θ) = Ŝθ e, L(θ) = Ŝ. Since the transformed Kendall s tau estimator is ran-based and could be indefinite (Zhao et al., 014b), the optimization in (.3) may not be convex even if R λ (θ) is a convex. 5

Remar.1. It is worth mentioning that the indefiniteness of Ŝ also maes(.3) unbounded from below, but as will be shown later, our proposed algorithm can still guarantee a unique sparse local solution with optimal statistical properties under suitable solutions. Remar.. To handle the possible nonconvexity, Liu et al. (01b) estimate Θ j graphical model estimation procedure proposed in Cai et al. (011) as follows, based on a min Θ j 1 subject to ŜΘ j I j λ for j = 1,...,d. (.5) Θ j R d (.5) is convex regardless the indefiniteness of Ŝ. But a major disadvantage of (.5) is the computation. Existing solvers can only solve (.5) up to moderate dimensions. We will present more empirical comparison between (.3) and (.5) in our numerical experiments. 3 Method For notational convenience, we rewrite the objective function F λ (θ) as F λ (θ) = L(θ) + H λ (θ) +λ θ 1. } {{ } L λ (θ) We call L λ (θ) the augmented loss function, which is smooth but possibly nonconvex. We first introduce the path-following optimization scheme, which is a multistage optimization framewor and also used in PISTA. 3.1 Path-following Optimization Scheme The path-following optimization scheme solves the regularized optimization problem (.1) using a decreasing sequence of N + 1 regularization parameters {λ K } N K=0, and yields a sequence of N + 1 output solutions { θ {K} } N K=0 from sparse to dense. We set the initial tuning parameter as λ 0 = L(0). By checing the KKT condition of (.1) for λ 0, we have min L λ (0) + λ 0 ξ = min L(0) + H λ (0) + λ 0 ξ = 0, (3.1) ξ 0 1 ξ 0 1 where the second equality comes from ξ 1 and H λ (0) = (h λ (0),h λ (0),...,h λ (0))T = 0 as introduced in.1. Since (3.1) indicates that 0 is a local solution to (.1) for λ 0, we tae the leading output solution as θ {0} = 0. Let η (0,1), we set λ K = ηλ K 1 for K = 1,...,N. We then solve (.1) for the regularization parameter λ K with θ {K 1} as the initial solution, which leads to the next output solution θ {K}. The path-following optimization scheme is illustrated in Algorithm 1. 6

Algorithm 1 Path-following optimization. It solves the problem (.1) using a decreasing sequence of regularization parameters {λ K } N K=0. More specifically, λ 0 = L(0) yields an all zero output solution θ {0} = 0. For K = 1,...,N, we set λ K = ηλ K 1, where η (0,1). We solve (.1) for λ K with θ {K 1} as an initial solution. Note that AISTA is the computational algorithm for obtaining θ K+1 using θ K as the initial solution. L min and { L {K} } N K=0 are corresponding step size parameters. More technical details on AISTA are presented are Algorithm 3. Algorithm: { θ (K) } N K=0 AP IST A({λ K} N K=0 ) Parameter: η, L min Initialize: λ 0 = L(0), θ {0} 0, L {0} L min For: K = 0,...,N 1 λ K+1 ηλ K, { θ {K+1}, L {K+1} } AIST A(λ K+1, θ {K}, L {K} ) End for Output: { θ (K) } N K=0 3. Accelerated Iterative Shrinage Thresholding Algorithm We then explain the accelerated iterative shrinage thresholding (AISTA) subroutine, which solves (.1) in each stage of the path-following optimization scheme. For notational simplicity, we omit the stage index K, and only consider the iteration index m of AISTA. Suppose that AISTA taes some initial solution θ [0] and an initial step size parameter L [0], and we want to solve (.1) with the regularization parameter λ. Then at the m th iteration of AISTA, we already have L [m] and θ [m]. Each iteration of AISTA contains two steps: The first one is the proximal gradient descent iteration, and the second one is the coordinate descent subroutine. (I) Proximal Gradient Descent Iteration: We consider the following quadratic approximation of F λ (θ) at θ = θ [m], Q λ,l [m+1](θ;θ [m] ) = L λ (θ [m] ) + (θ θ [m] ) T L λ (θ [m] ) + L[m+1] θ θ [m] + λ θ 1, where L [m+1] is the step size parameter such that Q λ,l [m+1](θ;θ [m] ) F λ (θ). We then tae a proximal gradient descent iteration and obtain θ [m] by θ [m+0.5] = argmin θ R d Q λ,l [m+1](θ;θ [m] ) = argmin θ R d where θ [m] = θ [m] L λ (θ [m] )/L [m+1]. For notational simplicity, we write L[m+1] θ θ [m] + λ θ 1, (3.) θ [m+0.5] = T λ,l [m+1](θ [m]) ). (3.3) For sparse column inverse operator, we can obtain a closed form solution to (3.) by soft thresholding θ [m+0.5] = S λ/l [m+1]( θ [m+0.5] ). 7

The step size 1/L [m+1] can be obtained by the bactracing line search. In particular, we start with a small enough L [0]. Then in each iteration of the middle loop, we choose the minimum nonnegative integer z such that L [m+1] = z L [m] satisfies F λ (θ [m+0.5] ) Q λ,l [m+1](θ [m+0.5] ;θ [m] ) for m = 0,1,,... (3.4) (II) Coordinate Descent Subroutine: Unlie the proximal gradient algorithm which repeats (3.3) until convergence at each stage of the path-following optimization scheme, AISTA exploits an additional coordinate descent subroutine to further boost the computational performance. More specifically, we define A = {j θ [m+0.5] j = 0} and solve the following optimization problem min θ F λ (θ) subject to θ A = 0 (3.5) using the cyclic coordinate descent algorithm (CCDA) initiated by θ [m+0.5]. For notational simplicity, we omit the stage index K and iteration index m, and only consider the iteration index t of CCDA. Suppose that the CCDA algorithm taes some initial solution θ (0) for solving (.1) with the regularization parameter λ. Without loss of generality, we denote A = {1,..., A }. At the t th iteration, we have θ (t). Then at the (t + 1) th iteration, we conduct the coordinate minimization cyclically over all active coordinates. Let w (t+1,) be an auxiliary solution of the (t + 1) th iteration with the first 1 coordinates updated. For = 1, we have w (t+1,1) = θ (t). We then update the th coordinate to obtain the next auxiliary solution w (t+1,+1). More specifically, let Lλ (θ) be the th entry of L λ (θ). We minimize the objective function with respect to each selected coordinate and eep all other coordinates fixed, w (t+1,+1) = argminl λ (θ ;w (t+1,) \ ) + r λ (θ ). (3.6) θ R Once we obtain w (t+1,+1), we can set w (t+1,+1) \ = w (t+1,) \ to obtain the next auxiliary solution w (t+1,+1). For sparse column inverse operator, let w (t+1,) w (t+1,+1) = e Ŝ T \ w(t+1,) \, we have 1 = argmin θ R Ŝθ + e Ŝ T \ w(t+1,) \ θ e θ + r λ (θ ) = argmin θ R ( 1 θ w (t+1,) ) + rλ (θ ), (3.7) where the last equality comes from the fact Ŝ = 1 for all = 1,...,d. By setting the subgradient of (3.7) equal to zero, we can obtain w (t+1,+1) as follows: For the l 1 norm regularization, we have w (t+1,+1) For the SCAD regularization, we have w (t+1,+1) = w (t+1,) S γλ/(γ 1) ( w (t+1,) ) 1 1/(γ 1) = S λ ( w (t+1,) ). if w (t+1,) γλ, if w (t+1,) [λ,γλ), S λ ( w (t+1,) ) if w (t+1,) < λ. 8

For the MCP regularization, we have w (t+1,+1) = w (t+1,) S λ ( w (t+1,) ) 1 1/γ if w (t+1,) γλ, if w (t+1,) < γλ. When all A coordinate updates in the (t +1) th iteration of CCDA finish, we set θ (t+1) = w (t+1, A +1). We summarize CCDA in Algorithm. Once CCDA terminates, we denote its output solution by θ [m+1], and start the next iteration of AISTA. We summarize AISTA in Algorithm 3. Algorithm The cyclic coordinate descent algorithm (CCDA). The cyclic coordinate descent algorithm cyclically iterates over the support of the initial solution. Without loss of generality, we assume A = {1,..., A }. Algorithm: θ CCDA(λ,θ (0) ). Initialize: t 0, A = supp(θ (0) ) Repeat: w (t+1,1) θ (t) For = 1,..., A w (t+1,+1) End for θ (t+1) w (t+1, A +1), t t + 1 Until convergence θ θ (t) argmin θ R L λ(θ ;w (t+1,) \ ) + r λ (θ ) and w (t+1,+1) \ w (t+1,) \ Algorithm 3 The accelerated iterative shrinage thresholding algorithm (AISTA). Within each iteration, we exploit an additional coordinate descent subroutine to improve the empirical computational performance. Algorithm: { θ, L} AIST A(λ,θ [0],L [0] ) Initialize: m 0 Repeat: z 0 Repeat: L [m+1] z L [m], θ [m+0.5] T λ,ω,l [m+1](θ [m] ), z z + 1 Until: Q λ,l [m+1](θ [m+0.5] ;θ [m] ) F λ (θ [m+1] ) θ [m+1] CCDA(λ,θ [m+0.5] ), m m + 1 Until convergence θ θ [m 0.5], L L [m] Output: { θ, L} 9

Remar 3.1. The bactracing line search procedure in PISTA has been extensively studied in existing optimization literature on the adaptive step size selection (Dennis and Schnabel, 1983; Nocedal and Wright, 006), especially for proximal gradient algorithms (Bec and Teboulle, 009b,a; NESTEROV, 013). Many empirical results have corroborated better computational performance than that using a fix step size. But unlie the classical proximal gradient algorithms, APISTA can efficiently reduce the objective value by the coordinate descent subroutine in each iteration. Therefore we can simply choose a constant step size parameter L such that L sup θ R d Λ max ( L(θ)). (3.8) The step size parameter L in (3.8) guarantees Q λ,l (θ;θ [m] ) F λ (θ) in each iteration of AISTA. For sparse column inverse operator, L(θ) = Ŝ does not depend on θ. Therefore we choose L = Λ max (Ŝ). Our numerical experiments show that choosing a fixed step not only simplifies the implementation, but also attains better empirical computational performance than the bactracing line search. See more details in 5. 3.3 Stopping Criteria Since θ is a local minimum if and only if the KKT condition min ξ θ 1 L λ (θ) + λξ = 0 holds, we terminate AISTA when ω λ (θ [m+0.5] ) = min L λ (θ [m+0.5] ) + λξ ε, (3.9) ξ θ [m+0.5] 1 where ε is the target precision and usually proportional to the regularization parameter. More specifically, given the regularization parameter λ K, we have ε K = δ K λ K for K = 1,...,N, (3.10) where δ K (0,1) is a convergence parameter for the K th stage of the path-following optimization scheme. Moreover, for CCDA, we terminate the iteration when θ (t+1) θ (t) δ 0 λ, (3.11) where δ 0 (0,1) is a convergence parameter. This stopping criterion is natural to the sparse coordinate descent algorithm, since we only need to calculate the value change of each coordinate (not the gradient). We will discuss how to choose δ K s and δ 0 in 4.1. 4 Theory Before we present the computational and statistical theories of APISTA, we introduce some additional assumptions. The first one is about the choice of regularization parameters. 10

Assumption 4.1. Let δ K s and η saitsify η [0.9,1) and max 0 K N δ K δ max = 1/4, where η is the rescaling parameter of the path-following optimization scheme, δ K s are the convergence parameters defined in (3.10), and δ 0 is the convergence parameter defined in (3.11). We have the regularization parameters λ 0 > λ 1... > λ N 8 L(θ ). Assumption 4.1 has been extensively studied in existing literature on high dimensional statistical theory of the regularized M-estimators (Rothman et al., 008; Zhang and Huang, 008; Negahban and Wainwright, 011; Negahban et al., 01). It requires the regularization parameters to be large enough such that irrelevant variables can be eliminated along the solution path. Though L(θ ) cannot be explicitly calculated (θ is unnown), we can exploit concentration inequalities to show that Assumption 4.1 holds with high probability (Ledoux, 005). In particular, we will verify Assumption 4.1 for sparse transellpitical graphical model estimation in Lemma 4.8. Before we proceed with our second assumption, we define the largest and smallest s-sparse eigenvalues of the Hessian matrix of the loss function as follows. Definition 4.1. Given an integer s 1, we define the largest and smallest s-sparse eigenvalues of L(θ) as Largest s-sparse Eigenvalue : ρ + (s) = Smallest s-sparse Eigenvalue : ρ (s) = sup v R d, v 0 s inf v R d, v 0 s v T L(θ)v v, v T L(θ)v v. Moreover, we define ρ (s) = ρ (s) α and ρ + (s) = ρ + (s) for notational simplicity, where α is defined in (.). The next lemma shows the connection between the sparse eigenvalue conditions and restricted strongly convex and smooth conditions. Lemma 4.1. Given ρ (s) > 0, for any θ,θ R d with supp(θ) supp(θ ) s, we have Moreover, if ρ (s) > α, then we have L(θ ) L(θ) + (θ θ) T L(θ) + ρ +(s) θ θ, L(θ ) L(θ) + (θ θ) T L(θ) + ρ (s) θ θ. L λ (θ ) L λ (θ) + (θ θ) T L λ (θ) + ρ +(s) θ θ, L λ (θ ) L λ (θ) + (θ θ) T L λ (θ) + ρ (s) θ θ, 11

and for any ξ θ 1, F λ (θ ) F λ (θ) + ( L λ (θ) + λξ) T (θ θ) + ρ (s) θ θ. The proof of Lemma 4.1 is provided in Wang et al. (014), therefore omitted. We then introduce the second assumption. Assumption 4.. Given θ 0 s, there exists an integer s satisfying where κ = ρ + (s + s)/ ρ (s + s). s (144κ + 50κ) s, ρ + (s + s) < +, and ρ (s + s) > 0, Assumption 4. requires that L λ (θ) satisfies the strong convexity and smoothness when θ is sparse. As will be shown later, APISTA can always guarantee the number of irrelevant coordinates with nonzero values not to exceed s. Therefore the restricted strong convexity is preserved along the solution path. We will verify that Assumption 4. holds with high probability for the transellpitical graphical model estimation in Lemma 4.9. Remar 4. (Step Size Initialization). We tae the initial step size parameter as L min ρ + (1). For sparse column inverse operator, we directly choose L min = ρ + (1) = 1. 4.1 Computational Theory We develop the computational theory of APISTA. For notational simplicity, we define S = {j θj 0} and S = {j θj = 0} for characterizing the the solution sparsity. We first start with the convergence analysis for the cyclic coordinate descent algorithm (CCDA). The next theorem presents its rate of convergence in term of the objective value. Theorem 4.3 (Geometric Rate of Convergence of CCDA). Suppose that Assumption 4. holds. Given a sparse initial solution satisfying θ (0) S 0 s, (3.5) is a strongly convex optimization problem with a unique global minimizer θ. Moreover, for t = 1,..., we have ( F λ (θ (t) (s ) F λ ( θ) + s)ρ +(s ) t + s) (s + s)ρ +(s [F + s) + ρ (1) ρ (s λ (θ (0) ) F λ ( θ)]. + s) The proof of Theorems 4.3 is provided in Appendix A. Theorem 4.3 suggests that when the initial solution is sparse, CCDA essentially solves a strongly convex optimization problem with a unique global minimizer. Consequently we can establish the geometric rate of convergence in term of the objective value for CCDA. We then proceed with the convergence analysis of AISTA. The next theorem presents its theoretical rate of convergence in term of the objective value. Theorem 4.4 (Geometric Rate of Convergence of AISTA). Suppose that Assumptions 4.1 and 4. hold. For any λ λ N, if the initial solution θ [0] satisfies θ [0] S 0 s, ω λ (θ [0] ) λ/, (4.1) 1

then we have θ [m] S 0 s for m = 0.5,1,1.5,,... Moreover, for m = 1,,..., we have ( F λ (θ [m] ) F λ ( θ λ ) 1 1 ) m [F 8κ λ (θ (0) ) F λ ( θ)], where θ λ is a unique sparse local solution to (.1) satisfying ω λ ( θ λ ) = 0 and θ λ S 0 s. The proof of Theorem 4.4 is provided in Appendix B. Theorem 4.4 suggests that all solutions of AISTA are sparse such that the restricted strongly convex and smooth conditions hold for all iterations. Therefore, AISTA attains the geometric rate of convergence in term of the objective value. Theorem 4.4 requires a proper initial solution to satisfy (4.1). This can be verified by the following theorem. Theorem 4.5 (Path-following Optimization Scheme). Suppose that Assumptions 4.1 and 4. hold. Given θ satisfying we have ω λk (θ) λ K /. θ S 0 s and ω λk 1 (θ) δ K 1 λ K 1, (4.) The proof of Theorem 4.5 is provided in Wang et al. (014), therefore omitted. Since θ {0} naturally satisfies (4.) for λ 1, by Theorem 4.5 and induction, we can show that the path-following optimization scheme always guarantees that the output solution of the (K 1) th stage is a proper initial solution for the K th stage, where K = 1,...,N. Eventually, we combine Theorems 4.3 and 4.4 with Theorem 4.5, and establish the global geometric rate of convergence in term of the objective value for APISTA in the next theorem. Theorem 4.6 (Global Geometric Rate of Convergence of APISTA). Suppose that Assumptions 4.1 and 4. hold. Recall that δ 0 and δ K s are defined in 3.3, κ and s are defined in Assumption 4., and α is defined in (.). We have the following results: (1) At the K th stage (K = 1,...,N), the number of coordinate descent iterations within each CCDA is at most C 1 log(c /δ 0 ), where ( C 1 = log 1 (s + s)ρ +(s ) + s) (s + s)ρ +(s + s) + ρ (1) ρ (s + s) 1s and C = ρ (s + s) ρ (1) ; () At the K th stage (K = 1,...,N), the number of the proximal gradient iterations in each AISTA is at most C 3 log(c 4 /δ K ), where ( C 3 = log 1 1 1 ) and C 8κ 4 = 10 κs ; (3) To compute all N + 1 output solutions, the total number of coordinate descent iterations in APISTA is at most C 1 log(c /δ 0 ) 13 N C 3 log(c 4 /δ K ); (4.3) K=1

(4) At the K th stage (K = 1,...,N), we have F λn ( θ {K} ) F λn ( θ λ N ) [1(K < N) + δ K ] 105λ K s ρ (s + s) ; The proof Theorem 4.6 is provided in Appendix C. We then present a more intuitive explanation about Result (3). To secure the generalization performance in practice, we usually tune the regularization parameter over a refined sequence based on cross validation. In particular, we solve (.1) using partial data with high precision for every regularization parameter. If we set δ K = δ opt λ K for K = 1,...N, where δ opt is a very small value (e.g. 10 8 ), then we can rewrite (4.3) as ( ) ( ) ( ( )) C C4 1 NC 1 log C 3 log = O N log, (4.4) δ 0 δ opt where δ 0 is some reasonably large value (e.g. 10 ) defined in 3.3. The iteration complexity in (4.4) depends on N. Once the regularization parameter is selected, we still need to solve (.1) using full data with some regularization sequence. But we only need high precision for the selected regularization parameter (e.g., λ N ), and for K = 1,...,N 1, we only solve (.1) for λ K up to an adequate precision, e.g., δ K = δ 0 for K = 1,...,N 1 and δ N = δ opt λ N. Since 1/δ opt is much larger than N, we can rewrite (4.3) as C 1 log ( C δ 0 )( (N 1)C 3 log ( C4 Now the iteration complexity in (4.5) does not depend on N. δ 0 ) δ opt ( )) ( ( )) C4 1 + C 3 log = O log. (4.5) δ opt δ opt Remar 4.7. To establish computational theories of APISTA with a fixed step size, we only need to slightly modify the proofs of Theorems 4.4 and 4.6 by replacing ρ + (s + s) and ρ + (s + s) by their upper bound L defined in (3.8). Then a global geometric rate of convergence can also be derived, but with a worse constant term. 4. Statistical Theory We then establish the statistical theory of the SCIO estimator obtained by APISTA under transelliptical models. We use Θ and Σ to denote the true latent precision and covariance matrices. We assume that Θ belongs to the following class of sparse, positive definite, and symmetric matrices: { U ψmax,ψ min (M,s ) = Θ R d d Θ = Θ T, max j Θ j 0 s, Θ 1 M, 0 < ψ 1 max Λ min (Θ) Λ max (Θ) ψ 1 min < }, where ψ max and ψ min are positive constants, and do not scale with (M,s,n,d). Since Σ = (Θ ) 1, we also have ψ min Λ min (Σ ) Λ max (Σ ) ψ max. We first verify Assumptions 4.1 and 4. in the next two lemmas for transelliptical models. 14

Lemma 4.8. Suppose that X T E d (Σ,ξ,{f j } d j=1 ). Given λ N = 8 πm logd/n, we have P(λ N 8 L(θ ) ) 1 1 d. The proof of Lemma 4.8 is provided in Appendix D. Lemma 4.8 guarantees that the selected regularization parameter λ N satisfies Assumption 4.1 with high probability. Lemma 4.9. Suppose that X T E d (Σ,ξ,{f j } d j=1 ). Given α = ψ min/, there exist universal positive constants c 1 and c such that for n 4ψ 1 min c (1 + c 1 )s logd, with probability at least 1 /d, we have s = c 1 s (144κ + 50κ)s, ρ (s + s) ψ min 4, ρ +(s + s) 5ψ max, 4 where κ is defined in Assumption 4.. The proof of Lemma 4.9 is provided in Appendix E. Lemma 4.9 guarantees that if the Lipschitz constant of h λ defined in (.) satisfies α = ψ min/, then the transformed Kendall s tau estimator Ŝ = L(θ) satisfies Assumption 4. with high probability. Remar 4.10. Since Assumptions 4.1 and 4. have been verified, by Theorem 4.6, we now that APISTA attains the geometric rate of convergence to a unique sparse local solution to (.3) in term of the objective value with high probability. Recall that we use θ to denote Θ j in (.4), by solving (.3) with respect to all d columns, we obtain Θ {N} ] [ ] = and Θ λn = Θ λ N 1,...,Θ λ N d, where Θ λ N j denotes the output solution of [ Θ{N} 1,..., Θ {N} d APISTA corresponding to λ N for the j th column (j = 1,..d), and Θ λ N j to denote the unique sparse local solution corresponding to λ N for the j th column (j = 1,..d), which APISTA converges to. We then present concrete rates of convergence of the estimator obtained by APISTA under the matrix l 1 and Frobenius norms in the following theorem. Theorem 4.11. [Parameter Estimation] Suppose that X T E d (Σ,ξ,{f j } d j=1 ), and α = ψ min/. For n 4ψmin 1 c (1 + c 1 )s logd, given λ N = 8 πm logd/n, we have ( Θ {N} logd Θ 1 = O P Ms n, 1 d Θ {N} M Θ F = O s ) logd P. n The proof of (4.11) is provided in Appendix F. The results in Theorem 4.11 show that the SCIO estimator obtained by APISTA achieves the same rates of convergence as those for subguassian distributions (Liu and Luo, 015). Moreover, when using the nonconvex regularization such as MCP and SCAD, we can achieve graph estimation consistency under the following assumption. Assumption 4.3. Suppose that X T E d (Σ,ξ,{f j } d j=1 ). Define E = {(,j) Θ j 0} as the support of Θ. There exists some universal constant c 3 such that s min (,j) E Θ j c 3M logd. n 15

Assumption 4.3 is a sufficient condition for sparse column inverse operator to achieve graph estimation consistency in high dimensions for transelliptical models. The violation of Assumption 4.3 may result in underselection of the nonzero entries in Θ. The next theorem shows that, with high probability, Θ λn and the oracle solution Θ o are identical. More specifically, let S j = supp(θ j ) for j = 1,...,d, Θ o = [ Θo 1,..., Θ o d] defined as follows, Θ o 1 S j j = argmin ΘT S j jŝs j S j Θ Sj j I T S j j Θ S j j and Θ o S Θ Sj j R S j j j = 0 for j = 1,...,d. (4.6) Theorem 4.1. [Graph Estimation] Suppose that X T E d (Σ,ξ,{f j } d j=1 ), α = ψ min/, and Assumption 4.3 holds. There exists a universal constant c 4 such that n 4ψ 1 min c (1 + c 1 )s logd, if we choose λ N = c 4 πm s logd/n, then we have P(Θ λn = Θ o ) 1 3 d. The proof of Theorem 4.1 is provided in Appendix G. Since Θ o shares the same support with Θ, Theorem 4.1 guarantees that the SCIO estimator obtained by APISTA can perfectly recover E with high probability. To the best of our nowledge, Theorem 4.1 is the first graph estimation consistency result for transelliptical models without any post-processing procedure (e.g. thresholding). Remar 4.13. In Theorem (4.1), we choose λ N = c 4 πm logd/n, which is different from the selected regularization parameter in Assumption 4.8. But as long as we have c 4 s 8, which is not an issue under the high dimensional scaling M,s,n,d and Ms logd/n 0, λ N 8 L(θ ) still holds with high probability. Therefore all computational theories in 4.1 hold for Θ λn in Theorem 4.1. 5 Numerical Experiments In this section, we study the computational and statistical performance of APISTA method through numerical experiments on sparse transelliptical graphical model estimation. All experiments are conducted on a personal computer with Intel Core i5 3.3 GHz and 16GB memory. All programs are coded in double precision C, called from R. The computation are optimized by exploiting the sparseness of vector and matrices. Thus we can gain a significant speedup in vector and matrix manipulations (e.g. calculating the gradient and evaluating the objective value). We choose the MCP regularization with varying β s for all simulations. 16

5.1 Simulated Data We consider the chain and Erdös-Rényi graph generation schemes with varying d = 00, 400, and 800 to obtain the latent precision matrices: Chain. Each node is assigned a coordinate j for j = 1,...,d. Two nodes are connected by an edge whenever the corresponding points are at distance no more than 1. Erdös-Rényi. We set an edge between each pair of nodes with probability 1/d, independently of the other edges. Two illustrative examples are presented in Figure 5.1. Let D be the adjacency matrix of the generated graph, and M be the rescaling operator that converts a symmetric positive semidefinite matrix to a correlation matrix. We calculate Σ = M [( D + (1 Λ min (D))I) 1 ]. We use Σ as the covariance matrix to generate n = 60logd independent observations from a multivariate t-distribution with mean 0 and degrees of freedom 3. We then adopt the power transformation g(t) = t 5 to convert to the t-distributed data to the transelliptical data. Note that the corresponding latent precision matrix is Ω = (Σ ) 1. We compare the following five computational methods: (1) APISTA: The computational algorithm proposed in 3. () F-APISTA: APISTA without the bactracing line search (using a fixed step size instead). (3) PISTA: The pathwise iterative shrinage thresholding algoritm proposed in Wang et al. (014). (4) CLIME: The sparse latent precision matrix estimation method proposed in Liu et al. (01b), which solves (.5) by the ADMM method (Alternating Direction Method of Multipliers, Li et al. (015); Liu et al. (014)). (5) SCIO(P): The SCIO estimator based on the positive semidefinite projection method proposed in Zhao et al. (014b). More specifically, we first project the possibly indefinite Kendall s tau matrix into the cone of all positive semidefinite matrices. Then we plug the obtained replacement into (.3), and solve it by the coordinate descent method proposed in Liu and Luo (015). Note that (4) and (5) have theoretical guarantees only when the l 1 norm regularization is applied. For (1)-(3), we set δ 0 = δ K = 10 5 for K = 1,...,N. We first compare the statistical performance in parameter estimation and graph estimation of all methods. To meet this end, we generate a validation set of the same size as the training set. 17

(a) Chain (b) Erdös-Rényi Figure 5.1: Different graph patterns. To ease the visualization, we only present graphs with d = 00. We use the regularization sequence with N = 100 and λ N = 0.5 logd/n 0.0645. The optimal regularization parameter is selected by λ = argmin Θ λ S I max, λ {λ 1,...,λ N } where Θ λ denotes the estimated latent precision matrix using the training set with the regularization parameter λ, and S denotes the estimated latent covariance matrix using the validation set. We repeat the simulation for 100 times, and summarize the averaged results in Tables 5.1 and 5.. For all settings, we set δ 0 = δ K = 10 5. We also vary β of the MCP regularization from 100 to 0/19, thus the corresponding α varies from 0.01 to 0.95. The parameter estimation performance is evaluated by the difference between the obtained estimator and the true latent prediction matrix under the Forbenius and matrix l 1 norms. The graph estimation performance is evaluated by the true positive rate (T. P. R.) and false positive rate (F. P. R.) defined as follows, j 1( Θ λ j 0) 1(Θ j 0) T. P. R. = j 1(Θ j 0), F. P. R. = j 1( Θ λ j 0) 1(Θ j = 0) j 1(Θ j = 0). Since the convergence of PISTA is very slow when α is large, we only present its results for α = 0.. APISTA and F-APISTA can wor for larger α s. Therefore they effectively reduces the estimation bias to attain the best statistical performance in both parameter estimation and graph estimation among all estimators. The SCIO(P) and CLIME methods only use l 1 norm without any bias reduction, their performance is worse than the other competitors. Moreover, due to the poor scalability of their solvers, SCIO(P) and CLIME fail to output valid results within 10 hours when d = 800. 18

Table 5.1: Quantitive comparison among different estimators on the chain model. Since APISTA and F-APISTA can output valid results for large α s, their estimator attains better performance than other competitors. The SCIO(P) and CLIME estimators use the l 1 norm regularization with no bias reduction. Thus their performance is worse than the other competitors in both parameter estimation and graph estimation. Method d Θ Θ F Θ Θ 1 T. P. R. F. P. R. α PISTA APISTA F-APISTA SCIO(P) CLIME 00 4.111(0.7856) 1.0517(0.1141) 1.0000(0.0000) 0.0048(0.0079) 0.0 400 6.4507(0.906) 1.0756(0.0717) 1.0000(0.0000) 0.0007(0.0004) 0.0 800 8.640(1.1456) 1.0434(0.0673) 1.0000(0.0000) 0.0003(0.0006) 0.0 00.516(0.677) 0.7665(0.1583) 0.9993(0.001) 0.0001(0.0001) 0.95 400 3.3664(0.735) 0.898(0.0986) 1.0000(0.0000) 0.000(0.0000) 0.67 800 5.044(0.7984) 0.931(0.16) 1.0000(0.0000) 0.000(0.0004) 0.50 00.5163(0.670) 0.7658(0.1559) 0.9994(0.0015) 0.0001(0.000) 0.95 400 3.369(0.70) 0.853(0.0959) 1.0000(0.0000) 0.000(0.0000) 0.67 800 5.037(0.7963) 0.9373(0.189) 1.0000(0.0000) 0.000(0.0005) 0.50 00 6.181(1.94) 1.45(0.0777) 1.0000(0.0000) 0.0165(0.00) 0.00 400 8.9991(0.9894) 1.55(0.0785) 1.0000(0.0000) 0.0058(0.0047) 0.00 00 6.4771(0.8617) 1.187(0.0358) 1.0000(0.0000) 0.016(0.0043) 0.00 400 9.11(0.9997) 1.177(0.069) 1.0000(0.0000) 0.0043(0.003) 0.00 We then compare the computational performance of all methods. We use a regularization sequence with N = 50, and λ N is proper selected such that the graphs obtained by all methods have approximately the same number of edges for each regularization parameter. In particular, the obtained graphs corresponding to λ N have approximately 0.1 d(d 1)/ edges. To mae a fair comparison, we choose the l 1 norm regularization for all methods. We repeat the simulation for 100 times, and the timing results are summarized in Tables 5.3 and 5.4. We see that F-APISTA method is up to 10 times faster than PISTA algorithm, and APISTA is up to 5 times after than PISTA. SCIO(P) and CLIME are much slower than the other three competitors. 5. Real Data We present a real data example to demonstrate the usefulness of the transelliptical graph obtained by the sparse column inverse operator (based on the transformed Kendall s tau matrix). We acquire closing prices from all stocs of the S&P 500 for all the days that the maret was open between January 1, 003 and January 1, 005, which results in 504 samples for the 45 stocs. We transform the dataset by calculating the log-ratio of the price at time t + 1 to price at time t. The 19

Table 5.: Quantitive comparison among different estimators on the Erdös-Rényi model. Since A-PISTA and F-APISTA can output valid results for large α s, their estimators attains better performance than other competitors. The SCIO(P) and CLIME estimators use the l 1 norm regularization with no bias reduction. Thus their performance is worse than the other competitors in both parameter estimation and graph estimation. Method d Θ Θ F Θ Θ 1 T. P. R. F. P. R. α PISTA APISTA F-APISTA SCIO(P) CLIME 00 3.647(0.135) 1.6807(0.675) 1.0000(0.0000) 0.0587(0.0013) 0.0 400 4.5609(0.7666).113(0.3358) 1.0000(0.0000) 0.095(0.0091) 0.0 800 5.0751(0.383).5718(0.86) 1.0000(0.0000) 0.0099(0.000) 0.0 00.888(0.1141) 1.1644(0.343) 1.0000(0.0000) 0.0193(0.0005) 0.33 400 3.06(0.733) 1.4974(0.778) 1.0000(0.0000) 0.0067(0.0100) 0.33 800 4.099(0.186) 1.6347(0.03) 1.0000(0.0000) 0.0036(0.0008) 0.50 00.890(0.1161) 1.1647(0.390) 1.0000(0.0000) 0.0197(0.0007) 0.33 400 3.51(0.70) 1.498(0.731) 1.0000(0.0000) 0.0060(0.010) 0.33 800 4.0984(0.1891) 1.6397(0.096) 1.0000(0.0000) 0.0034(0.0009) 0.50 00 3.477(0.5405) 1.513(0.33) 1.0000(0.0000) 0.0618(0.0170) 0.00 400 5.7144(0.8158) 1.9057(0.933) 0.9994(0.0017) 0.0341(0.0145) 0.00 00 3.697(0.6103) 1.4876(0.855) 1.0000(0.0000) 0.0581(0.0159) 0.00 400 5.906(0.8385) 1.846(0.817) 1.0000(0.0000) 0.030(0.011) 0.00 Table 5.3: Quantitive comparison of computational performance on the chain model (in seconds). We see that the F-APISTA method attains the best timing performance among all methods. The SCIO(P) and CLIME methods are much slower than the other three methods. d PISTA APISTA F-APISTA SCIO(P) CLIME 00 0.834(0.048) 0.693(0.0031) 0.1013(0.00).657(0.153) 8.593(0.5396) 400 3.878(0.0696) 1.103(0.0368) 0.4559(0.0308) 5.451(.575) 48.35(5.3494) 800 30.014(0.3514) 6.5970(0.338).483(0.605) 315.87(34.638) 460.1(45.11) 45 stocs are categorized into 10 Global Industry Classification Standard (GICS) sectors. We adopt the stability graphs obtained by the following procedure (Meinshausen and Bühlmann, 010; Liu et al., 010): (1) Calculate the graph path using all samples, and choose the regularization parameter at the 0

Table 5.4: Quantitive comparison of computational performance on the Erdös-Rényi model (in seconds). We see that the F-APISTA method attains the best timing performance among all methods. The SCIO(P) and CLIME methods are much slower than the other three methods. d PISTA APISTA F-APISTA SCIO(P) CLIME 00 0.5401(0.048) 0.048(0.0056) 0.1063(0.0110).71(0.13558) 7.135(0.7891) 400 3.0501(0.089) 0.998(0.0453) 0.4555(0.0071) 6.140(.1503) 45.160(4.906) 800 8.581(0.3517) 6.8417(0.7543).7037(0.145) 33.90(30.115) 44.57(50.978) (a) Transelliptical Graphs by SCIO (L) and CLIME (R) (b) Gaussian Graph by SCIO Figure 5.: Stoc Graphs. We see that both transelliptical graphs reveal more refined structures than the Gaussian graph. sparsity level 0.1; () Randomly choose 50% of all samples without replacement using the regularization parameter chosen in (1); (3) Repeat () 100 times and retain the edges that appear with frequencies no less than 95%. We choose the sparsity level 0.1 in (1) and subsampling ratio 50% in () based on two criteria: The resulting graphs need to be sparse to ease visualization, interpretation, and computation; The resulting graphs need to be stable. We then present the obtained stability graphs in Figure 5.. The nodes are colored according to the GICS sector of the corresponding stoc. We highlight a region in the transelliptical graph obtained by the SCIO method and by color coding we see that the nodes in this region belong to the same sector of the maret. A similar pattern is also found in the transelliptical graph obtained by the CLIME method. In contrast, this region is shown to be sparse in the Gaussian graph obtained by the SCIO method (based on the Pearson correlation matrix). Therefore we can see that the SCIO method is also capable of generating refined structures as the 1

CLIME method when estimating the transelliptical graph. 6 Discussions We compare F-APISTA with a closely related algorithm the path-following coordinate descent algorithm (PCDA 1 ) in timing performance. In particular, we give a failure example of PCDA for solving sparse linear regression. Let X R n d denote design matrix and y R n denote the response vector. We solve the following regularized optimization problem, 1 min θ n y Xθ + R λ(θ). We generate each row of the design matrix X i from a d-variate Gaussian distribution with mean 0 and covariance matrix Σ R d d, where Σ j = 0.75 if j and Σ = 1 for all j, = 1,...,d. We then normalize each column of the design matrix X j such that X j = n. The response vector is generated from the linear model y = Xθ +ɛ, where θ R d is the regression coefficient vector, and ɛ is generated from a n-variate Gaussian distribution with mean 0 and covariance matrix I. We set n = 60 and d = 1000. We set the coefficient vector as θ50 = 3, θ 500 =, θ 750 = 1.5, and θ j = 0 for all j 50,500,750. We then set α = 0.95, N = 100, λ N = 0.5 logd/n, and δ c = δ K = 10 5. We then generate a validation set using the same design matrix as the training set for the regularization selection. We denote the response vector of the validation set as ỹ R n. Let θ λ denote the obtained estimator using the regularization parameter λ. We then choose the optimal regularization parameter λ by λ = argmin ỹ X θ λ. λ {λ 1,...,λ N } We repeat 100 simulations, and summarize the average results in Table 6. We see that F-APISTA and PCDA attain similar timing results. But PCDA achieves worse statistical performance than F- APISTA in both support recovery and parameter estimation. This is because PCDA has no control over the solution sparsity. The overselection irrelevant variables compromise the restricted strong convexity, and mae PCDA attain some local optima with poor statistical properties. References Banerjee, O., El Ghaoui, L. and d Aspremont, A. (008). Model selection through sparse maximum lielihood estimation for multivariate gaussian or binary data. The Journal of Machine Learning Research 9 485 516. Bec, A. and Teboulle, M. (009a). Fast gradient-based algorithms for constrained total variation image denoising and deblurring problems. Image Processing, IEEE Transactions on 18 419 434. 1 In our numerical experiments, PCDA is implemented by the R pacage ncvreg.

Table 6.1: Quantitative comparison between F-APISTA and PCDA. We see that F-APISTA and PCDA attain similar timing results. But PCDA achieves worse statistical performance than F- APISTA in both support recovery and parameter estimation. Method θ θ θ S 0 θ S c 0 Correct Selection Timing F-APISTA 0.8001(0.9089).801(0.513) 0.890(.11) 667/1000 0.0181(0.005) PCDA 1.175(1.539).655(0.7051) 1.644(3.016) 517/1000 0.0195(0.001) Bec, A. and Teboulle, M. (009b). A fast iterative shrinage-thresholding algorithm for linear inverse problems. SIAM Journal on Imaging Sciences 183 0. Breheny, P. and Huang, J. (011). Coordinate descent algorithms for nonconvex penalized regression, with applications to biological feature selection. The Annals of Applied Statistics 5 3 53. Cai, T., Liu, W. and Luo, X. (011). A constrained l 1 minimization approach to sparse precision matrix estimation. Journal of the American Statistical Association 106 594 607. Dennis, J. J. E. and Schnabel, R. B. (1983). Numerical methods for unconstrained optimization and nonlinear equations, vol. 16. SIAM. Fan, J., Feng, Y. and Tong, X. (01). A road to classification in high dimensional space: the regularized optimal affine discriminant. Journal of the Royal Statistical Society: Series B 74 745 771. Fan, J., Feng, Y. and Wu, Y. (009). Networ exploration via the adaptive lasso and scad penalties. The Annals of Applied Statistics 3 51 541. Fan, J. and Li, R. (001). Variable selection via nonconcave penalized lielihood and its oracle properties. Journal of the American Statistical Association 96 1348 1360. Fan, J., Xue, L. and Zou, H. (014). Strong oracle optimality of folded concave penalized estimation. The Annals of Statistics 4 819 849. Friedman, J., Hastie, T., Höfling, H. and Tibshirani, R. (007). Pathwise coordinate optimization. The Annals of Applied Statistics 1 30 33. Friedman, J., Hastie, T. and Tibshirani, R. (008). Sparse inverse covariance estimation with the graphical lasso. Biostatistics 9 43 441. Friedman, J., Hastie, T. and Tibshirani, R. (010). Regularization paths for generalized linear models via coordinate descent. Journal of statistical software 33 1 13. 3