Optimal Regularized Dual Averaging Methods for Stochastic Optimization

Similar documents
Learning with stochastic proximal gradient

A Sparsity Preserving Stochastic Gradient Method for Composite Optimization

Stochastic Optimization

Adaptive Subgradient Methods for Online Learning and Stochastic Optimization John Duchi, Elad Hanzan, Yoram Singer

Composite Objective Mirror Descent

Stochastic and online algorithms

Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization

On the Generalization Ability of Online Strongly Convex Programming Algorithms

Large-scale Stochastic Optimization

Efficient Methods for Stochastic Composite Optimization

Optimization methods

Block stochastic gradient update method

arxiv: v4 [math.oc] 5 Jan 2016

Convex Optimization Lecture 16

Stochastic gradient methods for machine learning

Relative-Continuity for Non-Lipschitz Non-Smooth Convex Optimization using Stochastic (or Deterministic) Mirror Descent

Big Data Analytics: Optimization and Randomization

Proximal Minimization by Incremental Surrogate Optimization (MISO)

Statistical Machine Learning II Spring 2017, Learning Theory, Lecture 4

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

Optimization methods

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence:

Adaptive Online Gradient Descent

arxiv: v1 [math.oc] 1 Jul 2016

Adaptive Online Learning in Dynamic Environments

Convergence Rates for Deterministic and Stochastic Subgradient Methods Without Lipschitz Continuity

ECS289: Scalable Machine Learning

STA141C: Big Data & High Performance Statistical Computing

Fast proximal gradient methods

Stochastic Dual Coordinate Ascent Methods for Regularized Loss Minimization

A Unified Approach to Proximal Algorithms using Bregman Distance

Stochastic Gradient Descent with Only One Projection

Gradient Sliding for Composite Optimization

6. Proximal gradient method

Stochastic Gradient Descent. Ryan Tibshirani Convex Optimization

SVRG++ with Non-uniform Sampling

Stochastic gradient methods for machine learning

Accelerated Gradient Method for Multi-Task Sparse Learning Problem

Coordinate Descent and Ascent Methods

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices

Stochastic Optimization with Variance Reduction for Infinite Datasets with Finite Sum Structure

A Greedy Framework for First-Order Optimization

Logarithmic Regret Algorithms for Strongly Convex Repeated Games

Lecture 9: September 28

6. Proximal gradient method

Sparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda

Accelerating Online Convex Optimization via Adaptive Prediction

Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method

Stochastic Optimization Part I: Convex analysis and online stochastic optimization

Tutorial: PART 2. Online Convex Optimization, A Game- Theoretic Approach to Learning

Lasso: Algorithms and Extensions

Modern Stochastic Methods. Ryan Tibshirani (notes by Sashank Reddi and Ryan Tibshirani) Convex Optimization

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization /

Lecture 16: Perceptron and Exponential Weights Algorithm

Minimizing Finite Sums with the Stochastic Average Gradient Algorithm

Stochastic Composition Optimization

SADAGRAD: Strongly Adaptive Stochastic Gradient Methods

One Mirror Descent Algorithm for Convex Constrained Optimization Problems with Non-Standard Growth Properties

Stochastic Semi-Proximal Mirror-Prox

Summary and discussion of: Dropout Training as Adaptive Regularization

Computational and Statistical Learning Theory

On Stochastic Primal-Dual Hybrid Gradient Approach for Compositely Regularized Minimization

arxiv: v2 [cs.lg] 10 Oct 2018

Machine Learning. Support Vector Machines. Fabio Vandin November 20, 2017

Adaptive Gradient Methods AdaGrad / Adam. Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade

Contents. 1 Introduction. 1.1 History of Optimization ALG-ML SEMINAR LISSA: LINEAR TIME SECOND-ORDER STOCHASTIC ALGORITHM FEBRUARY 23, 2016

The Frank-Wolfe Algorithm:

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

Lower and Upper Bounds on the Generalization of Stochastic Exponentially Concave Optimization

Selected Topics in Optimization. Some slides borrowed from

Coordinate descent methods

Pavel Dvurechensky Alexander Gasnikov Alexander Tiurin. July 26, 2017

Stochastic Subgradient Method

An Asynchronous Mini-Batch Algorithm for. Regularized Stochastic Optimization

1 Regression with High Dimensional Data

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Lecture 23: November 21

ARock: an algorithmic framework for asynchronous parallel coordinate updates

Barzilai-Borwein Step Size for Stochastic Gradient Descent

Minimizing the Difference of L 1 and L 2 Norms with Applications

Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections

A Modular Analysis of Adaptive (Non-)Convex Optimization: Optimism, Composite Objectives, and Variational Bounds

1 Sparsity and l 1 relaxation

Linearly-Convergent Stochastic-Gradient Methods

Stochastic Proximal Gradient Algorithm

Lecture 25: November 27

Accelerated Training of Max-Margin Markov Networks with Kernels

DATA MINING AND MACHINE LEARNING

Lecture: Adaptive Filtering

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Finite-sum Composition Optimization via Variance Reduced Gradient Descent

Randomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity

Regret bounded by gradual variation for online convex optimization

Cubic regularization of Newton s method for convex problems with constraints

arxiv: v6 [math.oc] 24 Sep 2018

Math 273a: Optimization Subgradient Methods

Agenda. Fast proximal gradient methods. 1 Accelerated first-order methods. 2 Auxiliary sequences. 3 Convergence analysis. 4 Numerical examples

An Efficient Proximal Gradient Method for General Structured Sparse Learning

MATH 680 Fall November 27, Homework 3

Transcription:

Optimal Regularized Dual Averaging Methods for Stochastic Optimization Xi Chen Machine Learning Department Carnegie Mellon University xichen@cs.cmu.edu Qihang Lin Javier Peña Tepper School of Business Carnegie Mellon University {qihangl,jfp}@andrew.cmu.edu Abstract This paper considers a wide spectrum of regularized stochastic optimization problems where both the loss function and regularizer can be non-smooth. We develop a novel algorithm based on the regularized dual averaging RDA method, that can simultaneously achieve the optimal convergence rates for both convex and strongly convex loss. In particular, for strongly convex loss, it achieves the optimal rate of O + for iterations, which improves the rate O log for previous regularized dual averaging algorithms. In addition, our method constructs the final solution directly from the proximal mapping instead of averaging of all previous iterates. For widely used sparsity-inducing regularizers e.g., l -norm, it has the advantage of encouraging sparser solutions. We further develop a multistage extension using the proposed algorithm as a subroutine, which achieves the uniformly-optimal rate O + exp{ } for strongly convex loss. Introduction Many risk minimization problems in machine learning can be formulated into a regularized stochastic optimization problem of the following form: min x X {ϕx := fx + hx}. Here, the set of feasible solutions X is a convex set in R n, which is endowed with a norm and the dual norm. The regularizer hx is assumed to be convex, but could be non-differentiable. Popular examples of hx include l -norm and related sparsity-inducing regularizers. The loss function fx takes the form: fx := E ξ F x, ξ = F x, ξdp ξ, where ξ is a random vector with the distribution P. In typical regression or classification tasks, ξ is the input and response or class label pair. We assume that for every random vector ξ, F x, ξ is a convex and continuous function in x X. Therefore, fx is also convex. Furthermore, we assume that there exist constants L 0, M 0 and µ 0 such that µ x y fy fx y x, f x L x y + M x y, x, y X, where f x fx, the subdifferential of f. We note that this assumption allows us to adopt a wide class of loss functions. For example, if fx is smooth and its gradient f x = fx is Lipschitz continuous, we have L > 0 and M = 0 e.g., squared or logistic loss. If fx is nonsmooth but Lipschitz continuous, we have L = 0 and M > 0 e.g., hinge loss. If µ > 0, fx is strongly convex and µ is the so-called strong convexity parameter. In general, the optimization problem in Eq. is challenging since the integration in fx is computationally intractable for high-dimensional P. In many learning problems, we do not even know the underlying distribution P but can only generate i.i.d. samples ξ from P. A traditional approach is to

consider empirical loss minimization problem where the expectation in fx is replaced by its empirical average on a set of training samples {ξ,..., ξ m }: f emp x := m m i= F x, ξ i. However, for modern data-intensive applications, minimization of empirical loss with an off-line optimization solver could suffer from very poor scalability. In the past few years, many stochastic subgradient methods [6, 5, 8,, 4, 0, 9,, 7, 8] have been developed to directly solve the stochastic optimization problem in Eq., which enjoy low periteration complexity and the capability of scaling up to very large data sets. In particular, at the t-th iteration with the current iterate x t, these methods randomly draw a sample ξ t from P ; then compute the so-called stochastic subgradient Gx t, ξ t x F x t, ξ t where x F x t, ξ t denotes the subdifferential of F x, ξ t with respect to x at x t ; and update x t using Gx t, ξ t. These algorithms fall into the class of stochastic approximation methods. Recently, Xiao [] proposed the regularized dual averaging RDA method and its accelerated version AC-RDA based on esterov s primal-dual method [7]. Instead of only utilizing a single stochastic subgradient Gx t, ξ t of the current iteration, it updates the parameter vector using the average of all past stochastic subgradients {Gx i, ξ i } t i= and hence leads to improved empirical performances. In this paper, we propose a novel regularized dual averaging method, called optimal RDA or ORDA, which achieves the optimal expected convergence rate of E[ϕ x ϕx ], where x is the solution from ORDA and x is the optimal solution of Eq.. As compared to previous dual averaging methods, it has three main advantages:. For strongly convex fx, ORDA improves the convergence rate of stochastic dual averaging methods O log log O [7, ] to an optimal rate O +M + L O, where is the variance of the stochastic subgradient, is the number of iterations, and the parameters µ, M and L of fx are defined in Eq... ORDA is a self-adaptive and optimal algorithm for solving both convex and strongly convex fx with the strong convexity parameter µ as an input. When µ = 0, ORDA reduces to a variant of AC-RDA in [] with the optimal rate for solving convex fx. Furthermore, our analysis allows fx to be non-smooth while AC-RDA requires the smoothness of fx. For strongly convex fx with µ > 0, our algorithm achieves the optimal rate of while AC-RDA does not utilize the advantage of strong convexity. 3. Existing RDA methods [] and many other stochastic gradient methods e.g., [4, 0] can only show the convergence rate for the averaged iterates: x = t= ϱ tx t / t= ϱ t, where the {ϱ t } are nonnegative weights. However, in general, the average iterates x cannot keep the structure that the regularizer tends to enforce e.g., sparsity, low-rank, etc. For example, when hx is a sparsity-inducing regularizer l -norm, although x t computed from proximal mapping will be sparse as t goes large, the averaged solution could be non-sparse. In contrast, our method directly generates the final solution from the proximal mapping, which leads to sparser solutions. In addition to the rate of convergence, we also provide high probability bounds on the error of objective values. Utilizing a technical lemma from [3], we could show the same high probability bound as in RDA [] but under a weaker assumption. Furthermore, using ORDA as a subroutine, we develop the multi-stage ORDA which obtains the convergence rate of O +M + exp{ µ/l} for strongly convex fx. Recall that ORDA has the rate O +M + L for strongly convex fx. The rate of muli-stage ORDA improves the second term in the rate of ORDA from O L to O exp{ µ/l} and achieves the socalled uniformly-optimal rate [5]. Although the improvement is on the non-dominating term, multi-stage ORDA is an optimal algorithm for both stochastic and deterministic optimization. In particular, for deterministic strongly convex and smooth fx M = 0, one can use the same algorithm but only replaces the stochastic subgradient Gx, ξ by the deterministic gradient fx. Then, the variance of the stochastic subgradient = 0. ow the term +M in the rate equals to 0 and multi-stage ORDA becomes an optimal deterministic solver with the exponential rate

Algorithm Optimal Regularized Dual Averaging Method: ORDAx 0,, Γ, c Input Parameters: Starting point x 0 X, the number of iterations, constants Γ L and c 0. Parameters for fx: Constants L, M and µ for fx in Eq. and set µ = µ/τ. Initialization: Set θ t = t+ ; ν t = t+ ; γ t = ct + 3/ + τγ; z 0 = x 0. Iterate for t = 0,,,..., :. y t = θ tµ+θ t γ t θ t γ t+ θ t µ x t + θ tθ t µ+θ 3 t γ t θ t γ t+ θ t µ z t. Sample ξ t from the distribution P ξ and compute the stochastic subgradient Gy t, ξ t. 3. g t = θ t ν t t Gy i,ξ i ν i 4. z t+ = arg min x X { x, g t + hx + θ t ν t t 5. x t+ = arg min x X { x, Gy t, ξ t + hx + Output: x + µ τθ t } ν i + θ t ν t γ t+ x, x 0 } x, y t µ x,y i O exp{ µ/l}. This is the reason why such a rate is uniformly-optimal, i.e., optimal with respect to both stochastic and deterministic optimization. Preliminary and otations In the framework of first-order stochastic optimization, the only available information of fx is the stochastic subgradient. Formally speaking, stochastic subgradient of fx at x, Gx, ξ, is a vectorvalued function such that E ξ Gx, ξ = f x fx. Following the existing literature, a standard assumption on Gx, ξ is made throughout the paper : there exists a constant such that x X, + γt τ E ξ Gx, ξ f x. 3 A key updating step in dual averaging methods, the proximal mapping, utilizes the Bregman divergence. Let ωx : X R be a strongly convex and differentiable function, the Bregman divergence associated with ωx is defined as: x, y := ωx ωy ωy, x y. 4 One typical and simple example is ωx = x together with x, y = x y. One may refer to [] for more examples. We can always scale ωx so that x, y x y for all x, y X. Following the assumption in [0]: we assume that x, y grows quadratically with the parameter τ >, i.e., x, y τ x y with τ > for all x, y X. In fact, we could simply choose ωx with a τ-lipschitz continuous gradient so that the quadratic growth assumption will be automatically satisfied. 3 Optimal Regularized Dual Averaging Method In dual averaging methods [7, ], the key proximal mapping step utilizes the average of all past stochastic subgradients to update the parameter } vector. In particular, it takes the form: z t+ = arg min x X { g t, x + hx + β t t x, x 0, where β t is the step-size and g t = t t+ Gz i, ξ i. For strongly convex fx, the current dual averaging methods achieve a rate of O log, which is suboptimal. In this section, we propose a new dual averaging algorithm which adapts to both strongly and non-strongly convex fx via the strong convexity parameter µ and achieves optimal rates in both cases. In addition, for previous dual averaging methods, to guarantee the convergence, the final solution takes the form: x = + t=0 z t and hence is not sparse in nature for sparsityinducing regularizers. Instead of taking the average, we introduce another proximal mapping and generate the final solution directly from the second proximal mapping. This strategy will provide us sparser solutions in practice. It is worthy to note that in RDA, z has been proved to achieve the desirable sparsity pattern i.e., manifold identification property [3]. However, according to [3], the 3

convergence of ϕz to the optimal ϕx is established only under a more restrictive assumption that x is a strong local minimizer of ϕ relative to the optimal manifold and the convergence rate is quite slow. Without this assumption, the convergence of ϕz is still unknown. The proposed optimal RDA ORDA method is presented in Algorithm. To simplify our notations, we define the parameter µ = µ/τ, which scales the strong convexity parameter µ by τ, where τ is the quadratic growth constant. In general, the constant Γ which defines the step-size parameter γ t is set to L. However, we allow Γ to be an arbitrary constant greater than or equal to L to facilitate the introduction of the multi-stage ORDA in the later section. The parameter c is set to achieve the optimal rates for both convex and strongly convex loss. When µ > 0 or equivalently, µ > 0, c is set to 0 so that γ t τγ τl; while for µ = 0, c = τ+m. Since x,x x is unknown in practice, 0 one might replace x, x 0 in c by a tuning parameter. Here, we make a few more explanations of Algorithm. In Step, the intermediate point y t is a convex combination of x t and z t and when µ = 0, y t = θ t x t + θ t z t. The choice of the combination weights is inspired by [0]. Second, with our choice of θ t and ν t, it is easy to prove that t ν i = θ t ν t. Therefore, g t in Step 3 is a convex combination of {Gy i, ξ i } t. As compared to RDA which uses the average of past subgradients, g t in ORDA is a weighted average of all past stochastic subgradients and the subgradient from the larger iteration has a larger weight i.e., i+ Gy i, ξ i has the weight t+t+. In practice, instead of storing all past stochastic subgradients, g t could be simply updated based on g t : g t = θ t ν gt t θ t ν t + Gy t,ξ t ν t. We also note that since the error in the stochastic subgradient Gy t, ξ t will affect the sparsity of x t+ via the second proximal mapping, to obtain stable sparsity recovery performances, it would be better to construct the stochastic subgradient with a small batch of samples [, ]. This could help to reduce the noise of the stochastic subgradient. 3. Convergence Rate We present the convergence rate for ORDA. We start by presenting a general theorem without plugging the values of the parameters. To simplify our notations, we define t := Gy t, ξ t f y t. Theorem For ORDA, if we require c > 0 when µ = 0, then for any t 0: ϕx t+ ϕx θ t ν t γ t+ x, x 0 + θ tν t θtµ µ+γ t θ t t i + M + θ t ν t µ τθ i + θ iγ i θ τ il ν i where ẑ t = µ = 0. Taking the expectation on both sides of Eq.5: t x ẑ i, i ν i, 5 y t + θ tµ+γ t θ t µ+γ t z θt t, is a convex combination of y t and z t ; and ẑ t = z t when Eϕx t+ ϕx θ t ν t γ t+ x, x 0 + + M θ t ν t t. 6 µ τθ i + θ iγ i θ τ i L ν i The proof of Theorem is given in Appendix. In the next two corollaries, we establish the rates of convergence in expectation for ORDA by choosing different values for c based on µ. Corollary For convex fx with µ = 0, by setting c = Eϕx + ϕx 4τL x, x 0 + τ+m x,x 0 and Γ = L, we obtain: 8 + M τ x, x 0. 7 Based on Eq.6, the proof of Corollary is straightforward with the details in Appendix. Since x is unknown in practice, one could set c by replacing x, x 0 in c with any value D x, x 0. By doing so, Eq.7 remains valid after replacing all x, x 0 by D. For convex fx with µ = 0, the rate in Eq.7 has achieved the uniformly-optimal rate according to [5]. In fact, if fx is a deterministic and smooth function with = M = 0 e.g., smooth empirical loss, one only needs 4

to change the stochastic subgradient Gy t, ξ t to fy t. The resulting algorithm, which reduces to Algorithm 3 in [0], is an optimal deterministic first-order method with the rate O L x,x 0. We note that the quadratic growth assumption of x, y is not necessary for convex fx. If one does not assume this assumption and replaces the last step in ORDA by x t+ = arg min x X { x, Gy t, ξ t + hx + µ θ t + γt x y t }, we can achieve the same rate as in Eq.7 but just removing all τ from the right hand side. But the quadratic growth assumption is indeed required for showing the convergence for strongly convex fx as in the next corollary. Corollary For strongly convex fx with µ > 0, we set c = 0 and Γ = L and obtain that: Eϕx + ϕx 4τL x, x 0 + 4τ + M. 8 µ The dominating term in Eq.8, O µ, is optimal and better than the O log µ rate for previous dual averaging methods. However, ORDA has not achieved the uniformly-optimal rate, which takes the form of O +M µ +exp µ L. In particular, for deterministic smooth and strongly convex fx i.e., empirical loss with = M = 0, ORDA only achieves the rate of O L while the optimal deterministic rate should be O exp µ L [6]. Inspired by the multi-restart technique in [7, ], we present a multi-stage extension of ORDA in Section 4 which achieves the uniformlyoptimal convergence rate. 3. High Probability Bounds For stochastic optimization problems, another important evaluation criterion is the confidence level of the objective value. In particular, it is of great interest to find ϵ, δ as a monotonically decreasing function in both and δ 0, such that the solution x + satisfies Pr ϕx + ϕx ϵ, δ δ. In other words, we want to show that with probability at least δ, ϕx + ϕx < ϵ, δ. According to Markov inequality, for any ϵ > 0, Prϕx + ϕx ϵ Eϕx + ϕx ϵ. Therefore, we have ϵ, δ = Eϕx + ϕx δ. Under the basic assumption in Eq.3, namely E ξ Gx, ξ f x, and according to +M x,x Corollary and, ϵ, δ = O 0 δ for convex fx, and ϵ, δ = O for strongly convex fx. +M µδ However, the above bounds are quite loose. To obtain tighter bounds, we strengthen the basic assumption of the stochastic subgradient in Eq. 3 to the light-tail assumption [4]. In particular, we assume that E exp { Gx, ξ f x / } exp{}, x X. By further making the boundedness assumption x ẑ t D and utilizing a technical lemma from [3], we obtain a much tighter high probability bound with ϵ, δ = O convex fx. The details are presented in Appendix. ln/δd for both convex and strongly 4 Multistage ORDA for Stochastic Strongly Convex Optimization As we show in Section 3., for convex fx, ORDA achieves the uniformly-optimal rate. However, for strongly convex fx, although the dominating term of the convergence rate in Eq.8 is optimal, the overall rate is not uniformly-optimal. Inspired by the multi-stage stochastic approximation methods [7, 9, ], we propose the multi-stage extension of ORDA in Algorithm for stochastic strongly convex optimization. For each stage k K, we run ORDA in Algorithm as a sub-routine for k iterations with the parameter γ t = ct + 3/ + τγ with c = 0 and Γ = Λ k + L. Roughly speaking, we set k = k and Λ k = 4Λ k. In other words, we double the number of iterations for the next stage but reduce the step-size. The multi-stage ORDA has achieved uniformly-optimal convergence rate as shown in Theorem with the proof in Appendix. Due to our proof technique, instead of showing Eϕx ϕx ϵ as in ORDA, we show the number of iterations ϵ to achieve the ϵ-accurate solution: Eϕx ϵ ϕx ϵ. But the two convergence rates are 5

Algorithm Multi-stage ORDA for Stochastic Strongly Convex Optimization Initialization: x 0 X, a constant 0 ϕx 0 ϕx and the number of stages K. Iterate for k =,,..., K: { } τl. Set k = max 4 µ, k+9 τ +M µ 0. Set Λ k = 3/ k k µ +M τ 0 3. Generate x k by calling the sub-routine ORDA x k, k, Γ = Λ k + L, c = 0 Output: x K equivalent: the rate in Eq. 9 can be converted into O +M µ + exp{ µ τl, } which is the uniformly-optimal for stochastic strongly convex optimization. Theorem If we run multi-stage ORDA for K stages with K = log 0 ϵ for any given ϵ, we have Eϕ x K ϕx ϵ and the total number of iterations is upper bounded by: = K k 4 k= τl µ log 0 ϵ + 04τ + M. 9 µϵ 5 Related Works In the last few years, a number of stochastic gradient methods [6, 5, 8,, 4,, 0,, 7, 4, 3] have been developed to solve Eq., especially for a sparsity-inducing hx. In Table, we compare the proposed ORDA and its multi-stage extension with some widely used stochastic gradient methods using the following metrics. For the ease of comparison, we assume fx is smooth with M = 0.. The convergence rate for solving non-strongly convex fx and whether this rate has achieved the uniformly-optimal Uni-opt rate.. The convergence rate for solving strongly convex fx and whether the dominating term of rate is optimal, i.e., O and the overall rate is uniformly-optimal. 3. Whether the final solution x, on which the results of convergence are built, is generated from the weighted average of previous iterates Avg or from the proximal mapping Prox. For sparsity-inducing regularizers, the solution directly from the proximal mapping is often sparser than the averaged solution. 4. Whether an algorithm allows to use a general Bregman divergence in proximal mapping or it only allows the Euclidean distance x, y = x y. In Table, the algorithms in the first 7 rows are stochastic approximation algorithms where only the current stochastic gradient is used at each iteration. The last 4 rows are dual averaging methods where all past subgradients are used. Some algorithms in Table make a more restrictive assumption on the stochastic gradient: G > 0, E Gx, ξ G, x X. It is easy to verify that this assumption implies our basic assumption in Eq.3 by Jensen s inequality. As we can see from Table, the proposed ORDA possesses all good properties except that the convergence rate for strongly convex fx is not uniformly-optimal. Multi-stage ORDA further improves this rate to be uniformly-optimal. In particular, SAGE [8] achieves a nearly optimal rate since the parameter D in the convergence rate is chosen such that E x t x D for all t 0 and it could be much larger than x, x 0. In addition, SAGE requires the boundedness of the domain X, the smoothness of fx, and only allows the Euclidean distance in proximal mapping. As compared to AC-SA [0] and multi-stage AC-SA [], our methods do not require the final averaging step; and as shown in our experiments, ORDA has better empirical performances due to the usage of all past stochastic subgradients. Furthermore, we improve the rates of RDA and extend AC-RDA to an optimal algorithm for both convex and strongly convex fx. Another highly relevant work is [9]. Juditsky et al. [9] proposed multi-stage algorithms to achieve the optimal 6

FOBOS [6] COMID [5] Convex fx Strongly Convex fx Final x Bregman Rate Uni-opt Rate Opt Uni-opt G O O O G log O O Prox O G O O O G log O O Prox YES D + LD + LD + L SAGE [8] O EARLY O YES O Prox O AC-SA [0] O + L YES O YES O Avg YES M-AC-SA [] A A O + exp{ µ } YES YES Avg YES L G Epoch-GD [7] A A O YES O Avg O G RDA [] O O O G log O O Avg YES AC-RDA [] O + L YES A A A Avg YES ORDA O + L YES M-ORDA A A O O YES O Prox YES + exp{ µ } YES YES Prox YES L + L Table : Summary for different stochastic gradient algorithms. is short for x, x 0; AC for accelerated ; M for multi-stage and A stands for either not applicable or no analysis of the rate. strongly convex rate based on non-accelerated dual averaging methods. However, the algorithms in [9] assume that ϕx is a Lipschitz continuous function, i.e., the subgradient of ϕx is bounded. Therefore, when the domain X is unbounded, the algorithms in [9] cannot be directly applied. Recently, the paper [8] develops another stochastic gradient method which achieves the rate O G for strongly convex fx. However, for non-smooth fx, it requires the averaging of all iterates and this rate is not uniformly-optimal. 6 Simulated Experiments In this section, we conduct simulated experiments to demonstrate the performance of ORDA and its multi-stage extension M ORDA. We compare our ORDA and M ORDA only for strongly convex loss with several state-of-the-art stochastic gradient methods, including RDA and AC-RDA [], AC-SA [0], FOBOS [6] and SAGE [8]. For a fair comparison, we compare all different methods using solutions which have expected convergence guarantees. For all algorithms, we tune the parameter related to step-size e.g., c in ORDA for convex loss within an appropriate range and choose the one that leads to the minimum objective value. In this experiment, we solve a sparse linear regression problem: min x R n fx+hx where fx = E a,ba T x b + ρ x and hx = λ x. The input vector a is generated from 0, I n n and the response b = a T x + ϵ, where x i = for i n/ and 0 otherwise and the noise ϵ 0,. When ρ = 0, th problem is the well known Lasso [9] and when ρ > 0, it is known as Elastic-net []. The regularization parameter λ is tuned so that a deterministic solver on all the samples can correctly recover the underlying sparsity pattern. We set n = 00 and create a large pool of samples for generating stochastic gradients and evaluating objective values. The number of iterations is set to 500. Since we focus on stochastic optimization instead of online learning, we could randomly draw samples from an underlying distribution. So we construct the stochastic gradient using the mini-batch strategy [, ] with the batch size 50. We run each algorithm for 00 times and report the mean of the objective value and the F-score for sparsity recovery performance. F-score is defined as precision recall precision+recall where precision = p i= { x i=,x i =} / p i= { x i=} and recall = p i= { x i=,x i =} / p i= {x i =}. The higher the F-score is, the better the recovery ability of the sparsity pattern. The standard deviations for both objective value and the F-score in 00 runs are very small and thus omitted here due to space limitations. We first set ρ = 0 to test algorithms for non-strongly convex fx. The result is presented in Table the first two columns. We also plot the decrease of the objective values for the first 00 iterations in Figure. From Table, ORDA performs the best in both objective value and recovery 7

ρ = 0 ρ = Obj F Obj F RDA 0.87 0.67.57 0.67 AC-RDA 0.67 0.67. 0.67 AC-SA 0.66 0.67.0 0.67 FOBOS 0.98 0.83.9 0.84 SAGE 0.65 0.8.09 0.73 ORDA 0.56 0.9 0.97 0.87 M ORDA.A..A. 0.98 0.88 Table : Comparisons in objective value and F-score. Objective 8 7 6 5 4 3 0 RDA AC RDA AC SA FOBOS SAGE ORDA 50 00 50 00 Iteration Figure : Obj for Lasso. Objective 3 30 9 8 7 6 5 4 3 RDA AC RDA AC SA FOBOS SAGE ORDA 50 00 50 00 Iteration Figure : Obj for Elastic-et. ability of sparsity pattern. For those optimal algorithms e.g., AC-RDA, AC-SA, SAGE, ORDA, they achieve lower final objective values and the rates of the decrease are also faster. We note that for dual averaging methods, the solution generated from the first proximal mapping e.g., z t in ORDA has almost perfect sparsity recovery performance. However, since here is no convergence guarantee for that solution, we do not report results here. Then we set ρ = to test algorithms for solving strongly convex fx. The results are presented in Table the last two columns and Figure and 3. As we can see from Table, ORDA and M ORDA perform the best. Although M ORDA achieves the theoretical uniformlyoptimal convergence rate, the empirical performance of M ORDA is almost identical to that of ORDA. This observation is consistent with our theoretical analysis since the improvement of the convergence rate only appears on the non-dominating term. In addition, ORDA, M ORDA, AC-SA and SAGE with the convergence rate O achieve lower objective values as compared to other algorithms with the rate O log. For better visualization, we do Objective 30 8 6 4 0 00 00 300 400 500 Iteration ORDA M_ORDA Figure 3: ORDA v.s. M ORDA. not include the comparison between M ORA and ORDA in Figure. Instead, we present the comparison separately in Figure 3. From Figure 3, the final objective values of both algorithms are very close. An interesting observation is that, for M ORDA, each time when a new stage starts, it leads to a sharp increase in the objective value following by a quick drop. 7 Conclusions and Future Works In this paper, we propose a novel dual averaging method which achieves the optimal rates for solving stochastic regularized problems with both convex and strongly convex loss functions. We further propose a multi-stage extension to achieve the uniformly-optimal convergence rate for strongly convex loss. Although we study stochastic optimization problems in this paper, our algorithms can be easily converted into online optimization approaches, where a sequence of decisions {x t } t= are generated according to Algorithm or. We often measure the quality of an online learning algorithm via the so-called regret, defined as R x = t= F xt, ξ t + hx t F x, ξ t + hx. Given the expected convergence rate in Corollary and, the expected regret can be easily derived. For example, for strongly convex fx: ER x t= Eϕx t ϕx t= O t = Oln. However, it would be a challenging future work to derive the regret bound for ORDA instead of the expected regret. It would also be interesting to develop the parallel extensions of ORDA e.g., combining the distributed mini-batch strategy in [] with ORDA and apply them to some large-scale real problems. 8

References [] A. Cotter, O. Shamir,. Srebro, and K. Sridharan. Better mini-batch algorithms via accelerated gradient methods. In Advances in eural Information Processing Systems IPS, 0. [] O. Dekel, R. Gilad-Bachrach, O. Shamir, and L. Xiao. Optimal distributed online prediction using mini-batches. Technical report, Microsoft Research, 0. [3] J. Duchi, P. L. Bartlett, and M. Wainwright. Randomized smoothing for stochastic optimization. arxiv:03.496v, 0. [4] J. Duchi, E. Hazan, and Y. Singer. Adaptive subgradient methods for online learning and stochastic optimization. In Conference on Learning Theory COLT, 00. [5] J. Duchi, S. Shalev-Shwartz, Y. Singer, and A. Tewari. Composite objective mirror descent. In Conference on Learning Theory COLT, 00. [6] J. Duchi and Y. Singer. Efficient online and batch learning using forward-backward splitting. Journal of Machine Learning Research, 0:873 898, 009. [7] E. Hazan and S. Kale. Beyond the regret minimization barrier: an optimal algorithm for stochastic strongly-convex optimization. In Conference on Learning Theory COLT, 0. [8] C. Hu, J. T. Kwok, and W. Pan. Accelerated gradient methods for stochastic optimization and online learning. In Advances in eural Information Processing Systems IPS, 009. [9] A. Juditsky and Y. esterov. Primal-dual subgradient methods for minimizing uniformly convex functions. August 00. [0] G. Lan and S. Ghadimi. Optimal stochastic approximation algorithms for strongly convex stochastic composite optimization, part i: a generic algorithmic framework. Technical report, University of Florida, 00. [] G. Lan and S. Ghadimi. Optimal stochastic approximation algorithms for strongly convex stochastic composite optimization, part ii: shrinking procedures and optimal algorithms. Technical report, University of Florida, 00. [] J. Langford, L. Li, and T. Zhang. Sparse online learning via truncated gradient. Journal of Machine Learning Research, 0:777 80, 009. [3] S. Lee and S. J. Wright. Manifold identification of dual averaging methods for regularized stochastic online learning. In International Conference on Machine Learning ICML, 0. [4] A. emirovski, A. Juditsky, G. Lan, and A. Shapiro. Robust stochastic approximation approach to stochastic programming. SIAM Journal on Optimization, 94:574 609, 009. [5] A. emirovski and D. Yudin. Problem complexity and method efficiency in optimization. John Wiley ew York, 983. [6] Y. esterov. Introductory lectures on convex optimization: a basic course. Kluwer Academic Pub, 003. [7] Y. esterov. Primal-dual subgradient methods for convex problems. Mathematical Programming, 0: 59, 009. [8] A. Rakhlin, O. Shamir, and K. Sridharan. To average or not to average? making stochastic gradient descent optimal for strongly convex problems. In International Conference on Machine Learning ICML, 0. [9] R. Tibshirani. Regression shrinkage and selection via the lasso. J.R.Statist.Soc.B, 58:67 88, 996. [0] P. Tseng. On accelerated proximal gradient methods for convex-concave optimization. SIAM Journal on Optimization Submitted, 008. [] L. Xiao. Dual averaging methods for regularized stochastic learning and online optimization. Journal of Machine Learning Research, :543 596, 00. [] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. J. R. Statist. Soc. B, 67:30 30, 005. 9