Distributed Robust Optimization (DRO) Part I: Framework and Example

Similar documents
Alternative Decompositions for Distributed Maximization of Network Utility: Framework and Applications

Primal/Dual Decomposition Methods

Solving Dual Problems

A Distributed Newton Method for Network Utility Maximization, II: Convergence

Interior-Point Methods for Linear Optimization

Distributed Robust Optimization (DRO) Part II: Wireless Power Control

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Robust Network Codes for Unicast Connections: A Case Study

Interior Point Algorithms for Constrained Convex Optimization

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.

4y Springer NONLINEAR INTEGER PROGRAMMING

Routing. Topics: 6.976/ESD.937 1

3.10 Lagrangian relaxation

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

ELE539A: Optimization of Communication Systems Lecture 16: Pareto Optimization and Nonconvex Optimization

Distributed Approaches for Proportional and Max-Min Fairness in Random Access Ad Hoc Networks

Uniqueness of Generalized Equilibrium for Box Constrained Problems and Applications

Primal Solutions and Rate Analysis for Subgradient Methods

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Algorithms for constrained local optimization

12. Interior-point methods

Duality Theory of Constrained Optimization

CS261: A Second Course in Algorithms Lecture #12: Applications of Multiplicative Weights to Games and Linear Programs

Robust linear optimization under general norms

Lectures 6, 7 and part of 8

ELE539A: Optimization of Communication Systems Lecture 6: Quadratic Programming, Geometric Programming, and Applications

Machine Learning. Support Vector Machines. Manfred Huber

Lagrange duality. The Lagrangian. We consider an optimization program of the form

On the complexity of maximizing the minimum Shannon capacity in wireless networks by joint channel assignment and power allocation

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

The Multi-Path Utility Maximization Problem

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

Robust Performance of Spectrum Sensing in. Cognitive Radio Networks

This Dissertation. entitled. Event-triggered distributed algorithms for network optimization. Pu Wan

A strongly polynomial algorithm for linear systems having a binary solution

Lexicographic Max-Min Fairness in a Wireless Ad Hoc Network with Random Access

Node-based Distributed Optimal Control of Wireless Networks

Optimization and Stability of TCP/IP with Delay-Sensitive Utility Functions

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

A Hierarchy of Suboptimal Policies for the Multi-period, Multi-echelon, Robust Inventory Problem

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS

Outline. Relaxation. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Lagrangian Relaxation. Lecture 12 Single Machine Models, Column Generation

Constrained Optimization and Lagrangian Duality

Distributionally Robust Discrete Optimization with Entropic Value-at-Risk

CS 6820 Fall 2014 Lectures, October 3-20, 2014

Notes on Dantzig-Wolfe decomposition and column generation

Characterization of Convex and Concave Resource Allocation Problems in Interference Coupled Wireless Systems

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

Critical Reading of Optimization Methods for Logical Inference [1]

An event-triggered distributed primal-dual algorithm for Network Utility Maximization

Distributed power allocation for D2D communications underlaying/overlaying OFDMA cellular networks

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

5 Handling Constraints

Utility Maximization for Communication Networks with Multi-path Routing

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

The common-line problem in congested transit networks

Notes on Constrained Optimization

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Linear Programming: Simplex

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

Optimal Utility-Lifetime Trade-off in Self-regulating Wireless Sensor Networks: A Distributed Approach

Applications of Robust Optimization in Signal Processing: Beamforming and Power Control Fall 2012

A Distributed Newton Method for Network Utility Maximization, I: Algorithm

Assignment 1: From the Definition of Convexity to Helley Theorem

Ergodic Stochastic Optimization Algorithms for Wireless Communication and Networking

Topological properties

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

4TE3/6TE3. Algorithms for. Continuous Optimization

IE 5531 Midterm #2 Solutions

Lagrangian Duality Theory

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables

On the Approximate Linear Programming Approach for Network Revenue Management Problems

Utility Maximization for Communication Networks with Multi-path Routing

Optimization and Optimal Control in Banach Spaces

Dual Decomposition.

Tutorial on Convex Optimization: Part II

The Trust Region Subproblem with Non-Intersecting Linear Constraints

CS711008Z Algorithm Design and Analysis

DOWNLINK transmit beamforming has recently received

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Decomposition-based Methods for Large-scale Discrete Optimization p.1

Integer Programming ISE 418. Lecture 12. Dr. Ted Ralphs

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

Handout 8: Dealing with Data Uncertainty

Lec12p1, ORF363/COS323. The idea behind duality. This lecture

Optimization in Wireless Communication

Lecture 25: Subgradient Method and Bundle Methods April 24

Multiuser Downlink Beamforming: Rank-Constrained SDP

1 Seidel s LP algorithm

15-780: LinearProgramming

THE EXISTENCE AND USEFULNESS OF EQUALITY CUTS IN THE MULTI-DEMAND MULTIDIMENSIONAL KNAPSACK PROBLEM LEVI DELISSA. B.S., Kansas State University, 2014

Linear Programming Redux

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Applications. Stephen J. Stoyan, Maged M. Dessouky*, and Xiaoqing Wang

9. Dual decomposition and dual algorithms

ECE Optimization for wireless networks Final. minimize f o (x) s.t. Ax = b,

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization

Transcription:

Manuscript Click here to download Manuscript: DRO_Journal_PartI.ps Click here to view linked References Springer Optimization and Engineering manuscript No. (will be inserted by the editor) 0 0 0 Distributed Robust Optimization (DRO) Part I: Framework and Example Kai Yang Jianwei Huang Yihong Wu Xiaodong Wang Mung Chiang the date of receipt and acceptance should be inserted later Abstract Robustness of optimization models for network problems in communication networks has been an under-explored topic. Most existing algorithms for solving robust optimization problems are centralized, thus not suitable for networking problems that demand distributed solutions. This paper represents a first step towards a systematic theory for designing distributed and robust optimization models and algorithms. We first discuss several models for describing parameter uncertainty sets that can lead to decomposable problem structures and thus distributed solutions. These models include ellipsoid, polyhedron, and D-norm uncertainty sets. We then apply these models in solving a robust rate control problem in wireline networks. Three-way tradeoffs among performance, robustness, and distributiveness are illustrated both analytically and through simulations. In Part II of this two-part paper, we will present applications to wireless power control using the framework of distributed robust optimization. Introduction Despite the importance and success of using optimization theory to study communication and network problems, most work in this area makes the unrealistic assump- Kai Yang Bell Labs, Alcatel-Lucent, Murray Hill, NJ, USA E-mail: kai.yang@alcatel-lucent.com Jianwei Huang Department of Information Engineering, The Chinese University of Hong Kong, China E-mail: jwhuang@ie.cuhk.edu.hk Yihong Wu Department of Electrical Engineering, Princeton University, Princeton, NJ, USA E-mail: yihongwu@princeton.edu Xiaodong Wang Department of Electrical Engineering, Columbia University, New York, NY, USA E-mail: wangx@ee.columbia.edu Mung Chiang Department of Electrical Engineering, Princeton University, Princeton, NJ, USA E-mail: chiangm@princeton.edu

0 0 0 tion that the data defining the constraints and objective function of the optimization problem can be obtained precisely. We call the corresponding problems nominal. In many practical problems, these data are typically inaccurate, uncertain, or timevarying. Solving the nominal optimization problems may lead to poor or even infeasible solutions when deployed. Over the last ten years, robust optimization has emerged as a framework of tackling optimization problems under data uncertainty (e.g., [,,,,]). The basic idea of robust optimization is to seek a solution which remains feasible and near-optimal under the perturbation of parameters in the optimization problem. Each robust optimization problem is defined by three-tuple: a nominal formulation, a definition of robustness, and a representation of the uncertainty set. The process of making an optimization formulation robust can be viewed as a mapping from one optimization problem to another. A central question is as follows: when will important properties, such as convexity and decomposability, be preserved under such mapping? In particular, what kind of nominal formulation and uncertainty set representation will preserve convexity and decomposability in the robust version of the optimization problem? So far, almost all of the work on robust optimization focuses on determining what representations of data uncertainty preserves convexity, thus tractability through a centralized solution, in the robust counter part of the nominal problem for a given definition of robustness. For example, for the worst-case robustness, it has been shown that under the assumption of ellipsoid set of data uncertainty, a robust linear optimization problem can be converted into a second-order cone problem; and a robust second-order cone problem can be reformulated as a semi-definite optimization problem []. The reformulations can be solved efficiently using centralized algorithms such as the interior-point method. In this paper, motivated by needs in communication networking, we will focus instead on the distributiveness-preserving formulation of the robust optimization. The driving question thus becomes: how much more communication overhead is introduced in making the problem robust? To develop a systematic theory of Distributed Robust Optimization (DRO), we first show how to represent an uncertainty set in a way that not only captures the data uncertainty in the model but also leads to a distributively solvable optimization problem. Second, in the case where a fully distributed algorithm is not obtainable, we focus on the tradeoff between robustness and distributiveness. Distributed algorithms are often developed based on decomposability structure of the problem, which may disappear as the optimization formulation is made robust. While distributed computation has long been studied [], unlike convexity of a problem, distributiveness of an algorithm does not have a widely-agreed definition. It is often quantified by the amount of communication overhead required: how far and how frequent do the nodes have to pass message around? Zero communication overhead is obviously the most distributed, and we will see how the amount of overhead trades-off with the degree of robustness. In Section, we develop the framework of DRO, with a focus on the characterization of uncertainty sets that are useful for designing distributed algorithms. An example on robust rate control is given in Section, where we discuss various tradeoffs among robustness, distributiveness, and performance through both analysis and numerical studies. Conclusions to this part are given in Section. In Part II of the paper, extensive applications of DRO to wireless power control will be presented

0 0 0 General Framework. Optimization Models There are many uncertainties in the design of communication networks. These uncertainties stem from various sources and can be broadly grouped into two categories. The first type of uncertainties are related to the perturbation of a set of design parameters due to erroneous inputs such as errors in estimation or implementation. We call them perturbation errors. For example, power control is often used to maintain the quality of a wireless link, i.e., the signal-to-interference-plus-noise ratio (SINR). However, many power control schemes are optimized for estimated parameters, e.g., wireless channel quality or interference from other users. In reality, precise information on these parameters are rarely available. Consequently, the corresponding optimized power control scheme could easily become infeasible or exhibit poor performance. A good power control scheme should be capable of protecting the link quality against the perturbation errors. We show such an example on robust wireless power control in the second part of this paper. A common characteristic of such perturbation errors is that they can often be modeled as a continuous uncertainty set surrounding the basic point estimate of the parameter. The size of the uncertainty set could be used to characterize the level of perturbations the designer needs to protect against. The second type of uncertainties is termed as disruptive errors, as they are caused by the failure of communication links within the network. This type of errors can be modeled as a discrete uncertainty set. In Section III of this paper, we present an example related to disruptive errors on robust multipath rate control for service reliability. To make our discussions concrete, we will focus on a class of optimization problems with the following nominal form: maximization of a concave objective function over a given data set characterized by linear constraints, maximize f 0 (x) () subject to Ax b variables x, where A is an M N matrix, x is an N vector, and b is an M vector. We use to denote component-wise inequality. This class of problems can model a wide range of engineering systems (e.g., [,, 0,, ]). Generalization to nonlinear and convex constraint sets presents a direction for major future work. The uncertainty of Problem () may exist in the objective function f 0, matrix parameter A, and vector parameter b. In many cases, the uncertainty in objective function f 0 can be converted into uncertainty of the parameters defining the constraints []. Later in Section we show that it is also possible to convert the uncertainty in b into uncertainty in A in certain cases (although this could be difficult in general). In the rest of the paper, we will focus on studying the uncertainty in A. In many networking problems, structures and physical meaning of matrix A readily leads to distributed algorithms, e.g., in rate control where A is a given routing matrix, distributed subgradient algorithm is well-known to solve the problem, with an interesting correspondence with the practical protocol of TCP. Making the optimization robust may turn the linear constraints into nonlinear ones and increase the amount of message passing. Quantifying the tradeoff between robustness and distributedness is a main subject to study in this paper.

0 0 0 In the robust counterpart of Problem (), we require the constraints Ax b to be valid for any A A, where A denotes the uncertainty set of A, and the definition of robustness is in the worst-case sense []. Notice that we assume the uncertainty set A is either an accurate description or a conservative estimation of the uncertainty in practice. In other words, it is not possible to have parameters outside set A. In this case, the definition of robustness is the worst case robustness, i.e., the solution of the robust optimization problem is always feasible. However, this approach might be too conservative. A more meaningful choice of robustness is the chance-constrained robustness, i.e., the probability of infeasibility (or outage) is upper bounded. We can flexibly adjust the chance-constrained robustness of the robust solution by solving the worst-case robust optimization problem over a properly selected subset of the exact uncertainty set. We will discuss such an example in details in Section.. If we allow an arbitrary uncertainty set A, then the robust optimization problem is difficult to solve even in a centralized manner []. In this paper, we will focus on the study of constraint-wise (i.e. row-wise) uncertainty set, where the uncertainties between different rows in matrix A are decoupled. This restricted class of uncertainty set characterizes the data uncertainty in many practical problems, and it also allows us to convert the robust optimization problem into a formulation that is distributively solvable. Tackling more general forms of uncertainties is another direction of extension in the future. Denote the j th row of A be a T j, which lies in a compact uncertainty set A j. Then the robust optimization problem that we focus on in this paper can be written in the following form: maximize f 0 (x), () subject to a T j x b j, a j A j, j M, variables x. We show that the robust optimization problem () can be equivalently written in a form represented by protection functions instead of uncertainty sets. Denote the nominal counterpart of problem () with a coefficient matrix Ā (i.e., the values when there is no uncertainty), with the j th row s coefficient ā j A j. Then Proposition Problem () is equivalent to the following convex optimization problem: where maximize f 0 (x), () subject to ā T j x + g j (x) b j, j M, variables x, g j (x) = max a j A j (a j ā j ) T x, () is the protection function for the j th constraint, which depends on the uncertainty set A j and the nominal row ā j. Furthermore, g j (x) is a convex function since for any 0 t, we have that max a j A j (a j ā j ) T [tx + ( t)x ] t max a j A j (a j ā j ) T x + ( t) max a j A j (a j ā j ) T x. ()

0 0 0 Different forms of A j will lead to different protection function g j (x), which results in different robustness and performance tradeoff of the formulation. Next we consider several approaches in terms of modeling A j and the corresponding protection function g j (x).. Robust Formulation Defined By Polyhedron Uncertain Set In this case, the uncertainty set A j is a polyhedron characterized by a set of linear inequalities, i.e., A j {a j : D j a j c j }. The protection function is g j (x) = max a j:d ja j c j (a j ā j ) T x, () which involves a linear program (LP). We next show that the uncertainty set can be translated into a set of linear constraints. In the j th constraint in (), with x = ˆx fixed, we can characterize the set a j A j by comparing b j with the outcome of the following LP: v j = max a j:d ja j c j a T j ˆx. () If v j b j, then ˆx is feasible for (). However, this approach is not very useful since it requires solving one LP in () for each possible ˆx. Alternatively, we take the Lagrange dual problem of the LP in (), v j = min c T p j :D T j p j ˆx,p j 0 j p j. () If we can find a feasible solution ˆp j for (), and c T j ˆp j b j, then we must have v j ct j ˆp j b j. We can thus replace constraints in () by the following constraints: c T j p j b j, D T j p j x, p j 0, j M, () and we now have an equivalent and deterministic formulation for Problem (), where all the constraints are linear.. Robust Formulation Defined by D-norm Uncertainty Set D-norm approach [] is another method to model the uncertainty set, and has advantages such as guarantee of feasibility independent of uncertainty distributions and flexibility in trading off between robustness and performance. Consider the j th constraint a T j x b j in (). Denote the set of all uncertain coefficients in a j as E j. The size of E j is E j, which might be smaller than the total number of coefficients N (i.e., a ij for some i might not have uncertainty). For each a ij E j, assume the actual value falls into the range of [ā ij â ij, ā ij + â ij ], in which â ij is a given error bound. Also choose a nonnegative integer Γ j E j. The definition of robustness associated with the D-norm formulation is to maintain feasibility if at most Γ j out of all possible E j parameters are perturbed. Let s denote S i as the set of

0 0 0 Γ j uncertain coefficients. The above robustness definition can be characterized by the following protection function, g j (x) = max â ij x i. (0) S j:s j E j, S j =Γ j i S j If Γ j = 0, then g j (Γ j, x) = 0 and the j th constraint is reduced to the nominal constraint. If Γ j = E j, then g j (x) = i E j â ij x i and the j th constraint becomes Soyster s worst-case formulation []. The tradeoff between robustness and performance can be obtained by adjusting Γ j. Note that the nonlinearity of g j (Γ j, x) is difficult to deal with in the constraint. We can formulate it into the following linear integer programming problem, max â ij x i s ij, s.t. s ij Γ j. () {s ij {0,}} i Ej i E j i E j Thanks to its special structure, () has the same optimal objective function value as the following linear programming problem, max â ij x i s ij, s.t. s ij Γ j. () {0 s ij } i Ej i E j i E j Taking the dual of Problem (), we have min q jγ j + p ij, s.t. q j + p ij â ij x i, i E j. () {p ij 0} i Ej,q j 0 i E j Similar to Section., we can substitute () into the robust Problem () to obtain an equivalent formulation: maximize f 0 (x) () subject to ā ij x i + q j Γ j + p ij b j, j, i i E j q j + p ij â ij y i, i E j, j, y i x i y i, i, variables x, y 0, p 0, q 0.. Robust Formulation Defined by Ellipsoid Uncertainty Set Ellipsoid is commonly used to approximate complicated uncertainty sets for statistical reasons [] and for its ability to succinctly describe a set of discrete points in Euclidean geometry []. Here we consider the case where coefficient a j falls in an ellipsoid centered at the nominal ā j. Specifically, A j = {ā j + a j : a ij ǫ j }. () By (), the protection function is given by { g j (x) = max a ij x i : i i i a ij ǫ j }, ()

0 0 0 n Denote by x = i= x i as the l -norm (or the Euclidean norm) of x. By Cauchy- Schwartz inequality, a ij x i a ij x ǫ j x, and the equality is attained by choosing Therefore we conclude that i i a ij = x iǫ j x. g j (x) = ǫ j x. () Although the resulting constraint in Problem () is not readily decomposable using standard decomposition techniques, we will show in Part II of this paper that this leads to tractable formulations in some important applications (e.g., wireless power control) where users can obtain network information through local measurements without global message passing.. Distributed Algorithm for Robust Optimization under Linear Constraints We now consider possible distributed algorithms to solve the general robust optimization problem with linear constraints in (). Notice, however, the design of a truly distributed algorithm needs to take into account the setup of the practical system. Our motivation here is not to design a general distributed algorithm which is guaranteed to work equally well for all applications. Instead, we try to show how to solve () by exploring its special structures in DRO, and the resulting algorithm may lead to distributed algorithm for each of the individual applications, such as those in Section III below and Part II of the paper. Primal-dual Cutting-plane Method: The cutting-plane method has been used for solving the general nonlinear programming problems and mixed linear integer programming problems []. We assume the protection function here is convex and bounded as in the previous subsections. Also, we assume the feasible region of the robust optimization, X, is a convex and bounded set. At the k th iteration of this method, we use the following function to approximate g j (x), such that x X, we have ḡ j (x) = max l ã T jlx + b jl, l k, () ḡ j (x) g j (x). () The key is to appropriately choose ã jl and b jl for all l, i.e., generating the proper cutting planes. Substituting the protection function g j (x) with ḡ j (x), we obtain the following approximating problem. maximize f 0 (x), (0) subject to ā T j x + â T jlx + b jl b j, j M, l k, variables x.

0 0 0 The feasible set of (0) is still a polytope, and therefore we assume it can be distributively solved. Based on the obtained solution, we can generate a new cutting plane. The new cutting plane, together with previously obtained constraints, can help to attain a better approximation of the protection function. Notice, however, how to efficiently find a new cutting plane depends on which type of protection function we use. Details about how to find cutting planes for different convex set can be found in []. More importantly, for a collection of problems, we can attain a primal feasible solution as well as a dual upper bound to () via solving an approximating problem, as shown in the following theorem. Theorem Assume that ā ji 0, x 0, and g j (x) is a non-decreasing function in x i. Let x [ x, x,..., x N ]T denote the optimal solution to the approximating problem in (0). For the j th constraint, compute δ j = min[, arg max δ (δ)]. () i āji x i +gj(δ x ) b j,δ 0 In particular, for the ellipsoid, D-norm, and polyhedron uncertainty set, [ ] b j δ j = min, ā T j x + g j ( x. () ) It is noticeable that δ j is an auxiliary variable associated with the j th constraint in (). Likewise, for the i th variable, we can define an auxiliary variable δ i such that δ i minâji>0 δ j. Then { x [ x, x,..., x N ]T : x i = x i δ i } is a feasible solution to (). Furthermore, the optimal objective function value of () f 0 (x ) is lower bounded by f 0 ( x ) and upper bounded by f 0 ( x ), i.e., f 0 ( x ) f 0 (x ) f 0 ( x ). () Proof Equation () follows directly from the fact that the protection function under ellipsoid, D-norm or polyhedron uncertainty set scales linearly with δ j, i.e., g j (δ j x) = δ j g j (x). In addition, assume x is a feasible solution to (), it follows from () that x is also a feasible solution to (0), as ā T j x + ḡ j (x) ā T j x + g j (x) b j, j. () Hence, the feasible region of () is a subset of the feasible region of (0), which implies that f 0 (x ) f 0 ( x ). Furthermore, for the j th constraint of (), we have = i ā T j x + g j ( x ) δ i ā ji x i + g j ( x ) () δ j ā ji x i + g j (δ j x ) i () b j. () () holds as x i = x i δ i. () follows because ā ji > 0, δ j δ i, and g j ( x ) is a nondecreasing function in x i. () is obtained from the definition of δ j. Based on () to () we conclude that x is a feasible solution to () and therefore f 0 (x ) f 0 ( x ).

0 0 0 Comment To preserve the special structure of resulting problem we use a group of linear functions to approximate the protection function. However, it is clear from the proof that Theorem is applicable as long as the condition in () is met. In other words, for any feasible approximating function ḡ j (b j, x) satisfying () (not necessarily linear), Theorem implies that we can always compute an upper and lower bound to the optimal objective function of () by solving (0). Note that the key idea of the cutting-plane method is to choose a group of linear constraints to approximate the original problem in (). We show next that for any given convex set, we can always find a group of linear constraints such that () is satisfied. Such a group of linear constraints is related to the well known supporting hyperplanes. A hyperplane [] is said to support a set S in Euclidean space R n if it meets the following two conditions: ) S is entirely contained in one of the two closed half-spaces determined by the hyperplane ) S has at least one point on the hyperplane. Supporting hyperplane theorem [] is well known: If S is a closed convex set in Euclidean space R n, and x is a point on the boundary of S, then there exists a supporting hyperplane containing x. Since g j (x) is a convex function, the set X j {x 0 : a T j x + g j(x) b j } is a closed convex set. As a consequence, we obtain the following proposition on the existence of linear constraints satisfying (). Proposition Recall that X j = {x 0 : a T j x + g j(x) b j }. For any y 0 and y / X j. There exists at least one effective cutting plane e jl (x) ã T jl x + b jl such that ) () is satisfied, i.e., e jl (x) g j (x), x X j ; ) the effective cutting plane contains at least one point in X j ; ) e jl (x) separates X j from y, i.e., X j and y lie at different sides of e jl (x) and e jl (x) does not contain y. Proof Since y / X j, and X j is a closed convex set, we can project y onto the boundary of X j in Euclidean space R n. Assume the projection is ȳ. The supporting hyperplane theorem implies that we can find at least one supporting hyperplane containing ȳ. In addition, X j and y lie at different sides of this supporting hyperplane. Combing the fact that y does not belong to this hyperplane, we conclude that this supporting hyperplane is an effective cutting plane. Based on the above theorem, we have the primal-dual cutting-plane method in the sequel, which stops when the gap between the upper-bound and lower-bound of the objection function is small enough. Algorithm (Primal-dual cutting-plane method). Set time k=0 and set the threshold ǫ. For each j, let H j be the set containing the j th constraint of the nominal problem.. k = k +.. Obtain an approximating problem (0) based on all the constraints in {H j, j M}.. Solve the approximating problem. Calculate the sub-optimality gap η(k) = f 0( x (k)). If η(k) ǫ or k exceeds a given threshold, stop. Otherwise, for each f 0( x (k)) j, if x(k) violates the j th constraint, then find an effective cutting plane, add it into the set H j, and go to step.

0 0 0 0 Theorem The primal-dual cutting plane method monotonically converges to the optimal objective function value of (). In addition, the sub-optimality gap converges to 0 when k goes to infinity, i.e., lim k η(k) = 0. Proof Let H j (k) and X(k) denote respectively the value of H j and the feasible region of approximating problem (0) at the k th iteration of the above algorithm, we have X(k + ) X(k) as H j (k) H j (k + ). Since f 0 ( x (k)) = max x X(k) f 0 (x), we have f 0 ( x (k)) f 0 ( x (k + )), k >. () Hence the sequence f 0 ( x (k)) is monotonically non-increasing. In addition, since X j (k) X j, k, we have X j lim k X j (k) X j. Assume the cutting-plane algorithm does not converge to the optimal solution to (), in view of the fact that X = M j= X j, there must exist one y 0 in R n and a j {,..., n} such that y / X j but y X j. On the other hand, it follows from Proposition that any y / X j can be removed from the set of feasible solution of (0) by adding an effective cutting plane, which contradicts the assumption and therefore conclude the proof. Also, as x (k) converges to x, it follows from () that δ j converges to, which implies that both f 0 ( x (k)) and f 0 ( x (k)) converge to f 0 (x ), and therefore lim k η(k) = 0. Comment The generation of the effective cutting plan is carried out on a row by row basis, and can be performed in either a parallel (synchronous) or serial (asynchronous) manner. As such, it provides a lot of flexibilities that leads to a distributed realization. We later give an example on applying the primal-dual cutting-plane method to the distributed rate control for service reliability in Section. Note that Proposition not only proves the existence of an effective cutting plane but also gives an efficient way to calculate it. In practice, however, the operation of projection in Proposition might pose difficulties for a distributed implementation, since it is a quadratic optimization problem. On the other hand, in many cases there exist more than one effective cutting planes for x (k) [cf. Step of Algorithm ], we can then use an effective cutting plane different from the one proposed in Proposition to simplify the computation. For instance, we will show how to simply generate an effective cutting plan for a D-norm protection function in Section of this paper. Also, for the ellipsoid uncertainty set, [] has proposed an efficient approach to find a good polytope approximation for a given ellipsoid. A drawback of the cutting-plane approach is that it may result in a large number of linear constraints and bring about a lot of extra message passings. In practice we can employ efficient numerical algorithms such as the active-set method to reduce the number of message passing. Nonlinear Gauss-Seidel and Jacobi Method: The Gauss-Seidel (GS) and Jaobi methods were originally proposed as efficient means to solve linear systems, and have been extended to tackle unconstrained as well as constrained optimization problems []. GS methods and Jacobi methods have a variety of applications including CDMA multiuser detection [] and decoding of low-density parity-check code [0]. The general principle of both methods is to decompose a multidimensional optimization problem into multiple one-dimensional optimization problems, and solve these one-dimensional optimization problems in a successive or parallel manner.

0 0 0 Let x i (k) denote the tentative value of the i th variable obtained at the k th iteration of the GS algorithm. Also let X i (x i ) be the feasible region of the i th variable when the values are other variables are set as x i = (x,..., x i, x i+,..., x N ), i.e., X i (x i ) { } x i : ā T j x + g j (x) b j, j M, () where x = (x i, x i ). Then we can calculate x i (k + ) via the following GS iteration, x i (k + ) (0) = arg max x i X i(x (k+),...,x i (k+),x i+(k),...,x N (k)) f 0 (x (k + ),..., x i (k + ), x i, x k+ (k),..., x N (k)). Similarily, in the Jacobi iteration, x i (k + ) is found by x i (k + ) () = arg max x i X i(x (k),...,x i (k),x i+(k),...,x N (k)) f 0 (x (k),..., x i (k), x i, x i+ (k),...x N (k)). The difference between the GS method and the Jacobi method lies in the order of solving the one-dimensional problems. The GS method solves the one-dimensional problems sequentially, while the Jacobi method solve in parallel. They can also be combined together to form hybrid algorithms. We later show a distributed robust power control application, which is closely related to the nonlinear GS and Jacobi methods. Robust Multipath Rate Control for Service Reliability. Nominal and Robust Formulations Consider a wireline network where some links might fail because of human mistakes, software bugs, hardware defects, or natural hazard. Network operators typically reserve some bandwidth for some backup paths. When the primary paths fail, some or all of the traffic will be re-routed to the corresponding disjoint backup paths. Fast system recovery schemes are essential to ensure service availability in the presence of link failure. There are three key components for fast system recovery []: identifying a backup path disjoint from the primary path, computing network resource (such as bandwidth) in reservation prior to link failure, and detecting the link failure in realtime and re-route the traffic. The first component has been investigated extensively in graph theory. The third component has been extensively studied in system research community. Here we consider the robust rate control and bandwidth reservation in the face of possible failure of primary path, which is related to the second component. We first consider the nominal problem with no link failures. Following similar notation as in Kelly s seminal work [], we consider a network with S users, L links and T paths, indexed by s, l and t, respectively. Each user is a unique flow from one source node to one destination node. There could be multiple users between the same sourcedestination node pair. The network is characterized by the L T path-availability 0 matrix { dlt =, if link l is on path t, [D] lt = 0, otherwise.

0 0 0 and T S primary-path-choice nonnegative matrix { wts, if user s uses path t as the primary path, [W ] ts = 0, otherwise. where w ts > 0 indicates the percentage that user s allocates its rate to primary path t, and t w ts =. Let x, c, and y denote source rates, link capacities, and aggregated path rates, respectively. The nominal multi-path rate control problem is maximize s f s(x s) () subject to Dy c, W x y, variables x 0, y 0, where f s(x s) is user s utility function, which is increasing and strictly concave in x s. To represent Problem () in a more compact way, we denote R = DW as the link-source matrix { r [R] ls = ls = t T(l) w ts, if user s uses link l in one of its primary paths, 0, otherwise. where T(l) denotes the set of all paths associated with link l, i.e., T(l) = {t : d lt = }. The nominal problem can be compactly rewritten as maximize s subject to Rx c, variables x 0. f s(x s) () To ensure robust data transmission against the link failures, each user also determines a backup path when it joins the network. The nonnegative backup path choice matrix is { bts, if user s uses path t as the backup path, [B] ts = 0, otherwise. where b ts > 0 indicates the maximum percentage that user s allocates its rate to path t. The actual rate allocation will be a random variable between 0 and b ts, depending on whether the primary paths fail. We further assume that a path can only be used as either a primary path or a backup path for the same user but not both. The corresponding robust multi-path routing rate allocation problem is given by maximize f s(x s) () s subject to r ls x s + g t (b t, x) c l, l. s S(l) variables x 0. t T(l) Here S(l) denotes the set of users using the link l in one of their primary paths, and s S(l) r lsx s is the aggregate rate from users. Moverover, g t (b t, x) is the protection function for the traffic from users who use path t as their backup path, and b t is the t th row of matrix B.

0 0 0 There are many ways of characterizing the protection function. Here we take D- norm as an example. Let E t = {s : b ts > 0, s} denote the set of users who utilize path t as the backup path, and F t,γt denote a set such that F t,γt E t and F t,γt = Γ t. Here Γ t denotes the number of users who might experience path failures, and its value controls the tradeoff between robustness and performance. The protection function can then be written as g t (b t, x) = max b ts x s, t. () F t,γt E t s F t,γt Notice that Γ t is a parameter (instead of a variable) in the protection function (). A centralized algorithm such as the interior point method can be used to solve the service reliability problem (). In practice, however, a distributed solution is preferred in many situations. For example, for a large-scale communication network, multiple nodes and associated links might be removed or updated, and it could be difficult to collect the updated topology information of the whole network. As such, a centralized algorithm may not be applicable. Furthermore, a distributed algorithm can be used to dynamically adjust the rates according to changes in the network topology. Such scenarios motivate us to seek a distributed solution of the multipath rate control problem. Since we consider service reliability in this application, most likely such a update is only required infrequently.. Distributed Primal-dual Algorithms We now develop a fast distributed algorithm to to solve the robust optimization of multipath rate control based on a combination of active-set method [] and dualbased decomposition method. The proposed algorithm can be carried out in a fully distributed manner. We first show that the nonlinear constraints in Problem () can be replaced by a set of linear constraints: Proposition For any path t, the single constraint s S(l) r ls x s + t T(l) is equivalent to the following set of constraints g t (b t, x) c l, () r ls x s + ( b ts x s c l, Ft,Γt, t T(l) ) such that F t,γt E t.() s S(l) t T(l) s F t,γt Proof Let S(l) denote the set of users using link l on their primary paths. From (), we know s F t,γt b ts x s g t (b t, x) for all F t,γt E t. If x satisfies constraint (), we have r ls x s + b ts x s r ls x s + g t (b t, x ) c l, () s S(l) t T(l) s F t,γt s S(l) t T(l)

0 0 0 i.e., it also satisfies the set of constraints in (). On the other hand, if () is true, then r ls x s + g t (b t, x) s S(l) t T(l) = r ls x s + max b ts x s F t,γt E t s S(l) t T(l) s F t,γt () max r ls x s + b ts x s c l. (F t,γt, t T(l)):F t,γt E t s S(l) t T(l) s F t,γt () Therefore these two constraints are equivalent. We further define some short-hand notations. For each choice of paramters ( T(l) sets of users) F l = ( F t,γt E t, t T(l) ), we define a set corresponding to the choices of paths and users: Q Fl { (t, s) : t T(l), s F t,γt }. We further define G l = {Q Fl, F l }. The constraints in () can be rewritten as follows, r ls x s + b ts x s c l, Q Fl G l. () s S(l) (t,s) Q Fl Note the number of constraints in () is ( Et ) t T(l) Γ t, and increases quickly with Γ t and E t. This motives us to design an alternative method to solve (). We next present a sequential optimization approach. The basic idea is to iteratively generate a set Ḡl G l, and use the following set of constraints to approximate (): r ls x s + b ts x s c l, Q Fl Ḡl. () s S(l) (t,s) Q Fl This leads to a relaxation of Problem (): maximize f s(x s) () s subject to r ls x s + b ts x s c l, Q Fl Ḡl, l, s S(l) (t,s) Q Fl variables x 0. Let x denote an optimal solution to the relaxed problem () and x denote an optimal solution of the original problem (). If Ḡl = G l, then we have s fs( xs) = s fs(x s). Even if Ḡl G l, it is still possible that the two optimal objective values are the same as shown in the following theorem: Theorem s fs( xs) = s fs(x s) if the following condition holds max b ts x s = max b ts x s, l. () Q Fl Ḡl Q Fl G l (t,s) Q Fl (t,s) Q Fl

0 0 0 Proof Since Ḡl G l and x is the optimal source rate allocation of the relaxed problem in (), we have s fs( xs) s fs(x s). However, the condition in () implies that r ls x s + max b ts x s s Q Fl Ḡl (t,s) F l = r ls x s + max b ts x s c l, l, () Q s Fl G l (t,s) F l hence x is a feasible solution of (), and s fs( xs) s fs(x s). Thus we have s fs( xs) = s fs(x s). Next we develop a distributed algorithm (Algorithm ) to solve Problem () for a fixed Ḡl for each l, which is suboptimal for solving Problem (). We then design an optimal distributed algorithm (Algorithm ) that achieves the optimal solution of Problem () by iteratively using Algorithm. G By relaxing the constraints in Problem () using dual variables λ = {{λ li } l i= }L l=, we obtain the following Lagrangian, Z(λ, x) = f s(x s) + Ḡl λ li c l r ls x s b ts x s. s l i= s S(l) (t,s) Q Fl (i) For easy indexing, here we denote each set Ḡl = { Q Fl (i), i =,..., Ḡl }. The dual function is Z(λ) = max Z(λ, x). () x 0 The optimization over x in () can be decomposed into one problem for each user s: max f s(x s) Ḡl λ li r ls + λ li b ts x s. () x s 0 l i= (t,s) Q Fl (i) Notice that link l can be viewed as the composition of Ḡl sub-links and each sub-link is associated with a price (dual variable) λ li. Each user s determines its transmission rate x s by considering prices from both its primary path and backup path. The master dual problem is min λ 0 Z(λ), () which can be solved by the subgradient method. For each dual variable λ li, its subgradient can be calculated as ζ li (λ li ) = c l r ls x s b ts x s. () s S(l) (t,s) Q Fl (i) The value of λ li will be updated using the subgradient information correspondingly. The complete algorithm is given as in Algorithm. Algorithm (Distributed primal-dual subgradient algorithm). Set time k = 0, λ(0) = 0, µ(0) = 0, and thresholds ǫ.

0 0 0. Let k = k +.. Each user s determines x s(k) by solving Problem ().. Each user s passes the value of x s(k) to each link associated with this user.. Each link l calculates the subgradients ζ l (λ l (k)) = {ζ li (λ li (k)), t, i} as in (). and ( s(l, k) = c l c l, where c l = max Fl Ḡl s S(l) r lsx s(k) + ) (t,s) F l b ts x s(k).. Each link l updates the dual variables λ l (k + ) = max{λ l (k) + θ(k)ζ l (λ l (k)), 0}.. Each user s who uses link l on either its primary or backup paths calculates the associated dual prices by passing messages over path t from its destination to its source. User s also calculates [ ( s(k) = min l s(l, k) ) r ls >0, ( s(l, k) ) ] b ts>0, (t,s) Q Fl (i) and x s(k) = x s(k) s(k)by collecting messages from its associated links. If η i (k) = s fs( xs(k)) ǫ Z(λ k ) i (where Z(λ k ) is given in ()), stop. Otherwise go to step. Here θ(k) is the step-size at time k. If it meets conditions such as those given in [, pp. ], then this subgradient method is guaranteed to converge to the optimal solution. Common choices of stepsize include the constant step size (θ(k) = β), diminish stepsize (θ(k) = β k ), and square summable (θ(k) = β k ). The parameter ǫ i is a pre-defined small threshold. In addition, due to the special structure of (), we can characterize the sub-optimality gap of the obtained solution at each iteration, as shown below. Theorem Let P ({Ḡl}) be the optimal object function value of the relaxed problem (), λ k be the dual variables at the k th iteration of the subgradient algorithm, and [ x (k),..., x S (k)] be a feasible solution to (). Then P ({Ḡl}) is lower bounded by s fs( xs(k)) and upper bounded by Z(λk ), i.e., f s( x s(k)) P ({Ḡl}) Z(λ k ). () s Furthermore, the sub-optimality gap η i (k) satisfies lim k η i (k) = 0 for all i. Proof For the l th link, we have max r ls x s(k) + b ts x s(k) Q Fl Ḡl s S(l) (t,s) Q Fl = max r ls x s(k) s(k)x s(k) + b ts x s(k) s(k) () Q Fl Ḡl s S(l) (t,s) Q Fl max r ls x s(k) s(l, k)x s(k) + b ts x s(k) s(l, k) () Q Fl Ḡl s S(l) (t,s) Q Fl = s(l, k) max r ls x s(k)x s(k) + b ts x s(k) () Q Fl Ḡl s S(l) (t,s) Q Fl = c l c l c l = c l. () () follows from the definition of x s(k). () holds as s(k) s(l, k), r ls > 0, b ts > 0, (t, s) Q Fl (i). Hence, [ x (k),..., x S (k)] is a feasible solution to () and

0 0 0 s fs( xs(k)) P ({Ḡl}). The second part of the inequality () follows directly from the weak duality theorem, i.e., Z(λ k ) min λ Z(λ) P ({Ḡl}). () If a proper step-size is chosen, e.g., constant or diminish step size, the subgradient method converges to the optimal solution [], and therefore lim k s fs( xs(k)) = P (Ḡl). This combining with the strong duality theorem (i.e., P (Ḡl) = min λ Z(λ)) s implies that lim k fs( xs(k)) = 0. Z(λ k ) Comment The primary goal of computing η i (k) is to estimate the gap between the current solution and the optimal one. In order to reduce the amount of message-passing, we can calculate η i (k) less frequently, e.g., every 0 iterations. Also, s(k) approaches when the sub-optimality gap converges to zero. Hence each user s could use it as a metric to measure the near-optimality of the current feasible solution, which can be implemented in a distributed manner. During each iteration in Algorithm, we can calculate a lower bound to the optimal objective function value of the original problem () as stated in Theorem. The proof is similar to that of Theorem and is thus omitted. Theorem At the k th c step of Algorithm, let s(k) min l l,rls >0 ĉ l (k), where ĉ l (k) = max r ls x s(k) + b ts x s(k). () Q Fl G l s (t,s) Q Fl The optimal objective function value of () is upper bounded by Z(λ k ) and lowered bounded by s fs( xs(k)), where xs(k) = xs(k) s(k) for each user s. Furthermore, the sub-optimality gap of the solution [ x (k),..., x S (k)] can be defined as η o(k) s = fs( xs(k)). Based on Algorithm, we propose the optimal distributed Z(λ k ) algorithm to find the optimal solution of Problem (). The basic idea of Algorithm is to dynamically generate a set of active constraints to approximate the D-norm protection function. This method proceeds in a iterative manner.at each iteration, we use a subset of it to approximate the whole constraint set G l = {Q Fl }, and Algorithm is invoked to solve the relaxed problem. Algorithm (Distributed primal-dual cutting-plane algorithm for robust multipath rate control). For each link l, let Ḡl = {}.. Set the time k o = 0, and the sub-optimality gap for the outer loop iteration ǫ o.. Let k o = k o +.. Set the threshold of sub-optimality gap for the inner loop iteration ǫ i (k o).. Run Algorithm until η i (k i ) ǫ i (k o) or the number of iterations k i exceeds a prespecified threshold. If the sub-optimality gap η o(k i ) ǫ o or k o exceeds a pre-specified threshold, stop. Otherwise, go to the step.. Each user s passes the tentative data rate x s to every link associated with this user.. For each path t, rank {b ts x s} s Et in descending order and take the Γ t largest items to construct a new set F t,γt.

0 0 0. For each link l, construct set Q Fl = {(t, s) : t T(l), s F t,γt } and add it into Ḡ l. Go to step. Algorithm iteratively generates a group of relaxed problems to approximate the original problem (), and eventually converges to the optimal solution if both ǫ i and ǫ o are set as 0. Furthermore, the tradeoff between the performance and complexity can be controlled via setting ǫ i and ǫ o at different values. An intuitive way to accelerate the convergence speed of Algorithm is to set ǫ i as a relatively large value when k o is small and then gradually decreases with k o. As such, the algorithm can quickly find out useful cutting planes at the first several iterations and then gradually converge to the optimal solution. Notice that the amount of message passing at each step of Algorithm increases with Ḡl. On the other hand, due to the special structure of the protection function (), a lot of constraints in Ḡl should be inactive. This motivates us to consider the following distributed active-set method. We first show that inactive constraints are redundant in the theorem below. Theorem Assuming the optimal solution to the relaxation problem () is x [x, x,...x S ]T. For the set G l of each link l, if we remove the subset { Q Fl } corresponding to the inactive constraints at the optimal solution x, i.e., Q F l : Q Fl Ḡl, r ls x s + b ts x s < c l, l, () s S(l) (t,s) Q Fl from problem (), the resulting problem has the same optimal solution as (). Proof Assume that at least one constraint in Problem () is not active at its optimal solution x. Without loss of generality, assume that such inactive constraint corresponds to set Q F l(j) Ḡ l. Then we can remove the corresponding constraint and looks at the following problem: maximize f s(x s) () s subject to r ls x s + b ts x s c l, i Ḡl, l l, s S( l) variables x 0. s S(l) (t,s) Q Fl (i) r ls x s + b ts x s c l, i Ḡ l, i j, (t,s) Q F l(i) It remains to show that x is the optimal solution of (). Notice that the objective function of () is continuously differentiable and concave, and the constraints in () are continuously differentiable and convex functions, the following KKT conditions are

0 0 0 sufficient and necessary for optimality, f s(x s) Ḡl λ li r ls + b ts = 0, () l i= (t,s) Q Fl (i) r ls x s + b ts x s c l, () s (t,s) Q Fl (i) λ li(c li r ls x s b ts x s) = 0, i Ḡl, l, () (t,s) Q Fl (i) λ li 0, i Ḡl, l, () where f s(x) denotes the first order derivative of the function f s(x). Since r ls x s + b ts x s < c l, () s (t,s) Q F l(j) it follows from () that λ lj = 0. This implies that for the solution x, we can find a group of dual variables {λ li 0, li lj} to satisfy the KKT condition of (). Consequently, x is the optimal solution to (). Using induction, we can show that the optimal solution remains the same if we remove all inactive constraints. Theorem motivates us to design a new algorithm that reduces the amount of message passing in the distributed algorithm, as shown below. Algorithm (Distributed primal-dual active-set algorithm for robust multipath rate control). Each link l sets Ḡl as an empty set.. Set the number of outer loop iteration k o = 0, and the threshold of sub-optimality gap for the outer loop iteration ǫ o.. Let k o = k o +.. Set the sub-optimality gap for the inner loop iteration ǫ i (k o).. Algorithm is carried out until η i (k i ) ǫ i (k o) or the number of iterations k i exceeds a pre-specified threshold. If the sub-optimality gap η o(k i ) ǫ o or k o exceeds a pre-specified threshold, stop. Otherwise, go to the step.. Each user s passes the tentative data rate x s to every link associated with this user.. Rank {b ts x s} s Et in descending order for each path t, and take the Γ t largest items to construct a new set F t,γt.. For each link l, construct set Q Fl = {(t, s) : t T(l), s F t,γt }. Remove the redundant constraint set from Ḡl according to (), and add Q Fl into Ḡl. Go to step. We next prove that the above algorithm converges to the optimal solution of Problem (). Theorem Assume the objective function s fs(xs) is strictly concave. The distributed active-set algorithm converges globally and monotonically to the optimal solution of the robust multipath rate control problem in () in finite steps.

0 0 0 0 Proof Let x (q) [x (q), x (q), x (q),..., x S (q)]t be the solution obtained at the q th step of the active-set method. We first prove that if the active-set method does not converge to the optimal solution to () at the q th step, we have s fs(x (q)) > s fs(x (q + )). It follows from Theorem that x (q) is the optimal solution to the relaxed problem in () with only active constraints. Since the objective function is strictly concave, if x (q) is not an optimal solution to (), it must violate one of the constraints in (). Consequently we can add a cutting plane into the constraint set and remove x (q) from the feasible set of the relaxed problem at the (q + ) th step of the active-set method. This combing with the fact that the objective function is strictly concave implies s fs(x (q)) > s fs(x (q + )), i.e., the objective function value of the active-set method is monotonically decreasing before its convergence. Let C(q) be the set of active constraints at the q th step of the distributed primaldual active-set method. Assume the distributed active-set method does not converge before the q th step, we have s fs(x ()) > s fs(x ()) > > s fs(x (q)). Since the objective function is strictly concave,we have C( q) C(q), q < q, which implies that C() C() = C(q). () In view of the fact that the number of different C(q) is finite [cf. ()], we conclude that the proposed algorithm is guaranteed to converge when q goes to infinity.. Numerical Results Here we consider a simple network model with three nodes, links and paths, as shown in Fig.. Paths are single link paths that use links, respectively. Path consists of links and. The first paths are used as primary paths by users in the network. Path is used as the backup path by users, and path is used as the backup path by users. Each user s has a logarithmic utility function log (x s), where the unit of x s is bps. The capacity of each link is fixed at Mbps. Figure shows the convergence behavior of the Algorithm. Here Γ = Γ =, which means we allow maximum and users to experience failures and use path and path as their backup paths to transmit their data respectively. Two variations of active-set algorithms are considered in Figure. We call the first type of active-set method fixed active-set method, as we assume fixed number of inner iterations, namely, the iterations of the subgradient method, at each stage. The algorithm stops when η o ǫ o. The number of inner iterations is set as 00,,, respectively, corresponding to the first three subplots in Figure. In the second version of active-set method, i.e., the adaptive active-set method, the estimated sub-optimality gap, i.e., η i is are used as the stopping criteria. The inner iteration stops when η i ǫ i and the active-set algorithm stops when η o ǫ o. Also, ǫ i is set as 0. 0. ko where k o is the number of outer iterations. ǫ o is set as 0.000 in this simulation study. It is seen that in general the distributed active-set method can converge to a close-to-optimal solution within a few iterations. The active-set method (Algorithm ) with fixed number of iterations converge to the optimal solution if the maximum number of the inner iterations is chosen to be large enough (e.g., 00 and in this study), but the corresponding convergence time might be long (e.g., around 0 and 00, respectively). Reducing the maximum number of inner iterations to speeds up the convergence but fails to