A Distributed Newton Method for Network Optimization

Size: px
Start display at page:

Download "A Distributed Newton Method for Network Optimization"

Transcription

1 A Distributed Newton Method for Networ Optimization Ali Jadbabaie, Asuman Ozdaglar, and Michael Zargham Abstract Most existing wor uses dual decomposition and subgradient methods to solve networ optimization problems in a distributed manner, which suffer from slow convergence rate properties. This paper proposes an alternative distributed approach based on a Newton-type method for solving minimum cost networ optimization problems. The ey component of the method is to represent the dual Newton direction as the solution of a discrete Poisson equation involving the graph Laplacian. This representation enables using an iterative consensus-based local averaging scheme with an additional input term) to compute the Newton direction based only on local information. We show that even when the iterative schemes used for computing the Newton direction and the stepsize in our method are truncated, the resulting iterates converge superlinearly within an explicitly characterized error neighborhood. Simulation results illustrate the significant performance gains of this method relative to subgradient methods based on dual decomposition. I. INTRODUCTION The standard approach to distributed optimization in networs is to use dual decomposition and subgradient or firstorder) methods, which for some classes of problems yields iterative algorithms that operate on the basis of local information see [3], [4],[], and [5]). However, a major shortcoming of this approach, particularly relevant in today s largescale networs, is the slow convergence rate of the resulting algorithms. In this paper, we propose an alternative approach based on using Newton-type or second-order) methods for minimum cost networ optimization problems. We show that the proposed method can be implemented in a distributed manner and has faster convergence properties. Consider a networ represented by a directed graph G = N, E). Each edge e in the networ has a convex cost function φ e x e ), which captures the cost due to congestion effects as a function of the flow x e on this edge. The total cost of a flow vector x = [x e ] e E is given by the sum of the edge costs, i.e., e E φ ex e ). Given an external supply b i for each node i N, the minimum cost networ optimization problem is to find a minimum cost flow allocation vector that satisfies the flow conservation constraint at each node. A. Ozdaglar s research was supported by the National Science Foundation under Career grant DMI-5459 and the DARPA ITMANET program. Ali Jadbabaie s research was supported by Army Research lab s MAST Collaborative Technology Alliance, DARPA/DSO SToMP, and ONR MURI N Ali Jadbabaie and Michael Zargham are with the Department of Electrical and Systems Engineering and GRASP Laboratory, University of Pennsylvania. Asuman Ozdaglar is with the Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. We focus on feasible problems, i.e., we assume that the total in-flow to the networ is equal to the total out-flow, P i N b i =. This problem can be formulated as a convex optimization problem with linear equality constraints. The application of dual decomposition together with a dual subgradient algorithm then yields a distributed iterative solution method. Instead, we propose a distributed primal-dual Newton-type method that achieves a superlinear convergence rate to an error neighborhood). The challenges in using Newton-type methods in this context are twofold. First, the superlinear convergence rate of this type of methods is achieved by using a bactracing stepsize rule, which relies on global information in the computation of the norm of a residual function used in defining the stepsize). We solve this problem by using a consensus-based local averaging scheme for estimating the norm of the residual function. Second, the computation of the dual Newton step involves a matrix inversion, which requires global information. Our main contribution in this regard is to develop a distributed iterative scheme for the computation of the dual Newton step. The ey idea is to recognize that the dual Newton step is defined through a discrete Poisson equation with constant input), which involves the Laplacian of the graph, and therefore can be solved using a consensus-based scheme in a distributed manner. We show that the convergence rate of this scheme is governed by the spectral properties of the underlying graph. Hence, for fast-mixing graphs i.e., those with large spectral gaps), the dual Newton step can be computed efficiently using only local information. Since our method uses consensus-based schemes to compute the stepsize and the Newton direction in each, exact computation is not feasible. Another major contribution of our paper is to consider truncated versions of these consensusschemes at each and present convergence rate analysis of the constrained Newton method when the stepsize and the direction are estimated with some error. We show that when these errors are sufficiently small, the value of the residual function converges superlinearly to a neighborhood of the origin, whose size is explicitly quantified as a function of the errors and the parameters of the objective function and the constraints of the minimum cost networ optimization problem. Our analysis relates to wor on the convergence rate analysis of inexact Newton methods [8], []). These wors focus on providing conditions on the amount of error at each relative to the norm of the gradient of the current iterate that ensures superlinear convergence to the exact Consensus-based schemes have been used extensively in recent literature as distributed mechanisms for aligning or computing the average of) the values held by multiple agents; see [], [7],[8], and [9].

2 optimal solution essentially requiring the error to vanish in the limit). Hence, our focus on providing convergence rate results to an error neighborhood may be of independent interest. The rest of the paper is organized as follows: Section II defines the minimum cost networ optimization problem and shows that the dual decomposition and subgradient method can be implemented in a distributed manner. This section also presents the constrained primal-dual Newton method for this problem and introduces a distributed iterative scheme for computing the dual Newton step. Section III presents a convergence rate analysis for an inexact Newton method, for which there are errors associated with computation of the step and the stepsize. In section IV, simulations demonstrate that the Newton s method outperforms the subgradient method with respect to runtime. Section V contains our concluding remars. Basic Notation and Notions: A vector is viewed as a column vector, unless clearly stated otherwise. We denote by x i the i-th component of a vector x. When x i for all components i of a vector x, we write x. For a matrix A, we write A ij or [A] ij to denote the matrix entry in the i-th row and j-th column. We write x to denote the transpose of a vector x. The scalar product of two vectors x, y R m is denoted by x y. We use x to denote the standard Euclidean norm, x = x x. For a vector-valued function f : R n R m, the gradient matrix of f at x R n is denoted by fx). A vector a R m is said to be a stochastic vector when its components a i, i =,..., m, are nonnegative and their sum is equal to, i.e., m i= a i =. A square m m matrix A is said to be a stochastic matrix when each row of A is a stochastic vector. A stochastic matrix is called irreducible and aperiodic also nown as primitive) if all eigenvalues except the trivial eigenvalue at ) are subunit. One can associate a discrete-time Marov chain with a stochastic matrix and a graph G as follows: The state of the chain at time {,, }, denoted by X), is a node in N the node set of the graph) and the weight associated to each edge in the graph is the probability with which X maes a transition between two adjacent nodes. In other words, the transition from state i to state j happens with probability p ij, the weight of edge i, j). If π) with elements defined as π i ) = PX) = i) is the probability distribution of the state at time, the state distribution satisfies the recursion π +) T = π) T P. If the chain is irreducible and aperiodic then for all initial distributions, π converges to the unique stationary distribution π []. II. MINIMUM COST NETWORK OPTIMIZATION PROBLEM We consider a networ represented by a directed graph G = N, E) with node set N = {,..., N}, and edge set E = {,..., E}. We denote the flow vector by x = [x e ] e E, where x e denotes the flow on edge e. The flow conservation conditions at the nodes can be compactly expressed as Ax = b, where A is the N E node-edge incidence matrix of the graph, i.e., if edge j leaves node i A ij = if edge j enters node i otherwise, and the vector b denotes the external sources, i.e., b i > or b i < ) indicates b i units of external flow enters or exits) node i. We associate a cost function φ e : R R with each edge e, i.e., φ e x e ) denotes the cost on edge e as a function of the edge flow x e. We assume that the cost functions φ e are strictly convex and twice continuously differentiable. The minimum cost networ optimization problem can be written as E minimize φ e x e ) ) e= subject to Ax = b. In this paper, our goal is to investigate iterative distributed methods for solving problem ). In particular, we focus on two methods: first relies on solving the dual of problem ) using a subgradient method; second uses a constrained Newton method, where, at each, the Newton direction is computed iteratively using an averaging method. A. Dual Subgradient Method We first consider solving problem ) using a dual subgradient method. To define the dual problem, we form the Lagrangian function of problem ) L : R E R N R given by E Lx, λ) = φ e x e ) λ Ax b). e= The dual function qλ) is then given by qλ) = inf Lx, λ) x R E E ) = inf x R E φ e x e ) λ Ax + λ b = E inf x e R e= e= φ e x e ) λ A) e x e) + λ b. Hence, in view of the fact that the objective function and the constraints of problem ) are separable in the decision variables x e, the evaluation of the dual function decomposes into one-dimensional optimization problems. We assume that each of these optimization problems has an optimal solution, which is unique by the strict convexity of the functions φ e and is denoted by x e λ). Using the first order optimality conditions, it can be seen that for each e, x e λ) is given by x e λ) = φ e) λ i λ j ), ) where i, j N denote the end nodes of edge e. Thus, for each edge e, the evaluation of x e λ) can be done based on

3 local information about the edge cost function φ e and the dual variables of the incident nodes i and j. We can write the dual problem as maximize λ R N qλ). The dual problem can be solved by using a subgradient method: given an initial vector λ, the iterates are generated by λ + = λ α g for all, where g is a subgradient of the dual function qλ) at λ = λ given by g = Axλ ) b, xλ ) = argmin x R E Lx, λ ), i.e., for all e E, x e λ ) is given by Eq. ) with λ = λ. This method naturally lends itself to a distributed implementation: each node i updates its dual variable λ i using local subgradient) information g i obtained from edges e incident to that node, which in turn updates its primal variables x e λ) using dual variables of the incident nodes. Despite its simplicity and distributed nature, however, it is well-nown that the dual subgradient method suffers from slow rate of convergence see [6] and [5] for rate analysis and construction of primal solutions for dual subgradient methods), which motivates us to consider a Newton method for solving problem ). B. Equality-Constrained Newton Method We next consider solving problem ) using an infeasible start) equality-constrained Newton method see [4], Chapter ). We let fx) = E e= φ ex e ) for notational simplicity. Given an initial primal vector x, the iterates are generated by x + = x + α v where v is the Newton step given as the solution to the following system of linear equations: 3 ) ) ) fx ) A v fx ) =. A w Ax b We let H = fx ) and h = Ax b for notational convenience. Solving for v and w in the preceding yields v = H fx ) + A w ), and AH A )w = h AH fx ). 3) Since the matrix H is a diagonal matrix with entries [H ] ee = φ e x e ) ), given the vector w, the Newton step v can be computed using local information. However, the computation of the vector w at a given primal vector x cannot be implemented in a decentralized manner in view of the fact that solving equation 3) AH A ) requires global information. The following section provides an iterative scheme to compute the vector w using local information. 3 This is essentially a primal-dual method with the vectors v and w acting as primal and dual steps; see Section III. ) Distributed Computation of the Newton Direction: Consider the vector w defined in Eq. 3). The ey step in developing a decentralized iterative scheme for the computation of the vector w is to recognize that the matrix AH A is the weighted Laplacian of the underlying graph G = N, E), denoted by L. Hence, L can be written as L = AH A = D B. Here B is an N N matrix with entries { ) φ e B ) ij = x e ) if e = i, j) E, otherwise, and D is an N N diagonal matrix with entries D ) ii = j N i B ) ij, where N i denotes the set of neighbors of node i, i.e., N i = {j N i, j) E}. Letting s = h AH fx ) for notational convenience, Eq. 3) can be then rewritten as I D + I) B + I))w = D + I) s. This motivates the following iterative scheme nown as splitting) to solve for w. For any t, the iterates are generated by wt + ) = D + I) B + I)wt) + D + I) s, where we suppressed the indices for notational convenience). Note that the matrix P := D + I) B + I) is row stochastic, and when the graph of the networ is connected, it is irreducible and aperiodic i.e., a primitive matrix) [9]. Furthermore, P is diagonally similar to a symmetric matrix, therefore all of its eigenvalues are real and by Perron- Frobenius theorem [] they satisfy = λ P ) > λ P ) λ n P ) >. Aside from one eigenvalue at corresponding to the allone eigenvector ), all other eigenvalues are subunit [9]. As a result, the projection of the dynamics to the orthogonal complement of the span of is stable. Let V be the n n ) dimensional matrix whose orthonormal n columns span, the orthogonal subspace to. Denote wt) = V wt), s = V s, where V V = I n, and V V = I n T n. The projected dynamics can now be written as wt + ) = V P V wt) + V D + I) V s. The above equation depicts the dynamics of a stable linear system with a constant step) input, which from basic linear systems theory is nown to converge to the solution of equation 3). The exponential convergence rate is determined by the convergence rate of the random wal wt + ) = P wt) to its stationary distribution. More precisely, the speed of convergence of wt) to the stationary distribution, depends on the eigenstructure of the probability transition weight) matrix P.

4 ) Convergence Rates: It is well nown that the rate of convergence to the stationary distribution of a Marov chain is governed by the second largest eigenvalue modulus of matrix P defined as µp ) = max i=,,n { λ i P ) } = max{λ P ), λ n P ) }. To mae this statement more precise, let i be the initial state and define the total variation distance between the distribution at time and the stationary distributionπ as i ) = Pij πj. j V The rate of convergence to the stationary distribution is measured using the following quantity nown as the mixing time: T mix = max min{ : i ) < e for all }. i The following theorem indicates the relationship between the mixing time of a Marov chain and the second largest eigenvalue modulus of its probability transition matrix [], []. Theorem : The mixing time of a reversible Marov chain with transition probability matrix W and second largest eigenvalue modulus µ satisfies µ µ) ln ) T mix + log n µ. Therefore, the speed of convergence of the Marov chain to its stationary distribution is determined by the value of µ nown as the spectral gap; the larger the spectral gap, the faster the convergence. This suggests that for each, the dual Newton step w can be computed faster on graphs with large spectral gap. The above convergence results indicate that the dual Newton step can be computed in a decentralized fashion with a diffusion scheme, or more precisely, with a discrete Poisson equation. Since our distributed solution involves an iterative scheme, exact computation is not feasible. In what follows, we show that the Newton method has desirable convergence properties even when the Newton direction is computed with some error provided that the errors are sufficiently small. III. INEXACT NEWTON METHOD In this section, we consider the following convex optimization problem with equality constraints: minimize fx) 4) subject to Ax = b, where f : R n R is a twice continuously differentiable convex function, and A is an m n matrix. The minimum cost networ optimization problem ) is a special case of this problem with fx) = E e= φ ex e ) and A is the N E node-edge incidence matrix. We denote the optimal value of this problem by f. Throughout this section, we assume that the value f is finite and problem 4) has an optimal solution, which we denote by x. We consider an inexact infeasible start) Newton method for solving problem 4) see [4]). In particular, we let y = x, ν) R n R m, where x is the primal variable and ν is the dual variable, and study a primal-dual method which updates the vector y at as follows: y + = y + α d, 5) where α is a positive stepsize, and the vector d is an approximate constrained Newton direction given by Dry )d = ry ) + ɛ. 6) Here, the residual function r : R n R m R n R m is defined as rx, ν) = r dual x, ν), r pri x, ν)), 7) where and r dual x, ν) = fx) + A ν, 8) r pri x, ν)) = Ax b. 9) Moreover, Dry) R n+m) n+m) is the gradient matrix of r evaluated at y, and the vector ɛ is an error vector at. We assume that the error sequence {ɛ } is uniformly bounded from above, i.e., there exists a scalar ɛ such that ɛ ɛ for all. ) We adopt the following standard assumption: Assumption : Let r : R n R m R n R m be the residual function defined in Eqs. 7)-9). Then, we have: a) Lipschitz Condition) There exists some constant L > such that Dry) Drȳ) L y ȳ y, ȳ R n R m. b) There exists some constant M > such that A. Basic Relation Dry) M y R n R m. We use the norm of the residual vector ry) to measure the progress of the algorithm. In the next proposition, we present a relation between the iterates ry ), which holds for any stepsize rule. The proposition follows from a multidimensional extension of the descent lemma see [3]). Proposition : Let Assumption hold. Let {y } be a sequence generated by the method 5). For any stepsize rule α, we have ry + ) α ) ry ) + M Lα ry ) +α ɛ + M Lα ɛ. Proof: We consider two vectors w R n R m and z R n R m. We let ξ be a scalar parameter and define the function gξ) = rw+ξz). From the chain rule, it follows that gξ) = Drw + ξz)z. Using the Lipschitz continuity of the

5 residual function gradient [cf. Assumption a)], we obtain: rw + z) rw) = g) g) = = z gξ)dξ Drw + ξz)zdξ Drw + ξz) Drw))zdξ + Drw)zdξ Drw + ξz) Drw) z dξ + Drw)z Lξ z dξ + Drw)z = L z + Drw)z. We apply the preceding relation with w = y and z = α d and obtain ry + α d ) ry ) α Dry )d + L α d. By Eq. 6), we have Dry )d = ry ) + ɛ. Substituting this in the previous relation, this yields ry + α d ) α )ry ) + α ɛ + L α d. Moreover, using Assumption b), we have d = Dry ) ry ) + ɛ ) Dry ) ry ) + ɛ M ry ) + ɛ ). Combining the above relations, we obtain ry + ) α ) ry ) + M Lα ry ) + α ɛ + M Lα ɛ, establishing the desired relation. B. Inexact Bactracing Stepsize Rule We use a bactracing stepsize rule in our method to achieve the superlinear local convergence properties of the Newton method. However, this requires computation of the norm of the residual function ry). In view of the distributed nature of the residual vector ry) [cf. Eq. 7)], this norm can be computed using a distributed consensus-based scheme. Since this scheme is iterative, in practice the residual norm can only be estimated with some error. Let {y } be a sequence generated by the inexact Newton method 5). At each, we assume that we can compute the norm ry ) with some error, i.e., we can compute a scalar n that satisfies n ry ) γ/, ) for some constant γ. Hence, n is an approximate version of ry ), which can be computed for example using distributed iterative methods. For fixed scalars σ, /) and, ), we set the stepsize α equal to α = m, where m is the smallest nonnegative integer that satisfies n + σ m )n + B + γ. ) Here, γ is the maximum error in the residual function norm [cf. Eq. )], and B is a constant given by B = ɛ + M Lɛ, 3) where ɛ is the upper bound on the error sequence in the constrained Newton direction [cf. Eq. )] and M and L are the constants in Assumption. C. Global Convergence of the Inexact Newton Method The next two sections provide convergence rate estimates for the damped Newton phase and the local convergence phase of the inexact Newton method when there are errors in the Newton direction and the bactracing stepsize. ) Convergence Rate for Damped Newton Phase: We first show a strict decrease in the norm of the residual function if the errors ɛ and γ are sufficiently small, as quantified in the following assumption. Assumption : The errors B and e [cf. Eqs. 3) and )] satisfy B + γ 6M L, where is the constant used in the inexact bactracing stepsize rule, and M and L are the constants defined in Assumption. Under this assumption, the next proposition establishes a strict decrease in the norm of the residual function as long as ry) > M L. Proposition : Let Assumptions and hold. Let {y } be a sequence generated by the method 5) when the stepsize sequence {α } is selected using the inexact bactracing stepsize rule [cf. Eq. )]. Assume that ry ) > M L. Then, we have ry + ) ry ) 6M L. Proof: For any, we define ᾱ = M Ln + γ/). In view of the condition on n [cf. Eq. )], we have M L ry ) + γ) ᾱ M <, 4) L ry ) where the last inequality follows by the assumption ry ) > M L. Using the preceding relation and substituting α = ᾱ in the basic relation in Proposition, we obtain: ry + ) ry ) + ᾱ ɛ + M Lᾱ ɛ ) ᾱ ry ) M Lᾱ ry ) ry ) + ᾱ ɛ + M Lᾱ ɛ ᾱ ry ) M ry ) ) L M L ry ) ) ᾱ ɛ + M Lᾱ ɛ + ᾱ ry ) ) B + ᾱ ry ),

6 where the second inequality follows from the definition of ᾱ and the third inequality follows by combining the facts ᾱ <, ɛ ɛ for all, and the definition of B. The constant σ used in the definition of the inexact bactracing line search satisfies σ, /), therefore, it follows from the preceding relation that ry + ) σᾱ ) ry ) + B. Using condition ) once again, this implies n + σᾱ )n + B + γ, showing that the steplength α selected by the inexact bactracing line search satisfies α ᾱ. From condition ), we have which implies n + σα )n + B + γ, ry + ) σᾱ ) ry ) + B + γ. Combined with Eq. 4), this yields ry + ) σ M L ry ) + γ) By Assumption, we also have γ B + γ 6M L, ) ry ) + B + γ. which in view of the assumption ry ) > M L implies that γ ry ). Substituting this in the preceding relation and using the fact α, /), we obtain ry + ) ry ) 8M L + B + γ. Combined with Assumption, this yields the desired result. The preceding proposition shows that, under the assumptions on the size of the errors, at each, we obtain a minimum decrease in the norm of the residual function) of 6M L, as long as ry ) > /M L. This establishes that we need at most 6 ry ) M L, s until we obtain ry ) /M L. ) Convergence Rate for Local Convergence Phase: In this section, we show that when ry ) /M L, the inexact bactracing stepsize rule selects a full step α = and the norm of the residual function ry ) converges quadratically within an error neighborhood, which is a function of the parameters of the problem as given in Assumption ) and the error level in the constrained Newton direction. Proposition 3: Let Assumption hold. Let {y } be a sequence generated by the method 5) when the stepsize sequence {α } is selected using the inexact bactracing stepsize rule [cf. Eq. )]. Assume that there exists some such that ry ) M L. Then, the inexact bactracing stepsize rule selects α =. We further assume that for some δ, /), B + M LB δ 4M L, where B is the constant defined in Eq. 3). Then, we have ry +m ) m M L + B + δ m ) M 5) L m for all m >. As a particular consequence, we obtain lim sup ry m ) B + δ m M L. Proof: We first show that if ry ) M L for some >, then the inexact bactracing stepsize rule selects α =. Replacing α = in the basic relation of Proposition and using the definition of the constant B, we obtain ry + ) M L ry ) + B ry ) + B σ) ry ) + B, where to get the last inequality, we used the fact that the constant σ used in the inexact bactracing stepsize rule satisfies σ, /). Using the condition on n [cf. Eq. )], this yields n + σ)n + B + γ, showing that the steplength α = satisfies condition ) in the inexact bactracing stepsize rule. We next show Eq. 5) using induction on the m. Using α = in the basic relation of Proposition, we obtain ry + ) ry ) + B 4M L + B, where the second inequality follows from the assumption ry ) M L. This establishes relation 5) for m =. We next assume that 5) holds for some m >, and show that it also holds for m +. Eq. 5) implies that ry +m ) 4M L + B + δ 4M L. Using the assumption B + M LB ry +m ) + δ 4M L < M L, δ 4M L, this yields where the strict inequality follows from δ, /). Hence, the inexact bactracing stepsize rule selects α +m =. Using α +m = in the basic relation, we obtain M L ry +m+ ) M L ry +m ) ) + B. Using Eq. 5), this implies that

7 M L ry +m+ ) + M LB + δ m ) ) + M LB m m = + M LB m+ m + δ m m+ +M L B + δ m ) ) M + M LB. L m Using algebraic manipulations and the assumption B + M LB, this yields δ 4M L ry +m+ ) m+ M L + B + δ m+ ) M, L m+ completing the induction and therefore the proof of relation 5). Taing the limit superior in Eq. 5) establishes the final result. IV. SIMULATION RESULTS Our simulation results demonstrate that the decentralized Newton significantly outperforms the dual subgradient method algorithm in terms of runtime. Simulations were conducted as follows: Networ flow problems with conservation constraints and the cost function Φx) = E e= φ ex e ) where φ e x e ) = x e ) 4 were generated on Erdös-Rényi random graphs with n =,, 8, and 6 nodes and an expected node degree, np = 5. We limit the simulations to optimization problems which are well behaved in the sense that the Hessian matrix remains well conditioned, defined as λmax λ min. For this subclass of problem, the runtime of the Newton s method algorithm is significantly less than the subgradient method for frequency Runtime Distribution n=6, np=5) Newton Algorithm Subgradient Algorithm runtime Sec) Fig.. Runtime Histogram, 6 node graphs with mean node degree 5 Runtime sec) Newton Algorithm Subgradient Algorithm Mean Runtime 5 trials) 3 4 # of nodes = * x 4 the cost function is motivated by the Kuramoto model of coupled nonlinear oscillators []. Fig. 3. Average Runtime, 5 samples each fx) Ax s x x* Networ Flow Optimization: Random Graph with 8 nodes Subgradient, runtime =.6854sec) Newton, runtime =.65446sec) 6 3 Fig.. Sample convergence trajectory for a random graph with 8 nodes and mean node degree 5. all trials. Note that the stopping criterion is also tested in a distributed manner for both algorithms. In any particular experiment the Newton s method algorithm discovers the feasible direction in one and proceeds to find the optimal solution in two to four additional s. One such experiment is shown in Figure. A sample runtime distribution for 5 trials is presented in Figure. The fact that of course Newton s method outperform the subgradient scheme is not surprising. Perhaps the surprising fact is that Newton s method outperforms subgradient scheme, despite the fact the computation of the dual Newton step is based on a discrete Poisson equation essentially diffusion with input). On average the Newton s method terminates in less than half of the subgradient runtime and exhibits a tighter variance. This is a representative sample with respect to varying the number of nodes. As shown in Figure 3, the Newton s method algorithm completes in significantly less time on average for all of the graphs evaluated. We also tested the performance of the proposed method on graphs with different connectivity properties. In particular, we considered a complete fully connected) graph and a sparse

8 Ax s fx) Networ Flow Optimization: Random Graph with Subgradient, runtime = 8.79sec) Newton, runtime =.383sec) to compute the dual Newton direction. We show that even when the Newton direction and stepsize are computed with some error, the method achieves superlinear convergence rate to an error neighborhood. Our simulation experiments on different graphs suggested the superiority of the proposed Newton scheme to standard dual subgradient methods. In ongoing wor, we extend this approach to Networ Utility Maximization NUM) problems. Furthermore, we believe new results in [7] will enable us to solve the discrete Poisson equation for the dual Newton step using modified PageRan algorithms in a more scalable fashion [6]. x x* Ax s x x* 5 fx) Fig. 4. Sample convergence trajectory for a complete graph Networ Flow Optimization: Random Graph with nodes Fig. 5. Subgradient, runtime =.48sec) Newton, runtime =.59789sec) Sample convergence trajectory for a sparse graph graph. Figures 4 and 5 compare the performance of the subgradient scheme with the Newton method. The Newton method outperforms the subgradient scheme in both cases. As expected, the performance gains are more significant for the complete graph since the dual Newton step can be computed efficiently on graphs with large spectral gap. V. CONCLUSIONS This paper develops a distributed Newton-type method for solving minimum cost networ optimization problems that can achieve the superlinear convergence rate within some error neighborhood. We show that due to the sparsity structure of the incidence matrix of a networ, the computation of the dual Newton step can be performed using a discrete Poisson equation. This enables using distributed consensus schemes REFERENCES [] D. Aldous, Some inequalities for reversible Marov chains, Journal of the London Mathematical Society 5 98), [] A. Berman and R. J. Plemmons, Nonnegative matrices in the mathematical sciences, Academic Press, New Yor, 979. [3] D.P. Bertseas, A. Nedić, and A.E. Ozdaglar, Convex analysis and optimization, Athena Scientific, Cambridge, Massachusetts, 3. [4] S. Boyd and L. Vandenberghe, Convex optimization, Cambridge University Press, Cambridge, UK, 4. [5] M. Chiang, S.H. Low, A.R. Calderban, and J.C. Doyle, Layering as optimization decomposition: A mathematical theory of networ architectures, Proceedings of the IEEE 95 7), no., [6] F. Chung-Graham, PageRan as a discrete Green s function, preprint. [7], The heat ernel as the pageran of a graph, Proceedings of the National Academy of Sciences 4 7), no. 5, [8] R. Dembo, S. Eisenstat, and T. Steihaug, Inexact newton methods, SIAM Journal on Numerical Analysis 9 98), [9] R. Horn and C. R. Johnson, Matrix analysis, Cambridge University Press, New Yor, 985. [] A. Jadbabaie, J. Lin, and S. Morse, Coordination of groups of mobile autonomous agents using nearest neighbor rules, IEEE Transactions on Automatic Control 48 3), no. 6, 988. [] A. Jadbabaie, N. Motee, and M Barahona, On the stability of Kuramoto model of coupled nonlinear oscillators, Proceedings of the American Control Conference, June 4. [] C.T. Kelley, Iterative methods for linear and nonlinear equations, SIAM, Philadelphia, PA, 995. [3] F.P. Kelly, A.K. Maulloo, and D.K. Tan, Rate control for communication networs: shadow prices, proportional fairness, and stability, Journal of the Operational Research Society ), [4] S. Low and D.E. Lapsley, Optimization flow control, I: Basic algorithm and convergence, IEEE/ACM Transactions on Networing 7 999), no. 6, [5] A. Nedić and A. Ozdaglar, Approximate primal solutions and rate analysis for dual subgradient methods, SIAM Journal on Optimization, forthcoming 8). [6], Subgradient methods in networ resource allocation: Rate analysis, Proc. of CISS, 8. [7] R. Olfati-Saber and R.M. Murray, Consensus problems in networs of agents with switching topology and time-delays, IEEE Transactions on Automatic Control 49 4), no. 9, [8] A. Olshevsy and J.N. Tsitsilis, Convergence rates in distributed consensus averaging, Proceedings of IEEE CDC, 6. [9], Convergence speed in distributed consensus and averaging, SIAM Journal on Optimization, forthcoming 6). [] A. J. Sinclair, Improved bounds for mixing rates of Marov chains and multicommodity flow, Combinatorics, Probability and Computing 99), [] R. Sriant, Mathematics of Internet congestion control, Birhauser, 4.

A Distributed Newton Method for Network Utility Maximization

A Distributed Newton Method for Network Utility Maximization A Distributed Newton Method for Networ Utility Maximization Ermin Wei, Asuman Ozdaglar, and Ali Jadbabaie Abstract Most existing wor uses dual decomposition and subgradient methods to solve Networ Utility

More information

A Distributed Newton Method for Network Utility Maximization, II: Convergence

A Distributed Newton Method for Network Utility Maximization, II: Convergence A Distributed Newton Method for Network Utility Maximization, II: Convergence Ermin Wei, Asuman Ozdaglar, and Ali Jadbabaie October 31, 2012 Abstract The existing distributed algorithms for Network Utility

More information

A Distributed Newton Method for Network Utility Maximization, I: Algorithm

A Distributed Newton Method for Network Utility Maximization, I: Algorithm A Distributed Newton Method for Networ Utility Maximization, I: Algorithm Ermin Wei, Asuman Ozdaglar, and Ali Jadbabaie October 31, 2012 Abstract Most existing wors use dual decomposition and first-order

More information

Distributed Newton-type Algorithms for Network Resource Allocation. Ermin Wei

Distributed Newton-type Algorithms for Network Resource Allocation. Ermin Wei Distributed Newton-type Algorithms for Network Resource Allocation by Ermin Wei B.S., Computer Engineering B.S., Mathematics B.S., Finance Minor in German Language, Literature, and Culture University of

More information

Subgradient Methods in Network Resource Allocation: Rate Analysis

Subgradient Methods in Network Resource Allocation: Rate Analysis Subgradient Methods in Networ Resource Allocation: Rate Analysis Angelia Nedić Department of Industrial and Enterprise Systems Engineering University of Illinois Urbana-Champaign, IL 61801 Email: angelia@uiuc.edu

More information

Fast Linear Iterations for Distributed Averaging 1

Fast Linear Iterations for Distributed Averaging 1 Fast Linear Iterations for Distributed Averaging 1 Lin Xiao Stephen Boyd Information Systems Laboratory, Stanford University Stanford, CA 943-91 lxiao@stanford.edu, boyd@stanford.edu Abstract We consider

More information

Convergence Rate for Consensus with Delays

Convergence Rate for Consensus with Delays Convergence Rate for Consensus with Delays Angelia Nedić and Asuman Ozdaglar October 8, 2007 Abstract We study the problem of reaching a consensus in the values of a distributed system of agents with time-varying

More information

Dual Decomposition.

Dual Decomposition. 1/34 Dual Decomposition http://bicmr.pku.edu.cn/~wenzw/opt-2017-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/34 1 Conjugate function 2 introduction:

More information

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008 Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications

More information

Distributed Multiuser Optimization: Algorithms and Error Analysis

Distributed Multiuser Optimization: Algorithms and Error Analysis Distributed Multiuser Optimization: Algorithms and Error Analysis Jayash Koshal, Angelia Nedić and Uday V. Shanbhag Abstract We consider a class of multiuser optimization problems in which user interactions

More information

Network Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China)

Network Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China) Network Newton Aryan Mokhtari, Qing Ling and Alejandro Ribeiro University of Pennsylvania, University of Science and Technology (China) aryanm@seas.upenn.edu, qingling@mail.ustc.edu.cn, aribeiro@seas.upenn.edu

More information

Distributed Optimization over Networks Gossip-Based Algorithms

Distributed Optimization over Networks Gossip-Based Algorithms Distributed Optimization over Networks Gossip-Based Algorithms Angelia Nedić angelia@illinois.edu ISE Department and Coordinated Science Laboratory University of Illinois at Urbana-Champaign Outline Random

More information

Constrained Consensus and Optimization in Multi-Agent Networks

Constrained Consensus and Optimization in Multi-Agent Networks Constrained Consensus Optimization in Multi-Agent Networks The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher

More information

Solving Dual Problems

Solving Dual Problems Lecture 20 Solving Dual Problems We consider a constrained problem where, in addition to the constraint set X, there are also inequality and linear equality constraints. Specifically the minimization problem

More information

9. Dual decomposition and dual algorithms

9. Dual decomposition and dual algorithms EE 546, Univ of Washington, Spring 2016 9. Dual decomposition and dual algorithms dual gradient ascent example: network rate control dual decomposition and the proximal gradient method examples with simple

More information

Distributed Optimization over Random Networks

Distributed Optimization over Random Networks Distributed Optimization over Random Networks Ilan Lobel and Asu Ozdaglar Allerton Conference September 2008 Operations Research Center and Electrical Engineering & Computer Science Massachusetts Institute

More information

Distributed Control of the Laplacian Spectral Moments of a Network

Distributed Control of the Laplacian Spectral Moments of a Network Distributed Control of the Laplacian Spectral Moments of a Networ Victor M. Preciado, Michael M. Zavlanos, Ali Jadbabaie, and George J. Pappas Abstract It is well-nown that the eigenvalue spectrum of the

More information

Convergence of Rule-of-Thumb Learning Rules in Social Networks

Convergence of Rule-of-Thumb Learning Rules in Social Networks Convergence of Rule-of-Thumb Learning Rules in Social Networks Daron Acemoglu Department of Economics Massachusetts Institute of Technology Cambridge, MA 02142 Email: daron@mit.edu Angelia Nedić Department

More information

Consensus Problems on Small World Graphs: A Structural Study

Consensus Problems on Small World Graphs: A Structural Study Consensus Problems on Small World Graphs: A Structural Study Pedram Hovareshti and John S. Baras 1 Department of Electrical and Computer Engineering and the Institute for Systems Research, University of

More information

A SIMPLE PARALLEL ALGORITHM WITH AN O(1/T ) CONVERGENCE RATE FOR GENERAL CONVEX PROGRAMS

A SIMPLE PARALLEL ALGORITHM WITH AN O(1/T ) CONVERGENCE RATE FOR GENERAL CONVEX PROGRAMS A SIMPLE PARALLEL ALGORITHM WITH AN O(/T ) CONVERGENCE RATE FOR GENERAL CONVEX PROGRAMS HAO YU AND MICHAEL J. NEELY Abstract. This paper considers convex programs with a general (possibly non-differentiable)

More information

Alternative Characterization of Ergodicity for Doubly Stochastic Chains

Alternative Characterization of Ergodicity for Doubly Stochastic Chains Alternative Characterization of Ergodicity for Doubly Stochastic Chains Behrouz Touri and Angelia Nedić Abstract In this paper we discuss the ergodicity of stochastic and doubly stochastic chains. We define

More information

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors Ali Jadbabaie Department of Electrical and Systems Engineering University of Pennsylvania Philadelphia, PA 19104 jadbabai@seas.upenn.edu

More information

Complex Laplacians and Applications in Multi-Agent Systems

Complex Laplacians and Applications in Multi-Agent Systems 1 Complex Laplacians and Applications in Multi-Agent Systems Jiu-Gang Dong, and Li Qiu, Fellow, IEEE arxiv:1406.186v [math.oc] 14 Apr 015 Abstract Complex-valued Laplacians have been shown to be powerful

More information

Agreement algorithms for synchronization of clocks in nodes of stochastic networks

Agreement algorithms for synchronization of clocks in nodes of stochastic networks UDC 519.248: 62 192 Agreement algorithms for synchronization of clocks in nodes of stochastic networks L. Manita, A. Manita National Research University Higher School of Economics, Moscow Institute of

More information

DECENTRALIZED algorithms are used to solve optimization

DECENTRALIZED algorithms are used to solve optimization 5158 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 64, NO. 19, OCTOBER 1, 016 DQM: Decentralized Quadratically Approximated Alternating Direction Method of Multipliers Aryan Mohtari, Wei Shi, Qing Ling,

More information

Distributed Smooth and Strongly Convex Optimization with Inexact Dual Methods

Distributed Smooth and Strongly Convex Optimization with Inexact Dual Methods Distributed Smooth and Strongly Convex Optimization with Inexact Dual Methods Mahyar Fazlyab, Santiago Paternain, Alejandro Ribeiro and Victor M. Preciado Abstract In this paper, we consider a class of

More information

Consensus of Information Under Dynamically Changing Interaction Topologies

Consensus of Information Under Dynamically Changing Interaction Topologies Consensus of Information Under Dynamically Changing Interaction Topologies Wei Ren and Randal W. Beard Abstract This paper considers the problem of information consensus among multiple agents in the presence

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

Asynchronous Non-Convex Optimization For Separable Problem

Asynchronous Non-Convex Optimization For Separable Problem Asynchronous Non-Convex Optimization For Separable Problem Sandeep Kumar and Ketan Rajawat Dept. of Electrical Engineering, IIT Kanpur Uttar Pradesh, India Distributed Optimization A general multi-agent

More information

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization / Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear

More information

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem: CDS270 Maryam Fazel Lecture 2 Topics from Optimization and Duality Motivation network utility maximization (NUM) problem: consider a network with S sources (users), each sending one flow at rate x s, through

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

On the convergence of weighted-average consensus

On the convergence of weighted-average consensus On the convergence of weighted-average consensus Francisco Pedroche Miguel Rebollo Carlos Carrascosa Alberto Palomares arxiv:307.7562v [math.oc] 29 Jul 203 Abstract In this note we give sufficient conditions

More information

Uniqueness of Generalized Equilibrium for Box Constrained Problems and Applications

Uniqueness of Generalized Equilibrium for Box Constrained Problems and Applications Uniqueness of Generalized Equilibrium for Box Constrained Problems and Applications Alp Simsek Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology Asuman E.

More information

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with Appendix: Matrix Estimates and the Perron-Frobenius Theorem. This Appendix will first present some well known estimates. For any m n matrix A = [a ij ] over the real or complex numbers, it will be convenient

More information

On the simultaneous diagonal stability of a pair of positive linear systems

On the simultaneous diagonal stability of a pair of positive linear systems On the simultaneous diagonal stability of a pair of positive linear systems Oliver Mason Hamilton Institute NUI Maynooth Ireland Robert Shorten Hamilton Institute NUI Maynooth Ireland Abstract In this

More information

Chapter 8 Gradient Methods

Chapter 8 Gradient Methods Chapter 8 Gradient Methods An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Introduction Recall that a level set of a function is the set of points satisfying for some constant. Thus, a point

More information

Primal Solutions and Rate Analysis for Subgradient Methods

Primal Solutions and Rate Analysis for Subgradient Methods Primal Solutions and Rate Analysis for Subgradient Methods Asu Ozdaglar Joint work with Angelia Nedić, UIUC Conference on Information Sciences and Systems (CISS) March, 2008 Department of Electrical Engineering

More information

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies IEEE TRANSACTIONS ON AUTOMATIC CONTROL, SUBMITTED FOR PUBLICATION AS A TECHNICAL NOTE. 1 Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies Wei Ren, Student Member,

More information

Consensus Protocols for Networks of Dynamic Agents

Consensus Protocols for Networks of Dynamic Agents Consensus Protocols for Networks of Dynamic Agents Reza Olfati Saber Richard M. Murray Control and Dynamical Systems California Institute of Technology Pasadena, CA 91125 e-mail: {olfati,murray}@cds.caltech.edu

More information

Finite-Time Distributed Consensus in Graphs with Time-Invariant Topologies

Finite-Time Distributed Consensus in Graphs with Time-Invariant Topologies Finite-Time Distributed Consensus in Graphs with Time-Invariant Topologies Shreyas Sundaram and Christoforos N Hadjicostis Abstract We present a method for achieving consensus in distributed systems in

More information

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term; Chapter 2 Gradient Methods The gradient method forms the foundation of all of the schemes studied in this book. We will provide several complementary perspectives on this algorithm that highlight the many

More information

Coordination of Groups of Mobile Autonomous Agents Using Nearest Neighbor Rules

Coordination of Groups of Mobile Autonomous Agents Using Nearest Neighbor Rules University of Pennsylvania ScholarlyCommons Departmental Papers (ESE) Department of Electrical & Systems Engineering June 2003 Coordination of Groups of Mobile Autonomous Agents Using Nearest Neighbor

More information

Convex Optimization of Graph Laplacian Eigenvalues

Convex Optimization of Graph Laplacian Eigenvalues Convex Optimization of Graph Laplacian Eigenvalues Stephen Boyd Abstract. We consider the problem of choosing the edge weights of an undirected graph so as to maximize or minimize some function of the

More information

ADMM and Fast Gradient Methods for Distributed Optimization

ADMM and Fast Gradient Methods for Distributed Optimization ADMM and Fast Gradient Methods for Distributed Optimization João Xavier Instituto Sistemas e Robótica (ISR), Instituto Superior Técnico (IST) European Control Conference, ECC 13 July 16, 013 Joint work

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

Consensus Tracking for Multi-Agent Systems with Nonlinear Dynamics under Fixed Communication Topologies

Consensus Tracking for Multi-Agent Systems with Nonlinear Dynamics under Fixed Communication Topologies Proceedings of the World Congress on Engineering and Computer Science Vol I WCECS, October 9-,, San Francisco, USA Consensus Tracking for Multi-Agent Systems with Nonlinear Dynamics under Fixed Communication

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Optimal Scaling of a Gradient Method for Distributed Resource Allocation 1

Optimal Scaling of a Gradient Method for Distributed Resource Allocation 1 JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS: Vol. 129, No. 3, pp. 469 488, June 2006 ( C 2006) DOI: 10.1007/s10957-006-9080-1 Optimal Scaling of a Gradient Method for Distributed Resource Allocation

More information

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies IEEE TRANSACTIONS ON AUTOMATIC CONTROL, SUBMITTED FOR PUBLICATION AS A TECHNICAL NOTE. Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies Wei Ren, Student Member,

More information

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018 MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S

More information

Convex Optimization. Lecture 12 - Equality Constrained Optimization. Instructor: Yuanzhang Xiao. Fall University of Hawaii at Manoa

Convex Optimization. Lecture 12 - Equality Constrained Optimization. Instructor: Yuanzhang Xiao. Fall University of Hawaii at Manoa Convex Optimization Lecture 12 - Equality Constrained Optimization Instructor: Yuanzhang Xiao University of Hawaii at Manoa Fall 2017 1 / 19 Today s Lecture 1 Basic Concepts 2 for Equality Constrained

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

On Spectral Properties of the Grounded Laplacian Matrix

On Spectral Properties of the Grounded Laplacian Matrix On Spectral Properties of the Grounded Laplacian Matrix by Mohammad Pirani A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Science

More information

DLM: Decentralized Linearized Alternating Direction Method of Multipliers

DLM: Decentralized Linearized Alternating Direction Method of Multipliers 1 DLM: Decentralized Linearized Alternating Direction Method of Multipliers Qing Ling, Wei Shi, Gang Wu, and Alejandro Ribeiro Abstract This paper develops the Decentralized Linearized Alternating Direction

More information

MATH 205C: STATIONARY PHASE LEMMA

MATH 205C: STATIONARY PHASE LEMMA MATH 205C: STATIONARY PHASE LEMMA For ω, consider an integral of the form I(ω) = e iωf(x) u(x) dx, where u Cc (R n ) complex valued, with support in a compact set K, and f C (R n ) real valued. Thus, I(ω)

More information

Lecture: Local Spectral Methods (1 of 4)

Lecture: Local Spectral Methods (1 of 4) Stat260/CS294: Spectral Graph Methods Lecture 18-03/31/2015 Lecture: Local Spectral Methods (1 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough. They provide

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Search Directions for Unconstrained Optimization

Search Directions for Unconstrained Optimization 8 CHAPTER 8 Search Directions for Unconstrained Optimization In this chapter we study the choice of search directions used in our basic updating scheme x +1 = x + t d. for solving P min f(x). x R n All

More information

On Linear Copositive Lyapunov Functions and the Stability of Switched Positive Linear Systems

On Linear Copositive Lyapunov Functions and the Stability of Switched Positive Linear Systems 1 On Linear Copositive Lyapunov Functions and the Stability of Switched Positive Linear Systems O. Mason and R. Shorten Abstract We consider the problem of common linear copositive function existence for

More information

Decentralized Estimation of the Algebraic Connectivity for Strongly Connected Networks

Decentralized Estimation of the Algebraic Connectivity for Strongly Connected Networks Decentralized Estimation of the Algebraic Connectivity for Strongly Connected Networs Hasan A Poonawala and Mar W Spong Abstract The second smallest eigenvalue λ 2 (L) of the Laplacian L of a networ G

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information

Lecture 3: Huge-scale optimization problems

Lecture 3: Huge-scale optimization problems Liege University: Francqui Chair 2011-2012 Lecture 3: Huge-scale optimization problems Yurii Nesterov, CORE/INMA (UCL) March 9, 2012 Yu. Nesterov () Huge-scale optimization problems 1/32March 9, 2012 1

More information

Fastest Distributed Consensus Averaging Problem on. Chain of Rhombus Networks

Fastest Distributed Consensus Averaging Problem on. Chain of Rhombus Networks Fastest Distributed Consensus Averaging Problem on Chain of Rhombus Networks Saber Jafarizadeh Department of Electrical Engineering Sharif University of Technology, Azadi Ave, Tehran, Iran Email: jafarizadeh@ee.sharif.edu

More information

An Interior-Point Method for Large Scale Network Utility Maximization

An Interior-Point Method for Large Scale Network Utility Maximization An Interior-Point Method for Large Scale Network Utility Maximization Argyrios Zymnis, Nikolaos Trichakis, Stephen Boyd, and Dan O Neill August 6, 2007 Abstract We describe a specialized truncated-newton

More information

Stability Analysis of Stochastically Varying Formations of Dynamic Agents

Stability Analysis of Stochastically Varying Formations of Dynamic Agents Stability Analysis of Stochastically Varying Formations of Dynamic Agents Vijay Gupta, Babak Hassibi and Richard M. Murray Division of Engineering and Applied Science California Institute of Technology

More information

ON SPATIAL GOSSIP ALGORITHMS FOR AVERAGE CONSENSUS. Michael G. Rabbat

ON SPATIAL GOSSIP ALGORITHMS FOR AVERAGE CONSENSUS. Michael G. Rabbat ON SPATIAL GOSSIP ALGORITHMS FOR AVERAGE CONSENSUS Michael G. Rabbat Dept. of Electrical and Computer Engineering McGill University, Montréal, Québec Email: michael.rabbat@mcgill.ca ABSTRACT This paper

More information

An event-triggered distributed primal-dual algorithm for Network Utility Maximization

An event-triggered distributed primal-dual algorithm for Network Utility Maximization An event-triggered distributed primal-dual algorithm for Network Utility Maximization Pu Wan and Michael D. Lemmon Abstract Many problems associated with networked systems can be formulated as network

More information

THE EFFECT OF DETERMINISTIC NOISE 1 IN SUBGRADIENT METHODS

THE EFFECT OF DETERMINISTIC NOISE 1 IN SUBGRADIENT METHODS Submitted: 24 September 2007 Revised: 5 June 2008 THE EFFECT OF DETERMINISTIC NOISE 1 IN SUBGRADIENT METHODS by Angelia Nedić 2 and Dimitri P. Bertseas 3 Abstract In this paper, we study the influence

More information

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints Journal of Computational and Applied Mathematics 161 (003) 1 5 www.elsevier.com/locate/cam A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality

More information

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 9, SEPTEMBER 2010 1987 Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE Abstract

More information

1520 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 49, NO. 9, SEPTEMBER Reza Olfati-Saber, Member, IEEE, and Richard M. Murray, Member, IEEE

1520 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 49, NO. 9, SEPTEMBER Reza Olfati-Saber, Member, IEEE, and Richard M. Murray, Member, IEEE 1520 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 49, NO. 9, SEPTEMBER 2004 Consensus Problems in Networks of Agents With Switching Topology and Time-Delays Reza Olfati-Saber, Member, IEEE, and Richard

More information

Randomized Gossip Algorithms

Randomized Gossip Algorithms 2508 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 52, NO 6, JUNE 2006 Randomized Gossip Algorithms Stephen Boyd, Fellow, IEEE, Arpita Ghosh, Student Member, IEEE, Balaji Prabhakar, Member, IEEE, and Devavrat

More information

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained

More information

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME)

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME) WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME ITAI BENJAMINI, GADY KOZMA, LÁSZLÓ LOVÁSZ, DAN ROMIK, AND GÁBOR TARDOS Abstract. We observe returns of a simple random wal on a finite graph to a fixed node,

More information

11. Equality constrained minimization

11. Equality constrained minimization Convex Optimization Boyd & Vandenberghe 11. Equality constrained minimization equality constrained minimization eliminating equality constraints Newton s method with equality constraints infeasible start

More information

On Quantized Consensus by Means of Gossip Algorithm Part I: Convergence Proof

On Quantized Consensus by Means of Gossip Algorithm Part I: Convergence Proof Submitted, 9 American Control Conference (ACC) http://www.cds.caltech.edu/~murray/papers/8q_lm9a-acc.html On Quantized Consensus by Means of Gossip Algorithm Part I: Convergence Proof Javad Lavaei and

More information

3.10 Lagrangian relaxation

3.10 Lagrangian relaxation 3.10 Lagrangian relaxation Consider a generic ILP problem min {c t x : Ax b, Dx d, x Z n } with integer coefficients. Suppose Dx d are the complicating constraints. Often the linear relaxation and the

More information

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS HONOUR SCHOOL OF MATHEMATICS, OXFORD UNIVERSITY HILARY TERM 2005, DR RAPHAEL HAUSER 1. The Quasi-Newton Idea. In this lecture we will discuss

More information

Computational Finance

Computational Finance Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples

More information

Newton s Method for Constrained Norm Minimization and Its Application to Weighted Graph Problems

Newton s Method for Constrained Norm Minimization and Its Application to Weighted Graph Problems Newton s Method for Constrained Norm Minimization and Its Application to Weighted Graph Problems Mahmoud El Chamie 1 Giovanni Neglia 1 Abstract Due to increasing computer processing power, Newton s method

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Markov Chains and Spectral Clustering

Markov Chains and Spectral Clustering Markov Chains and Spectral Clustering Ning Liu 1,2 and William J. Stewart 1,3 1 Department of Computer Science North Carolina State University, Raleigh, NC 27695-8206, USA. 2 nliu@ncsu.edu, 3 billy@ncsu.edu

More information

Convergence rates for distributed stochastic optimization over random networks

Convergence rates for distributed stochastic optimization over random networks Convergence rates for distributed stochastic optimization over random networs Dusan Jaovetic, Dragana Bajovic, Anit Kumar Sahu and Soummya Kar Abstract We establish the O ) convergence rate for distributed

More information

On Quantized Consensus by Means of Gossip Algorithm Part II: Convergence Time

On Quantized Consensus by Means of Gossip Algorithm Part II: Convergence Time Submitted, 009 American Control Conference (ACC) http://www.cds.caltech.edu/~murray/papers/008q_lm09b-acc.html On Quantized Consensus by Means of Gossip Algorithm Part II: Convergence Time Javad Lavaei

More information

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal IPAM Summer School 2012 Tutorial on Optimization methods for machine learning Jorge Nocedal Northwestern University Overview 1. We discuss some characteristics of optimization problems arising in deep

More information

Dual Proximal Gradient Method

Dual Proximal Gradient Method Dual Proximal Gradient Method http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/19 1 proximal gradient method

More information

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

Primal-dual Subgradient Method for Convex Problems with Functional Constraints Primal-dual Subgradient Method for Convex Problems with Functional Constraints Yurii Nesterov, CORE/INMA (UCL) Workshop on embedded optimization EMBOPT2014 September 9, 2014 (Lucca) Yu. Nesterov Primal-dual

More information

Lecture Notes: Geometric Considerations in Unconstrained Optimization

Lecture Notes: Geometric Considerations in Unconstrained Optimization Lecture Notes: Geometric Considerations in Unconstrained Optimization James T. Allison February 15, 2006 The primary objectives of this lecture on unconstrained optimization are to: Establish connections

More information

Consensus-Based Distributed Optimization with Malicious Nodes

Consensus-Based Distributed Optimization with Malicious Nodes Consensus-Based Distributed Optimization with Malicious Nodes Shreyas Sundaram Bahman Gharesifard Abstract We investigate the vulnerabilities of consensusbased distributed optimization protocols to nodes

More information

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications Weijun Zhou 28 October 20 Abstract A hybrid HS and PRP type conjugate gradient method for smooth

More information

WE consider an undirected, connected network of n

WE consider an undirected, connected network of n On Nonconvex Decentralized Gradient Descent Jinshan Zeng and Wotao Yin Abstract Consensus optimization has received considerable attention in recent years. A number of decentralized algorithms have been

More information

A DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS

A DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS Working Paper 01 09 Departamento de Estadística y Econometría Statistics and Econometrics Series 06 Universidad Carlos III de Madrid January 2001 Calle Madrid, 126 28903 Getafe (Spain) Fax (34) 91 624

More information

Maria Cameron. f(x) = 1 n

Maria Cameron. f(x) = 1 n Maria Cameron 1. Local algorithms for solving nonlinear equations Here we discuss local methods for nonlinear equations r(x) =. These methods are Newton, inexact Newton and quasi-newton. We will show that

More information

On Optimal Frame Conditioners

On Optimal Frame Conditioners On Optimal Frame Conditioners Chae A. Clark Department of Mathematics University of Maryland, College Park Email: cclark18@math.umd.edu Kasso A. Okoudjou Department of Mathematics University of Maryland,

More information

Hegselmann-Krause Dynamics: An Upper Bound on Termination Time

Hegselmann-Krause Dynamics: An Upper Bound on Termination Time Hegselmann-Krause Dynamics: An Upper Bound on Termination Time B. Touri Coordinated Science Laboratory University of Illinois Urbana, IL 680 touri@illinois.edu A. Nedić Industrial and Enterprise Systems

More information