Theory and Internet Protocols

Similar documents
CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Today: Linear Programming (con t.)

An introductory example

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 17: Duality and MinMax Theorem Lecturer: Sanjeev Arora

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

CSC373: Algorithm Design, Analysis and Complexity Fall 2017 DENIS PANKRATOV NOVEMBER 1, 2017

Lectures 6, 7 and part of 8

CS711008Z Algorithm Design and Analysis

Another max flow application: baseball

1 Seidel s LP algorithm

Input: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function

CS 6820 Fall 2014 Lectures, October 3-20, 2014

BBM402-Lecture 20: LP Duality

Algorithmic Game Theory and Applications. Lecture 7: The LP Duality Theorem

Zero-Sum Games Public Strategies Minimax Theorem and Nash Equilibria Appendix. Zero-Sum Games. Algorithmic Game Theory.

1 Overview. 2 Extreme Points. AM 221: Advanced Optimization Spring 2016

Massachusetts Institute of Technology 6.854J/18.415J: Advanced Algorithms Friday, March 18, 2016 Ankur Moitra. Problem Set 6

The Simplex Algorithm

Discrete Optimization

Lecture 6: Conic Optimization September 8

Lecture Notes 3: Duality

Lecture #21. c T x Ax b. maximize subject to

Spring 2017 CO 250 Course Notes TABLE OF CONTENTS. richardwu.ca. CO 250 Course Notes. Introduction to Optimization

Topics in Theoretical Computer Science April 08, Lecture 8

Algorithms and Theory of Computation. Lecture 13: Linear Programming (2)

Part IB Optimisation

LINEAR PROGRAMMING III

A Redundant Klee-Minty Construction with All the Redundant Constraints Touching the Feasible Region

4. Duality and Sensitivity

Numerical Optimization

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Linear Programming. Chapter Introduction

Lecture Note 18: Duality

Optimization WS 13/14:, by Y. Goldstein/K. Reinert, 9. Dezember 2013, 16: Linear programming. Optimization Problems

Polynomial-time Computation of Exact Correlated Equilibrium in Compact Games

OPTIMISATION /09 EXAM PREPARATION GUIDELINES

Example Problem. Linear Program (standard form) CSCI5654 (Linear Programming, Fall 2013) Lecture-7. Duality

Linear Programming Inverse Projection Theory Chapter 3

Week 8. 1 LP is easy: the Ellipsoid Method

Chapter 1 Linear Programming. Paragraph 5 Duality

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

Week 3 Linear programming duality

OPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES

IE 5531: Engineering Optimization I

CO 250 Final Exam Guide

LINEAR PROGRAMMING II

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Advanced Linear Programming: The Exercises

Linear and Integer Optimization (V3C1/F4C1)

CS 573: Algorithmic Game Theory Lecture date: January 23rd, 2008

1 Basic Game Modelling

Efficient Nash Equilibrium Computation in Two Player Rank-1 G

4.5 Simplex method. LP in standard form: min z = c T x s.t. Ax = b

2.098/6.255/ Optimization Methods Practice True/False Questions

3. Linear Programming and Polyhedral Combinatorics

Lecture 15: October 15

Lecture 7 Duality II

16.1 L.P. Duality Applied to the Minimax Theorem

Lecture 8. Strong Duality Results. September 22, 2008

7. Lecture notes on the ellipsoid algorithm

Integer Programming ISE 418. Lecture 12. Dr. Ted Ralphs

Lecture 5. x 1,x 2,x 3 0 (1)

Introduction to Semidefinite Programming I: Basic properties a

Duality of LPs and Applications

Linear and Combinatorial Optimization

Lecture notes for Analysis of Algorithms : Markov decision processes

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

Math 16 - Practice Final

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

"SYMMETRIC" PRIMAL-DUAL PAIR

Resource Constrained Project Scheduling Linear and Integer Programming (1)

CSC304: Algorithmic Game Theory and Mechanism Design Fall 2016

Two Lectures on the Ellipsoid Method for Solving Linear Programs

Introduction to Linear Programming

A primal-simplex based Tardos algorithm

Today: Linear Programming

Polynomiality of Linear Programming

Complexity of linear programming: outline

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

Lecture 9 : PPAD and the Complexity of Equilibrium Computation. 1 Complexity Class PPAD. 1.1 What does PPAD mean?

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Duality. Peter Bro Mitersen (University of Aarhus) Optimization, Lecture 9 February 28, / 49

Optimisation and Operations Research

Approximate Farkas Lemmas in Convex Optimization

Lecture 1 Introduction

IE 521 Convex Optimization Homework #1 Solution

Semidefinite Programming

COT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748

A Simpler and Tighter Redundant Klee-Minty Construction

Understanding the Simplex algorithm. Standard Optimization Problems.

6.854 Advanced Algorithms

MATH 4211/6211 Optimization Linear Programming

OPTIMISATION 3: NOTES ON THE SIMPLEX ALGORITHM

CS261: A Second Course in Algorithms Lecture #12: Applications of Multiplicative Weights to Games and Linear Programs

1 Primals and Duals: Zero Sum Games

Optimization: Then and Now

Farkas Lemma, Dual Simplex and Sensitivity Analysis

CSCI5654 (Linear Programming, Fall 2013) Lecture-8. Lecture 8 Slide# 1

Review Solutions, Exam 2, Operations Research

Transcription:

Game Lecture 2: Linear Programming and Zero Sum Nash Equilibrium Xiaotie Deng AIMS Lab Department of Computer Science Shanghai Jiaotong University September 26, 2016

1 2 3 4

Standard Form (P) Outline max c T x subject to Ax b, x 0. A is a m n matrix, c is a column vector of n components, b of m component. We denote it as LP.

Dual LP (D) Outline LP has a corresponding dual form: min b T y subject to y T A c T, y 0. y is a column vector of m component. We denote it as DLP.

Weak Duality Outline LP DLP Proof: Let x be feasible in LP and y in DLP. By prime feasibility, Ax b. Left multiplying y 0: y T Ax y T b. By dual feasibility and x 0: y T Ax c T x it results c T x y T b = b T y, claim holds.

Farkas Lemma Outline Exactly one of the following holds: 1 z 0 : Bz = f 2 w : w T B 0, w T f < 0 No more than one is easy: proof by contradiction: Left multiplying both side of Item 1 by w w T Bz = w T f A contradiction derives as LHS 0 and RHS < 0.

Prove Farkas Lemma Proof of either-or Item 1 states f is a positive linear combination of the columns of B. That is, it is in the convex cone of the column vectors of B. If item 1 does not hold, there is a hyperplane c T z = 0 separating f and column vectors of B. c T f < 0 c T B 0 Claim holds by setting w = c.

Strong Duality as an LP If strong duality holds then there exists (x, y) such that c T x + b T y 0 Ax b A T y c x, y 0 Rewrite (and name it SDLP): c T +b T A 0 0 A T ( ) x y (x, y) T 0 0 b c

Apply Farkas Lemma If SDLP does not hold, by Farkas Lemma, p, q, s 0 s. t. (p T, q T, s T ) c T +b T A 0 0 0 A T (p T, q T, s T ) 0 b < 0 c

If SDLP is not True Outline Simplification 1 pc T + q T A 0 2 +pb T s T A T 0 3 q T b s T c < 0 Right multiplying 1 by s: pc T s + q T As 0. Right multiplying 2 by q: pb T q s T A T q 0. Add them to derive p(b T q c T s) 0. Combining with 3 derives p = 0.

If SDLP is not True Outline Further Simplification 1 q T A 0 2 As 0 3 q T b s T c < 0 By 3, either q T b < 0 or c T s > 0. If q T b < 0 then a feasible DLP implies it is not bounded from below. If c T s > 0 then a feasible LP implies it is not bounded from above.

Full Strong Duality Outline Exactly one of the following holds: 1 DLP = LP 2 LP is feasible and unbounded, DLP is not feasible 3 DLP is feasible and unbounded, LP is not feasible 4 Neither LP nor DLP is feasible and unbounded. Exercise: finish the full proof.

Algorithms Outline Simplex Method by Georgia Dantzig Given an initial solution, the algorithm goes from one basis solution (vertex of polytope) to another. It is practical and runs fast in the average but known to take an exponential time. Ellipsoid Method by Leonid Khachiyan. Given an initial ellipsoid containing the polytope, evaluate if the centre is in the polytope. If not, find a separating hyperplane passing through the centre. Find an ellipsoid containing an half of the original ellipsoid. It is the first polynomial time algorithm for finding a solution. Interior Method by Narendra Karmarkar Given an initial interior point in the polygon, move toward the optimum solution using a barrier function to avoid being to close to the boundary. It is the first practical polynomial time algorithm for finding a Lecture 2: Linear Programming solution. and Zero Sum Nash EquilibriumXiaotie GameDeng

Two Player Games in Strategic Form A Game of two players can be described by their payoff matricess 1 Given two players: the row player and the column player 2 Strategies of the row players are R and the strategies of the column player are C. 3 Let the payoff matrix for the row player be A and the column player be B. 4 if the strategy of the row player be i, the column player be j, then let A(i, j) be the payoff to the row player be A(i, j), and to the column player be B(i, j). A mixed strategy σ [0, 1] R satisfies σ T e = 1 where e = (1, 1,, 1) T

Max-Min Decision Outline Row Player s Mixed Strategy 1 max x min n j=1 x T A j 2 e T x = 1, x 0. Column Player s Mixed Strategy 1 max y min n i=1 B i y 2 e T y = 1, y 0.

LP and Dual LP Outline Row Player: max v 1 v x T A j, 2 e T x = 1, x 0. Column Player: min w 1 w A i y 2 e T y = 1, y 0.

Recall Strong Duality as an LP c T +b T A 0 0 A T ( ) x y (x, y) T 0 0 b c Rewrite it as 0 A T c A 0 b c T b T 0 x y 1 (x, y, 1) T 0 0 0 0

Solve Linear Program Using Zero Sum Game Multiplying z on both side and rescale: x = x z, y = y z, z = z and x + y + z = 1 We use the unprimed x, y, z for the primed x, y, z for simplicity. We construct a payoff matrix for the row player R. The payoff for the column player is its negation. x y z x 0 A T -c y A 0 b z c T b T 0 Reference: http://www.optimization-online.org/db FILE/2010/06/2659.pdf

Properties of Asymmetric Zero Sum Game The asymmetric payoff matrix, combining with the zero sum game condition, implies that the game value is zero. Proof: By zero sum: A = B and asymmetry: A = A T. The row player s utility is x T Ay Assume the game value v < 0 for the row player. Interprete v as the minimum value the row player can achieve no matter what is the strategy of the column player. Since the column s payoff is x T By = x T Ay = x T A T y = y T Ax, it has the same minimum value v. As the sum of their values must be zero, 2v = 0 implies v = 0.

How to obtain solution for LP? Solve Zero Sum Game, we have 0 A T c x A 0 b y c T b T 0 z x + y + z = 1, (x, y, z ) T 0 0 0 0 Dividing z on both side to obtain a solution for SDLP.

Proof of Correctness The payoff matrix is asymmetric. Combining it with the zero sum game condition, it implies that the game value is zero. It depends on z 0. If z = 0, we have A T y 0 and Ax 0, as well as c T x b T y, which implies unbounded LP or unbounded DLP unless c T x = b T y = 0. Bonus assignment: Find out how to deal with the case z = 0.

Solve Zero Sum Game using Linear Program A Game of two players can be described by their payoff matrices [A, B] 1 zero sum game: A + B = 0 2 Row player s problem max{x T Ay : x T e = 1, x 0} for fixed y 0, e T y = 1. the opponent will minimise the above item, therefore, the row player will choose max{w : x T A w e, e T x = 1, x 0} 3 Column player s problem min{x T Ay : y 0, e T y = 1} for fixed x 0, e T x = 1. min{z : Ay z e, e T y = 1, y 0} Those give the primal and dual of a linear program.

Solving Linear Program

Solving Linear Program Give a primal linear program, how do we solve it? 1 Simplex: going from one node to another 2 Ellipsoid: Make a ball contain a solution, cut ball in half, cover the half ball by a ball(ellipsoid), then continue 3 Interior point: move from one feasible solution to another. Polynomial time solution: ellipsoid and interior point.

Optimal solution can be reduced to feasibility 1 The optimum solution is found at a vertex of the polyhedra defined by {Ax b, x 0}. 2 A vertex can be written as a solution to a set of linear equations. 3 The representation of the vertex has at most the number of bits polynomial in the input data. 4 Do binary search on the optimum value of the linear program. The number iterations is bound by the number of bits of the vertex s representation.

2D Example Outline For the 2D LP find a feasible solution in Ax b, x 0. 1 Make a ball containing a point in the polyhedra 2 Do a linear transformation to move the ball to a unit circle centered at the origin. Ask if the center is a solution, if yes, done. Cut the unit circle into two halves at the center so that the feasible region lies at one side. Decide which sides of the cutting line the feasible region lies. Create a new ellipse cover the corresponding half of the unit circle. Repeat.

Estimate the number of iterations Equation of the center {(x, y) : x 2 + y 2 1} Equation for ellipse covering upper half of center {(x, y) : x2 + (y c)2 1} a 2 b 2 Calculate the new ellipse by setting it to touch three points on the unit circle ( 1, 0), (0, 1), (1, 0). The ratio will be 1/8 1/4 which is less than 1. The size converges geometrically and it is close to the exact solution with high precision (in value exponentially closer) after a polynomial number of iterations. We then round it to the exact solution.

Solving Linear Program Exercise: Solve the game of Rock-Paper-Sissors by LP Find your favorite LP solver and demonstrate how to solve the above. References http://www.optimizationonline.org/db FILE/2010/06/2659.pdf