Optimization 4. GAME THEORY

Similar documents
3. THE SIMPLEX ALGORITHM

Understanding the Simplex algorithm. Standard Optimization Problems.

Part IB Optimisation

3 The Simplex Method. 3.1 Basic Solutions

Review Solutions, Exam 2, Operations Research

Optimization - Examples Sheet 1

Part 1. The Review of Linear Programming

Chapter 1 Linear Programming. Paragraph 5 Duality

II. Analysis of Linear Programming Solutions

c) Place the Coefficients from all Equations into a Simplex Tableau, labeled above with variables indicating their respective columns

OPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES

OPTIMISATION /09 EXAM PREPARATION GUIDELINES

Lectures 6, 7 and part of 8

Lecture 7 Duality II

Introduction to Linear Programming

1 Primals and Duals: Zero Sum Games

1 Review Session. 1.1 Lecture 2

An introductory example

A Review of Linear Programming

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

Lecture 11: Post-Optimal Analysis. September 23, 2009

December 2014 MATH 340 Name Page 2 of 10 pages

"SYMMETRIC" PRIMAL-DUAL PAIR

The use of shadow price is an example of sensitivity analysis. Duality theory can be applied to do other kind of sensitivity analysis:

Foundations of Operations Research

SAMPLE QUESTIONS. b = (30, 20, 40, 10, 50) T, c = (650, 1000, 1350, 1600, 1900) T.

Chap6 Duality Theory and Sensitivity Analysis

15-780: LinearProgramming

Simplex method(s) for solving LPs in standard form

NOTES ON COOPERATIVE GAME THEORY AND THE CORE. 1. Introduction

Linear Programming: Chapter 5 Duality

The augmented form of this LP is the following linear system of equations:

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

Duality. Peter Bro Mitersen (University of Aarhus) Optimization, Lecture 9 February 28, / 49

Solutions to Exercises of Section III.1. a (13/4, 3) (22/4, 3/4). b (4, 10/4) (21/4, 2)

Math 210 Finite Mathematics Chapter 4.2 Linear Programming Problems Minimization - The Dual Problem

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Michælmas 2012 Operations Research III/IV 1

Slide 1 Math 1520, Lecture 10

AM 121: Intro to Optimization

Algorithmic Game Theory and Applications. Lecture 7: The LP Duality Theorem

Zero-Sum Games Public Strategies Minimax Theorem and Nash Equilibria Appendix. Zero-Sum Games. Algorithmic Game Theory.

COMP3121/9101/3821/9801 Lecture Notes. Linear Programming

Math 16 - Practice Final

Optimality, Duality, Complementarity for Constrained Optimization

Slack Variable. Max Z= 3x 1 + 4x 2 + 5X 3. Subject to: X 1 + X 2 + X x 1 + 4x 2 + X X 1 + X 2 + 4X 3 10 X 1 0, X 2 0, X 3 0

Farkas Lemma, Dual Simplex and Sensitivity Analysis

4.3 Minimizing & Mixed Constraints

Lecture Notes 3: Duality

Chapter 5 Linear Programming (LP)

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Ω R n is called the constraint set or feasible set. x 1

(P ) Minimize 4x 1 + 6x 2 + 5x 3 s.t. 2x 1 3x 3 3 3x 2 2x 3 6

3. Duality: What is duality? Why does it matter? Sensitivity through duality.


Linear Programming Duality P&S Chapter 3 Last Revised Nov 1, 2004

Linear Programming Duality

CSC304: Algorithmic Game Theory and Mechanism Design Fall 2016

56:270 Final Exam - May

Network Flows. CS124 Lecture 17

Linear Programming: Simplex

Lecture 2: The Simplex method

Dual Basic Solutions. Observation 5.7. Consider LP in standard form with A 2 R m n,rank(a) =m, and dual LP:

Optimisation and Operations Research

Sensitivity Analysis

Fundamentals of Operations Research. Prof. G. Srinivasan. Indian Institute of Technology Madras. Lecture No. # 15

Worked Examples for Chapter 5

Today: Linear Programming (con t.)

maxz = 3x 1 +4x 2 2x 1 +x 2 6 2x 1 +3x 2 9 x 1,x 2

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem

MVE165/MMG631 Linear and integer optimization with applications Lecture 5 Linear programming duality and sensitivity analysis

Summary of the simplex method

Part IB Optimisation

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Note 3: LP Duality. If the primal problem (P) in the canonical form is min Z = n (1) then the dual problem (D) in the canonical form is max W = m (2)

min 4x 1 5x 2 + 3x 3 s.t. x 1 + 2x 2 + x 3 = 10 x 1 x 2 6 x 1 + 3x 2 + x 3 14

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

ORF 522. Linear Programming and Convex Analysis

Prelude to the Simplex Algorithm. The Algebraic Approach The search for extreme point solutions.

M340(921) Solutions Problem Set 6 (c) 2013, Philip D Loewen. g = 35y y y 3.

March 13, Duality 3

MATH 445/545 Test 1 Spring 2016

21. Solve the LP given in Exercise 19 using the big-m method discussed in Exercise 20.

The simplex algorithm

Another max flow application: baseball

The Dual Simplex Algorithm

Lecture 10: Linear programming. duality. and. The dual of the LP in standard form. maximize w = b T y (D) subject to A T y c, minimize z = c T x (P)

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

Duality Theory, Optimality Conditions

The Simplex Method. Standard form (max) z c T x = 0 such that Ax = b.

Week 3 Linear programming duality

2018 년전기입시기출문제 2017/8/21~22

Lecture #21. c T x Ax b. maximize subject to

that nds a basis which is optimal for both the primal and the dual problems, given

Computational Game Theory Spring Semester, 2005/6. Lecturer: Yishay Mansour Scribe: Ilan Cohen, Natan Rubin, Ophir Bleiberg*

Today s class. Constrained optimization Linear programming. Prof. Jinbo Bi CSE, UConn. Numerical Methods, Fall 2011 Lecture 12

The Simplex Algorithm

Math Models of OR: The Revised Simplex Method

Brief summary of linear programming and duality: Consider the linear program in standard form. (P ) min z = cx. x 0. (D) max yb. z = c B x B + c N x N

END3033 Operations Research I Sensitivity Analysis & Duality. to accompany Operations Research: Applications and Algorithms Fatih Cavdur

Transcription:

Optimization GAME THEORY DPK Easter Term Saddle points of two-person zero-sum games We consider a game with two players Player I can choose one of m strategies, indexed by i =,, m and Player II can choose one of n strategies, indexed by j =,, n; if Player I plays strategy i and Player II plays strategy j the payoff to Player I is a i,j and the game is zero-sum in that what Player I wins Player II loses The payoff matrix A = (a i,j ) is given, and known to each player For example, consider the game in which Player I has choices of strategy and Player II has and the payoff matrix is: A = 5 5 To analyze the game, consider the worst outcome that can happen for Player I if he picks each of his strategies, or Looking along row we see that the minimum he can get, if II picks option, is ; if I picks his worst possibility is, which occurs when II picks or, and if I picks, his worst is 5 when II picks These are illustrated below, and if we consider similarly the worst possibilities for Player II, we get the column maxima in the table Player II chooses j Strategy Row min Player I chooses i 5 5 5 Col max 5 By taking the maximum of the row minima we see that Player I is guaranteed not to get less than the amount by choosing strategy, while, by considering the minimum of the column maxima, Player II is guaranteed not to lose more than by choosing his strategy 9

The upshot is that they will settle on the (,) element which is worth to Player I ( to Player II) and either player may be worse off if they deviate from the strategies indicated The amount that they settle on, here, is known as the value of the game and the element (,) of the matrix is a saddle point The matrix has a saddle point when min j max i a i,j = max i min j a i,j It is always the case that min j max i a i,j max i min j a i,j but not all matrices have a saddle point as the next example below illustrates For the row player I, we say that strategy i dominates strategy i if a i,j a i,j for all j =,, n; for the column player II, strategy j dominates strategy j if a i,j a i,j for all i =,, m A player will never play a strategy that is dominated by another (except possibly in the special case where the payoffs are identical for all outcomes) In the example above, for the row player strategy, dominates, while for the column player strategies and dominate Example Undercut In a version of the game Undercut, each player selects a number from,,, The players reveal their numbers and the player with the smaller number wins, unless the numbers are either adjacent, when the player with the larger number wins, or equal, when the game is tied with payoff zero The payoff matrix is Strategy Row min Col max Since all the row minima are negative and all the column maxima are positive there can be no saddle point This is an example of a symmetric game; its payoff matrix is an anti-symmetric matrix, that is A = A Mixed strategies In order to study games for which there is no saddle point, we widen the sets of strategies available to each player A mixed strategy for Player I is a set of probabilities

p = (p,, p m ), where p,, p m, and i= p i = ; if Player I uses the mixed strategy p he chooses row i with probability p i A mixed strategy for Player II is a set of n probabilities q = (q,, q n ), where q,, q n, and q j = with the analogous interpretation The expected payoff to Player I if Player II uses j is i p ia i,j, so Player I seeks to maximize [ min j ] p i a i,j i= This problem is equivalent to subject to j= p i =, p i, i m i= P: maximize v subject to p i a i,j v, j n, i= p i =, p i, i m i= If we write e = (,,, ) (we will use the same notation, e, for vectors of lengths m and n with all components equal to it will be clear from the context which is meant), then this problem may be written in matrix form as P: maximize v subject to A p ve, e p =, p This is of course a linear programming problem and it is easy to see that the dual problem is Player II s problem as he will wish to solve D: minimize v subject to which in matrix form is n a i,j q j v, i m, j= n q j =, q j, j n, j= D: minimize v subject to Aq ve, e q =, q Theorem If p, q and v satisfy: A p ve, e p =, p Aq ve, e q =, q v = p Aq (primal feasibility) (dual feasibility) (complementary slackness)

then p and q are optimal for P and D respectively, and v, the common value of the two problems, is the value of the game Proof This is just Lagrangian sufficiency again Suppose that v, p are feasible for P and w, q are feasible for D, then v v + q ( A p ve ) = q A p = w + p (Aq we) w, with equality throughout if we have v = w = v, p = p and q = q since we have complementary slackness, which (along with e p = = e q) is equivalent to q ( A p ve ) = = p (Aq ve), which completes the argument Remark The necessity part of the duality result for the pair of linear programming problems P and D is of course also true: that is, for any matrix A there exist p, q and v satisfying the conditions of Theorem This fact is sometimes known as the Fundamental Theorem of Matrix Games, but as we have seen it is just a consequence of linear programming duality It s conclusion is also equivalent to the statement that min max q p p Aq = max min p q p Aq, where the minimum is over q, e q = and the maximum over p, e p = Finding solutions for games There are several approaches to tackling the problem of finding solutions to games: Direct solution of a player s optimization problem For small matrices it may be possible to solve one of the players problems directly; for example, consider the payoff matrix

For Player it is clear that row dominates row so that an optimal row strategy will put no weight on row, and if we let p = p, p = p, p = the problem becomes one of trying to find the maximum v such that v and p satisfy p + ( p) = p v, p + ( p) = + p v, p + ( p) = + 7 p v, p Plotting the three lines as functions of p gives v + 7 p + p p p which shows that the largest value of v is when the first two lines meet (at p = ) giving v = 7 Thus we have the optimal strategy for Player I is (,, ) Notice that the line + 7 p is strictly greater that v at p =, thus Player II s optimal strategy must be of the form (q, q, ) and looking at the top row we get (using complementary slackness) that q + ( q) = v = 7 which gives q = Determining strategies satisfying the conditions of the Theorem In some cases features of the matrix may be used to help determine strategies which may be checked to be optimal using the sufficient conditions of the Theorem For example, for the Undercut game introduced above the matrix is, and since the game is symmetric (A = A ) the value is necessarily and any strategy optimal for Player I must also be optimal for Player II Suppose that p = (p, p, p, p )

is an optimal strategy, then if p >, complementary slackness will give (from the first row) p + p + p = or p + ( p p ) = () which gives p = ( p )/ If p >, we would have from the second row, p p + p = or p p + ( p p p ) =, () which gives p = ( + p )/9, after using the expression for p Similarly, from the third row, if p >, we must have p + p p =, () and we deduce that p = ( 8p )/ But we also have p = p p p which gives p = ( p )/9 Equating the two expressions for p shows that we must have p = This proves p = (,,, ) is an optimal strategy and it is the unique optimal strategy since its first three components are strictly positive, the argument above shows that any optimal strategy must satisfy (), () and () which has only one solution Using linear programming Although Player I s problem is a linear programming problem, it is not immediately in a form to which we could apply the simplex algorithm because v is unconstrained in sign However, if we add a constant c > min i,j a i,j, to each element of the matrix A, all the elements of A may be taken as positive, and we can assume that v > ; this will not change the optimal strategies and will just add c to the value Now set x i = p i /v and the row player s problem which is maximize v subject to A p ve, e p =, p, becomes which is equivalent to maximize v subject to A x e, e x = /v, x, minimize e x subject to A x e, x

This may now be tackled using the simplex algorithm The column player s problem may be reduced in the same way, by putting y i = q i /v, to maximize e y subject to Ay e, y ; this latter problem is somewhat simpler to solve by this method as it does not require the use of the Phase I procedure to find an initial bfs as the row player s problem would Consider the Undercut game and add to each element to obtain the matrix 5 5 5 ; setting up the tableau (with slack variables z, z, z, z ) we have * * * * y y y y z z z z z 5 5 z 5 z z Payoff Then a sequence of pivots as indicated, gives y y y y z z z z z y z z Payoff 7 5 7 5 5 7 y y y y z z z z z 8 y y z 7 9 Payoff 5

The final pivot, bringing in y and dropping z, gives the optimal solution y y y y z z z z y 7 8 y 5 8 y 5 z 5 8 8 8 5 8 7 8 8 7 6 7 5 8 Payoff 5 6 The optimal value of the linear programming problem is, which means the value of the game is (remember we added to all the components of the original symmetric game whose value was ) We can read off the optimal solution y =, y = 6, y = and y = These values must be scaled by the value to get the optimal strategy q =, q =, q = and q =, as we had obtained before Remarks Notice the symmetry between the payoff row and the right-hand side in the final tableau; minus the entries in the payoff row are the optimal solution to the dual problem because of the symmetry of the original game the optimal solutions to the primal and dual are the same, one giving the row player s strategy, the other the column player s For any game, when we solve one player s problem by linear programming, we may identify the other player s optimal strategy from the final payoff row It is clear from this section that solving a game by linear programming may not be tractable by hand except for the smallest problems However, for larger problems and with the aid of a computer, it may be the only way to fly 6 May