Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem

Similar documents
Algorithmic Game Theory and Applications. Lecture 7: The LP Duality Theorem

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Zero-Sum Games Public Strategies Minimax Theorem and Nash Equilibria Appendix. Zero-Sum Games. Algorithmic Game Theory.

Dot Products, Transposes, and Orthogonal Projections

Kevin James. MTHSC 3110 Section 2.1 Matrix Operations

Game Theory: Lecture 3

CSC304 Lecture 5. Game Theory : Zero-Sum Games, The Minimax Theorem. CSC304 - Nisarg Shah 1

CS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:

Lectures 6, 7 and part of 8

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 17: Duality and MinMax Theorem Lecturer: Sanjeev Arora

Matrices: 2.1 Operations with Matrices

Computing Minmax; Dominance

Game Theory. 2.1 Zero Sum Games (Part 2) George Mason University, Spring 2018

MS&E 246: Lecture 12 Static games of incomplete information. Ramesh Johari

Today: Linear Programming (con t.)

Game Theory and Algorithms Lecture 2: Nash Equilibria and Examples

Lecture 6: April 25, 2006

22A-2 SUMMER 2014 LECTURE Agenda

. =. a i1 x 1 + a i2 x 2 + a in x n = b i. a 11 a 12 a 1n a 21 a 22 a 1n. i1 a i2 a in

MTH5112 Linear Algebra I MTH5212 Applied Linear Algebra (2017/2018)

Linear Algebra and Matrix Inversion

Computing Minmax; Dominance

Prisoner s Dilemma. Veronica Ciocanel. February 25, 2013

Computational Game Theory Spring Semester, 2005/6. Lecturer: Yishay Mansour Scribe: Ilan Cohen, Natan Rubin, Ophir Bleiberg*

Matrices BUSINESS MATHEMATICS

Part IB Optimisation

Chapter 9. Mixed Extensions. 9.1 Mixed strategies

16.1 L.P. Duality Applied to the Minimax Theorem

Math 123, Week 2: Matrix Operations, Inverses

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

Optimization 4. GAME THEORY

Matrix Arithmetic. j=1

Understanding the Simplex algorithm. Standard Optimization Problems.

CS 246 Review of Linear Algebra 01/17/19

An introductory example

I = i 0,

EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium

Mon Feb Matrix algebra and matrix inverses. Announcements: Warm-up Exercise:

Topic 15 Notes Jeremy Orloff

Iterated Strict Dominance in Pure Strategies

MATH4250 Game Theory 1. THE CHINESE UNIVERSITY OF HONG KONG Department of Mathematics MATH4250 Game Theory

Optimization - Examples Sheet 1

Matrices. 1 a a2 1 b b 2 1 c c π

Introduction to Determinants

Volume in n Dimensions

b 1 b 2.. b = b m A = [a 1,a 2,...,a n ] where a 1,j a 2,j a j = a m,j Let A R m n and x 1 x 2 x = x n

CS261: A Second Course in Algorithms Lecture #12: Applications of Multiplicative Weights to Games and Linear Programs

Duality. Peter Bro Mitersen (University of Aarhus) Optimization, Lecture 9 February 28, / 49

Announcements Monday, October 02

ICS 6N Computational Linear Algebra Matrix Algebra

Announcements Wednesday, October 10

1 Seidel s LP algorithm

Interval values for strategic games in which players cooperate

Vector, Matrix, and Tensor Derivatives

Dynamic Programming: Matrix chain multiplication (CLRS 15.2)

Lower Semicontinuity of the Solution Set Mapping in Some Optimization Problems

Phys 201. Matrices and Determinants

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture 24 Scribe: Sachin Ravi May 2, 2013

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

ftdt. We often drop the parentheses and write Tf instead of Tf. We could also write in this example Tfx 0

15-780: LinearProgramming

MAT 1302B Mathematical Methods II

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

Game Theory. 4. Algorithms. Bernhard Nebel and Robert Mattmüller. May 2nd, Albert-Ludwigs-Universität Freiburg

April 29, 2010 CHAPTER 13: EVOLUTIONARY EQUILIBRIUM

Game Theory and Control

[ Here 21 is the dot product of (3, 1, 2, 5) with (2, 3, 1, 2), and 31 is the dot product of

Lecture December 2009 Fall 2009 Scribe: R. Ring In this lecture we will talk about

CS261: A Second Course in Algorithms Lecture #8: Linear Programming Duality (Part 1)

We could express the left side as a sum of vectors and obtain the Vector Form of a Linear System: a 12 a x n. a m2

1 Motivation. Game Theory. 2 Linear Programming. Motivation. 4. Algorithms. Bernhard Nebel and Robert Mattmüller May 15th, 2017

Problem 1: (3 points) Recall that the dot product of two vectors in R 3 is

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

Matrix Algebra & Elementary Matrices

MATH 320, WEEK 7: Matrices, Matrix Operations

Matrix Inverses. November 19, 2014

Matrix Multiplication

The Distribution of Optimal Strategies in Symmetric Zero-sum Games

Mixed Nash Equilibria

April 25 May 6, 2016, Verona, Italy. GAME THEORY and APPLICATIONS Mikhail Ivanov Krastanov

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

Msc Micro I exam. Lecturer: Todd Kaplan.

Elementary maths for GMT

1 Basic Game Modelling

Normal-form games. Vincent Conitzer

Introduction to Game Theory Lecture Note 2: Strategic-Form Games and Nash Equilibrium (2)

4: Dynamic games. Concordia February 6, 2017

12. Perturbed Matrices

General-sum games. I.e., pretend that the opponent is only trying to hurt you. If Column was trying to hurt Row, Column would play Left, so

MS&E 246: Lecture 4 Mixed strategies. Ramesh Johari January 18, 2007

Notes on Blackwell s Comparison of Experiments Tilman Börgers, June 29, 2009

Algorithmic Game Theory and Applications. Lecture 5: Introduction to Linear Programming

Industrial Organization Lecture 3: Game Theory

Matrices and Vectors

Matrices. Chapter Definitions and Notations

Selfishness vs Altruism vs Balance

Two matrices of the same size are added by adding their corresponding entries =.

Game Theory: Lecture 2

Matrix Arithmetic. a 11 a. A + B = + a m1 a mn. + b. a 11 + b 11 a 1n + b 1n = a m1. b m1 b mn. and scalar multiplication for matrices via.

Math 240 Calculus III

Transcription:

Algorithmic Game Theory and Applications Lecture 4: 2-player zero-sum games, and the Minimax Theorem Kousha Etessami

2-person zero-sum games A finite 2-person zero-sum (2p-zs) strategic game Γ, is a strategic game where: For players i {1,2}, the payoff functions u i : S R are such that for all s = (s 1, s 2 ) S, u 1 (s) + u 2 (s) = 0 1 I.e., u 1 (s) = u 2 (s). u i (s 1,s 2 ) can conveniently be viewed as a m 1 m 2 payoff matrix A i, where: A 1 = u 1 (1, 1)..... u 1 (1,m 2 ).... u 1 (m 1, 1)... u 1 (m 1,m 2 ) Note, A 2 = A 1. Thus we may assume only one function u(s 1,s 2 ) is given, as one matrix, A = A 1. Player 1 wants to maximize u(s), while Player 2 wants to minimize it (i.e., to maximize its negative).

matrices and vectors As just noted, a 2p-zs game can be described by an m 1 m 2 matrix: A = 2 a 1,1.... a 1,m2.... a i,j.... a m1,1.... a m1,m 2 where a i,j = u(i,j). For any (n 1 n 2 )-matrix A we ll either use a i,j or (A) i,j to denote the entry in the i th row and j th column of A. For (n 1 n 2 ) matrices A and B, let A B denote that for all i, j, a i,j b i,j. Let A > B denote that for all i, j, a i,j > b i,j. For a matrix A, let A 0 denote that every entry is 0. Likewise, let A > 0 mean every entry is > 0.

more review of matrices and vectors Recall matrix multiplication: given (n 1 n 2 )-matrix A and (n 2 n 3 )-matrix B, the product AB is an (n 1 n 3 )-matrix C, where 3 c i,j = n 2 k=1 a i,k b k,j Fact: matrix multiplication is associative : i.e., (AB)C = A(BC) (Note: for the multiplications to be defined, the dimensions of the matrices A, B, and C need to be consistent : (n 1 n 2 ), (n 2 n 3 ), and (n 3 n 4 ), respectively.) Fact: For matrices A, B, C, of appropriate dimensions, if A B, and C 0, then AC BC, and likewise, CA CB. (C s dimensions might be different in each case.)

more on matrices and vectors For a (n 1 n 2 ) matrix B, let B T denote the (n 2 n 1 ) transpose matrix, where (B T ) i,j := (B) j,i. y(1) We can view a column vector, y =.., as a y(m) (m 1)-matrix. Then, y T would be a (1 m)-matrix, i.e., a row vector. Typically, we think of vectors as column vectors and explicitly transpose them if we need to. We ll call a length m vector an m-vector. Multiplying a (n 1 n 2 )-matrix A by a n 2 -vector y is just a special case of matrix multiplication: Ay is a n 1 -vector. Likewise, pre-multiplying A, by a n 1 -row vector y T, is also just a special case of matrix multiplication: y T A is a n 2 -row vector. For a column (row) vector y, we use (y) j to denote the entry (y) j,1 (respectively, (y) 1,j ). 4

A matrix view of zero-sum games Suppose we have a 2p-zs game given by a (m 1 m 2 )-matrix, A. Suppose Player 1 chooses a mixed strategy x 1, and Player 2 chooses mixed strategy x 2 (assume x 1 and x 2 are given by column vectors). Consider the product x T 1 Ax 2 If you do the calculation, x T 1 Ax 2 = m 1 i=1 m 2 j=1 5 (x 1 (i) x 2 (j)) a i,j But note that (x 1 (i) x 2 (j)) is precisely the probability of the pure combination s = (i,j). Thus, for the mixed profile x = (x 1, x 2 ) x T 1 Ax 2 = U 1 (x) = U 2 (x) where U 1 (x) is the expected payoff (which Player 1 is trying to maximize, and Player 2 is trying to minimize).

minmaximizing strategies Suppose Player 1 chooses a mixed strategy x 1 X 1, by trying to maximize the worst that can happen. The worst that can happen would be for Player 2 to choose x 2 which minimizes (x 1) T Ax 2. Definition: x 1 X 1 is a minmaximizer for Player 1 if 6 min (x 1) T Ax 2 = max min (x 1 ) T Ax 2 x 2 X 2 x 1 X 1 x 2 X 2 Similarly, x 2 X 2 is a maxminimizer for Player 2 if max x 1 X 1 (x 1 ) T Ax 2 = min x 2 X 2 max x 1 X 1 x T 1 Ax 2 Note that min x 2 X 2 (x 1) T Ax 2 (x 1) T Ax 2 max x 1 X 1 x T 1 Ax 2 Amazingly, von Neumann (1928) showed equality holds!

The Minimax Theorem Theorem(von Neumann) Let a 2p-zs game Γ be given by an (m 1 m 2 )-matrix A of real numbers. There exists a unique value v R, such that there exists x = (x 1, x 2) X such that 1. ((x 1) T A) j v, for j = 1,...,m 2. 2. (Ax 2) j v, for j = 1,...,m 1. 3. And (thus) v = (x 1) T Ax 2 and max min (x 1 ) T Ax 2 = v = min max x T 1 Ax 2 x 1 X 1 x 2 X 2 x 2 X 2 x 1 X 1 4. In fact, the above conditions all hold precisely when x = (x 1,x 2) is any Nash Equilibrium. Equivalently, they hold precisely when x 1 is any minmaximizer and x 2 is any maxminimizer. 7

Note: some remarks (1.) says x 1 guarantees Player 1 at least expected profit v, and (2.) says x 2 guarantees Player 2 at most expected loss v. We call any such x = (x 1,x 2) a minimax profile. We call the unique v the minimax value of game Γ. It is obvious that the maximum profit that Player 1 can guarantee for itself should be the minimum loss that Player 2 can guarantee for itself, i.e., that 8 max min (x 1 ) T Ax 2 min max x T 1 Ax 2 x 1 X 1 x 2 X 2 x 2 X 2 x 1 X 1 What is not obvious at all is why these two values should be the same!

Proof of the Minimax Theorem The Minimax Theorem follows directly from Nash s Theorem (but historically, it predates Nash). Proof: Let x = (x 1, x 2) X be a NE of the 2-player zero-sum game Γ, with matrix A. Let v := (x 1) T Ax 2 = U 1 (x ) = U 2 (x ). Since x 1 and x 2 are best responses to each other, we know that for i {1,2} 9 U i (x i; π i,j ) U i (x ) But 1. U 1 (x 1; π 1,j ) = (Ax 2) j. Thus, (Ax 2) j v = U 1 (x ) for all j = 1,...,m 1. 2. U 2 (x 2; π 2,j ) = ((x 1) T A) j. Thus, ((x 1) T A) j v = U 2 (x ) for all j = 1,...,m 2.

3. max x1 X 1 (x 1 ) T Ax 2 v because (x 1 ) T Ax 2 is a weighted average of (Ax 2) j s. Similarly, v min x2 X 2 (x 1) T Ax 2 because (x 1) T Ax 2 is a weighted average of ((x 1) T A) j s. Thus 10 max (x 1 ) T Ax 2 v min (x 1) T Ax 2 x 1 X 1 x 2 X 2 We earlier noted the opposite inequalities, so, min max x T 1 Ax 2 = v = max min (x 1 ) T Ax 2 x 2 X 2 x 1 X 1 x 1 X 1 x 2 X 2 4. We didn t assume anything about the particular Nash Equilibrium we chose. So, for every NE, x, letting v = (x 1) T Ax 2, max min (x 1 ) T Ax 2 = v = v = min max x T 1 Ax 2 x 1 X 1 x 2 X 2 x 2 X 2 x 1 X 1 Moreover, if x = (x 1, x 2) satisfies conditions (1) and (2) for some v, then x must be a Nash Equilibrium. (This will be a homework exercise.) Q.E.D. (Minimax Theorem)

remarks and food for thought Thus, for 2-player zero-sum games, Nash Equilibria and Minimax profiles are the same thing. Let us note here Useful Corollary for Minimax: In a minimax profile x = (x 1,x 2), 1. if x 2(j) > 0 then ((x 1) T A) j = (x 1) T Ax 2 = v. 2. if x 1(j) > 0 then (Ax 2) j = (x 1) T Ax 2 = v. This is an immediate consequence of the Useful Corollary for Nash Equilibria. If you were playing a 2-player zero-sum game (say, as player 1) would you always play a minmaximizer strategy? What if you were convinced your opponent is an idiot? Notice, we have no clue yet how to compute the minimax value and a minimax profile. That is about to change. 11

minimax as an optimization problem Consider the following optimization problem : Maximize v Subject to constraints: (x T 1 A) j v for j = 1,...,m 2, x 1 (1) +... + x 1 (m 1 ) = 1, x 1 (j) 0 for j = 1,...,m 1 It follows from the minimax theorem that an optimal solution (x 1,v ) would give precisely the minimax value v, and a minmaximizer x 1 for Player 1. We are optimizing a linear objective, under linear constraints (or linear inequalities ). 12 That s what Linear Programming is. Fortunately, we have good algorithms for it. Next time, we start Linear Programming.