Computing Solution Concepts of Normal-Form Games. Song Chong EE, KAIST

Similar documents
Algorithmic Game Theory

Computing Minmax; Dominance

Computing Minmax; Dominance

Game Theory. 4. Algorithms. Bernhard Nebel and Robert Mattmüller. May 2nd, Albert-Ludwigs-Universität Freiburg

1 Motivation. Game Theory. 2 Linear Programming. Motivation. 4. Algorithms. Bernhard Nebel and Robert Mattmüller May 15th, 2017

Basic Game Theory. Kate Larson. January 7, University of Waterloo. Kate Larson. What is Game Theory? Normal Form Games. Computing Equilibria

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Game Theory: Lecture 3

Normal-form games. Vincent Conitzer

Lecture December 2009 Fall 2009 Scribe: R. Ring In this lecture we will talk about

Iterated Strict Dominance in Pure Strategies

General-sum games. I.e., pretend that the opponent is only trying to hurt you. If Column was trying to hurt Row, Column would play Left, so

Game Theory Fall 2003

Optimality, Duality, Complementarity for Constrained Optimization

Advanced Microeconomics

Game Theory for Linguists

Notes on Coursera s Game Theory

Distributed Optimization. Song Chong EE, KAIST

NASH IMPLEMENTATION USING SIMPLE MECHANISMS WITHOUT UNDESIRABLE MIXED-STRATEGY EQUILIBRIA

6.891 Games, Decision, and Computation February 5, Lecture 2

Efficient Nash Equilibrium Computation in Two Player Rank-1 G

New Complexity Results about Nash Equilibria

M340(921) Solutions Practice Problems (c) 2013, Philip D Loewen

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Today: Linear Programming (con t.)

1 Basic Game Modelling

Lectures 6, 7 and part of 8

Long-Run versus Short-Run Player

Review of topics since what was covered in the midterm: Topics that we covered before the midterm (also may be included in final):

Game Theory DR. ÖZGÜR GÜRERK UNIVERSITY OF ERFURT WINTER TERM 2012/13. Strict and nonstrict equilibria

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games

Chapter 9. Mixed Extensions. 9.1 Mixed strategies

C31: Game Theory, Lecture 1

Near-Potential Games: Geometry and Dynamics

MS&E 246: Lecture 4 Mixed strategies. Ramesh Johari January 18, 2007

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

Zero-Sum Games Public Strategies Minimax Theorem and Nash Equilibria Appendix. Zero-Sum Games. Algorithmic Game Theory.

Exclusion Method for Finding Nash Equilibrium in Multiplayer Games

Symmetries and the Complexity of Pure Nash Equilibrium

Title: The Castle on the Hill. Author: David K. Levine. Department of Economics UCLA. Los Angeles, CA phone/fax

Equilibria in Games with Weak Payoff Externalities

Two-Player Kidney Exchange Game

"Arrow s Theorem and the Gibbard-Satterthwaite Theorem: A Unified Approach", by Phillip Reny. Economic Letters (70) (2001),

Network Flows. CS124 Lecture 17

Massachusetts Institute of Technology 6.854J/18.415J: Advanced Algorithms Friday, March 18, 2016 Ankur Moitra. Problem Set 6

Weak Dominance and Never Best Responses

First Prev Next Last Go Back Full Screen Close Quit. Game Theory. Giorgio Fagiolo

Computational complexity of solution concepts

CS 598RM: Algorithmic Game Theory, Spring Practice Exam Solutions

Algorithmic Game Theory and Economics: A very short introduction. Mysore Park Workshop August, 2012

Ex Post Cheap Talk : Value of Information and Value of Signals

CO759: Algorithmic Game Theory Spring 2015

Static (or Simultaneous- Move) Games of Complete Information

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem

Mathematical Economics - PhD in Economics

Definitions and Proofs

FPTAS for Computing a Symmetric Leontief Competitive Economy Equilibrium

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India August 2012

Near-Potential Games: Geometry and Dynamics

Discrete Mathematics

Potential Games. Krzysztof R. Apt. CWI, Amsterdam, the Netherlands, University of Amsterdam. Potential Games p. 1/3

Endre Boros b Vladimir Gurvich d ;

Mixed Strategies. Krzysztof R. Apt. CWI, Amsterdam, the Netherlands, University of Amsterdam. (so not Krzystof and definitely not Krystof)

Introduction. 1 University of Pennsylvania, Wharton Finance Department, Steinberg Hall-Dietrich Hall, 3620

Microeconomics for Business Practice Session 3 - Solutions

Nash Equilibrium: Existence & Computation

CS364A: Algorithmic Game Theory Lecture #13: Potential Games; A Hierarchy of Equilibria

1 Equilibrium Comparisons

Theoretical Computer Science

CMSC 474, Introduction to Game Theory Maxmin and Minmax Strategies

Game Theory and Algorithms Lecture 2: Nash Equilibria and Examples

Exact and Approximate Equilibria for Optimal Group Network Formation

Exact and Approximate Equilibria for Optimal Group Network Formation

Game theory: Models, Algorithms and Applications Lecture 4 Part II Geometry of the LCP. September 10, 2008

An introductory example

CSC304: Algorithmic Game Theory and Mechanism Design Fall 2016

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

Lecture 6: April 25, 2006

האוניברסיטה העברית בירושלים

Motivating examples Introduction to algorithms Simplex algorithm. On a particular example General algorithm. Duality An application to game theory

Spring 2016 Network Science. Solution of Quiz I

15-780: LinearProgramming

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 17: Duality and MinMax Theorem Lecturer: Sanjeev Arora

1 Column Generation and the Cutting Stock Problem

A Polynomial-time Nash Equilibrium Algorithm for Repeated Games

Computational Aspects of Shapley s Saddles

MS&E 246: Lecture 17 Network routing. Ramesh Johari

Computing Equilibria by Incorporating Qualitative Models 1

Price of Stability in Survivable Network Design

of a bimatrix game David Avis McGill University Gabriel Rosenberg Yale University Rahul Savani University of Warwick

Economics and Computation

Mechanism Design for Bounded Agents

A Technique for Reducing Normal Form Games to Compute a Nash Equilibrium

Game Theory and Algorithms Lecture 7: PPAD and Fixed-Point Theorems

April 29, 2010 CHAPTER 13: EVOLUTIONARY EQUILIBRIUM

The Effect of Changes in Risk Attitude on Strategic Behavior

NETS 412: Algorithmic Game Theory March 28 and 30, Lecture Approximation in Mechanism Design. X(v) = arg max v i (a)

Linear Programming: Simplex

SF2972 Game Theory Exam with Solutions March 15, 2013

Combinatorial Agency of Threshold Functions

Transcription:

Computing Solution Concepts of Normal-Form Games Song Chong EE, KAIST songchong@kaist.edu

Computing Nash Equilibria of Two-Player, Zero-Sum Games Can be expressed as a linear program (LP), which means that equilibria can be computed in polynomial time Let U i be the expected utility for player i in equilibrium (the value of the game). U i = U 2. The minimax theorem tells us that U 1 holds constant in all equilibria and that it is the same as the value that player 1 achieves under a minmax strategy by player 2 Using this result, construct an LP (player 2 s program) that gives us player 2 s mixed strategy in equilibrium 1

Computing Nash Equilibria of Two-Player, Zero-Sum Games In the same fashion, construct an LP (player 1 s program) to give us player 1 s mixed strategies. This program reverses the roles of player 1 and player 2 in the constraints; the objective is to maximize U 1, as player 1 wants to maximize his own payoffs. This corresponds to the dual of player 2 s program 2

Computing Nash Equilibria of Two-Player, Zero-Sum Games Finally, construct a formulation equivalent to our first linear program by introducing slack variables r 1 j for every j A 1 and then replacing the inequality constraints with equality constraints 3

Computing Nash Equilibria of Two-Player, General-Sum Games Unfortunately, the problem of finding a Nash equilibrium of a two-player, general-sum game cannot be formulated as a linear program. This is because the two players interests are no longer diametrically opposed. Thus, we cannot state our problem as an optimization problem However, the problem of finding a Nash equilibrium of a two-player, general-sum game can be formulated as a linear complementarity problem (LCP). The LCP has no objective function at all, and is thus a constraint satisfaction problem, or a feasibility program, rather than an optimization problem 4

Linear Complementarity Problem (LCP) We would still have a linear program if we do not include the last nonlinear constraint. However, we would also have a flaw in our formulation: U 1 and U 2 would be allowed to take unboundedly large values, because all of these constraints remain satisfied when both U i and r i j are increased by the same constant, for any given i and j. This is why we add the nonlinear constraint, called the complementarity condition Why does the complementarity condition fix our problem formulation? The condition requires that the slack variable r i j is zero exactly when its corresponding action a i j is a best response to the mixed strategy s i So, the condition captures the fact that, in equilibrium, all strategies that are played with positive probability must yield the same expected payoff, while all strategies that lead to lower expected payoffs are not played. That is, each player plays a best response to the other player s mixed strategy, which is the definition of a Nash equilibrium The Lemke Howson algorithm is the best-known algorithm designed to solve the LCP problem and is guaranteed to find a sample Nash equilibrium Theorem 4.2.1 The problem of finding a sample Nash equilibrium of a general-sum finite game with two or more players is PPAD-complete (a less familiar complexity class) PPAD-complete has shown to be neither NP-complete nor easier than NP-complete. However, it is generally believed that PPAD P and, in the worst case, it will take exponential time 5

Beyond Sample NE Computation In two-player, general-sum games, instead of just searching for a sample equilibrium, we may want to find an equilibrium with a specific property as follows 1. (Uniqueness) Given a game G, does there exist a unique equilibrium in G? 2. (Pareto optimality) Given a game G, does there exist a strictly Pareto efficient equilibrium in G? 3. (Guaranteed payoff) Given a game G and a value v, does there exist an equilibrium in G in which some player i obtains an expected payoff of at least v? 4. (Guaranteed social welfare) Given a game G, does there exist an equilibrium in which the sum of agents utilities is at least k? 5. (Action inclusion) Given a game G and an action a i A i for some player i N, does there exist an equilibrium of G in which player i plays action a i with strictly positive probability? 6. (Action exclusion) Given a game G and an action a i A i for some player i N, does there exist an equilibrium of G in which player i plays action a i with zero probability? 6

Beyond Sample NE Computation Unfortunately, all of these questions are hard in the worst case Theorem 4.2.3 The following problems are NP-hard when applied to Nash equilibria: uniqueness, Pareto optimality, guaranteed payoff, guaranteed social welfare, action inclusion, and action exclusion We may also want to determine all Nash equilibria Theorem 4.2.4 Computing all of the Nash equilibria of a two-player, general-sum game requires worst-case time that is exponential in the number of actions for each player 7

Computing Nash Equilibria of n-player, General-Sum Games For n-player games where n 3, the problem of finding a Nash equilibrium can no longer be represented even as an LCP While it does allow a formulation as a nonlinear complementarity problem, such problems are often hopelessly impractical to solve exactly 8

Computing Maxmin and Minmax Strategies for Two-Player, General-Sum Games In a two-player, general-sum game, a maxmin strategy for player i is a strategy that maximizes his worst-case payoff, presuming that the other player j follows the strategy that will cause the greatest harm to i A minmax strategy for j against i is such a maximum-harm strategy Maxmin and minmax strategies can be computed in polynomial time because they correspond to Nash equilibrium strategies in related zero-sum games How to compute a maxmin strategy for player 1? Let G be an arbitrary two-player game G = ( 1,2, A 1 A 2, (u 1, u 2 )) and define a related zerosum game G = ( 1,2, A 1 A 2, (u 1, u 1 )) By the minmax theorem, every strategy for player 1 which is part of a Nash equilibrium strategy profile for G is a maxmin strategy for player 1 in G By definition, player 1 s maxmin strategy is independent of player 2 s utility function. Thus, player 1 s maxmin strategy is the same in G and in G Problem of finding a maxmin strategy in G thus reduces to finding a Nash equilibrium of G, which can be solved by an LP The computation of minmax strategies follows the same way 9

Identifying Dominated Strategies Iterated removal of strictly dominated strategies The same set of strategies will survive regardless of the elimination order, and all Nash equilibria of the original game will be contained in this set Thus, this method can be used to narrow down the set of strategies to consider before attempting to identify a sample Nash equilibrium In the worst case this procedure will have no effect many games have no dominated strategies. In practice, however, it can make a big difference Iterated removal of weakly or very weakly dominated strategies The set of strategies that survive iterated removal can differ depending on the order in which dominated strategies are removed. As a consequence, removing weakly or very weakly dominated strategies can eliminate some equilibria of the original game However, since no new equilibria are ever created by this elimination and every game has at least one equilibrium, at least one of the original equilibria always survives. This is enough if all we want to do is to identify a sample Nash equilibrium Furthermore, iterative removal of weakly or very weakly dominated strategies can eliminate a larger set of strategies than iterative removal of strictly dominated strategies and so will often produce a smaller game The complexity of determining whether a given strategy can be removed depends on Domination by a pure or mixed strategies Strict, weak or very weak domination Only domination or survival under iterated removal of dominated strategies 10

Domination by a Pure Strategy Algorithm for determining whether s i is strictly dominated by any pure strategy Need to check every pure strategy a i for player i and every pure-strategy profile a i for the other players, meaning that we do not need to check every mixed strategy profile of the other players This is because if u i s i, a i < u i a i, a i for every pure-strategy profile a i A i, then there cannot exist any mixed-strategy profile s i S i for which u i s i, s i u i a i, s i due to the linearity of expectation The case of weak or very weak dominance can also be tested using essentially the same algorithm with some modification For all of the definitions of domination, the complexity of the procedure is O A, linear in the size of the normal-form game 11

Domination by a Mixed Strategy Cannot use a simple algorithm like the one used to test domination by a pure strategy because mixed strategies cannot be enumerated. However, it turns out that we can still answer the question in polynomial time by solving a linear program Without loss of generality, assume that player i s utilities are strictly positive A feasibility program to describe strict domination by a mixed strategy This formulation, however, does not constitute a linear program because the constraints in linear programs must be weak inequalities 12

Instead, we must use the LP that follows Domination by a Mixed Strategy This linear program simulates the strict inequality of previous formulation through the objective function Because no constraints restrict the p j s from above, this LP will always be feasible. However, in the optimal solution the p j s may not sum to 1 In the optimal solution, we will have σ j Ai p j u i a j, a i = u i (s i, a i ) for at least some a i A i A strictly dominating mixed strategy therefore exists if and only if the optimal solution to the LP has objective function value strictly less than 1 In this case, we can add a positive amount to each p j in order to cause the weak inequality constraint to hold in its strict version everywhere while achieving the condition σ j p j = 1 13

Domination by a Mixed Strategy A feasibility program to describe very weak domination by a mixed strategy An LP to check weak domination by a mixed strategy can be derived by adding an objective function to the above feasibility program 14

Domination by a Mixed Strategy Because of constraint (4.49), any feasible solution will have a nonnegative objective value If the optimal solution has a strictly positive objective, the mixed strategy given by the p j s achieves strictly positive expected utility for at least one a i A i, meaning that s i is weakly dominated by this mixed strategy Observe that all of the linear programs above can be modified to check whether a strategy s i is strictly dominated by any mixed strategy that only places positive probability on some subset of i s actions T A i. This can be achieved simply by replacing all occurrences of A i by T in the linear programs 15

Iterated Dominance We only consider pure strategies as candidates for removal; indeed, as it turns out, it never helps to remove dominated mixed strategies when performing iterated removal It is important, however, that we consider the possibility that pure strategies may be dominated by mixed strategies For all three flavors of domination, it requires only polynomial time to iteratively remove dominated strategies until the game has been maximally reduced A single step of this process consists of checking whether a pure strategy of a player is dominated by any other mixed strategy, which can be answered in polynomial time by solving solve a linear program Each step removes one pure strategy for one player, so there can be at most σ i N A i steps 16

Iterated Dominance Computational questions regarding which strategies remain in reduced games 1. (Strategy elimination) Does there exist some elimination path under which the strategy s i is eliminated? 2. (Reduction identity) Given action subsets A i A i for each player i, does there exist a maximally reduced game where each player has the actions in A i? 3. (Reduction size) Given constants k i for each player i, does there exist a maximally reduced game where each player i has exactly k i actions Theorem 4.5.1 For iterated strict dominance, the strategy elimination, reduction identity and reduction size problems are in P. For iterated weak dominance, these problems are NP-complete 17