MS&E 246: Lecture 4 Mixed strategies. Ramesh Johari January 18, 2007

Similar documents
MS&E 246: Lecture 17 Network routing. Ramesh Johari

MS&E 246: Lecture 12 Static games of incomplete information. Ramesh Johari

MS&E 246: Lecture 18 Network routing. Ramesh Johari

1 The General Definition

Game Theory DR. ÖZGÜR GÜRERK UNIVERSITY OF ERFURT WINTER TERM 2012/13. Strict and nonstrict equilibria

Economics 703 Advanced Microeconomics. Professor Peter Cramton Fall 2017

Mean-field equilibrium: An approximation approach for large dynamic games

Lecture Notes on Game Theory

Non-zero-sum Game and Nash Equilibarium

First Prev Next Last Go Back Full Screen Close Quit. Game Theory. Giorgio Fagiolo

6.891 Games, Decision, and Computation February 5, Lecture 2

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

1 Lattices and Tarski s Theorem

Basic Game Theory. Kate Larson. January 7, University of Waterloo. Kate Larson. What is Game Theory? Normal Form Games. Computing Equilibria

Game Theory and Algorithms Lecture 7: PPAD and Fixed-Point Theorems

Game theory Lecture 19. Dynamic games. Game theory

Mixed Nash Equilibria

Game Theory Lecture 10+11: Knowledge

Lecture 6: April 25, 2006

Outline for today. Stat155 Game Theory Lecture 17: Correlated equilibria and the price of anarchy. Correlated equilibrium. A driving example.

Econ 618: Correlated Equilibrium

Game Theory and Rationality

Computing Minmax; Dominance

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games

Industrial Organization Lecture 3: Game Theory

CS364A: Algorithmic Game Theory Lecture #13: Potential Games; A Hierarchy of Equilibria

Belief-based Learning

: Cryptography and Game Theory Ran Canetti and Alon Rosen. Lecture 8

1 Equilibrium Comparisons

Game Theory: introduction and applications to computer networks

Computing Solution Concepts of Normal-Form Games. Song Chong EE, KAIST

6.207/14.15: Networks Lecture 11: Introduction to Game Theory 3

Game Theory. Professor Peter Cramton Economics 300

Recap Social Choice Fun Game Voting Paradoxes Properties. Social Choice. Lecture 11. Social Choice Lecture 11, Slide 1

Randomized Nash Equilibrium

Quantum Games. Quantum Strategies in Classical Games. Presented by Yaniv Carmeli

Mixed Strategies. Krzysztof R. Apt. CWI, Amsterdam, the Netherlands, University of Amsterdam. (so not Krzystof and definitely not Krystof)

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem

Game Theory Lecture 7 Refinements: Perfect equilibrium

Computing Minmax; Dominance

Iterated Strict Dominance in Pure Strategies

Outline for today. Stat155 Game Theory Lecture 16: Evolutionary game theory. Evolutionarily stable strategies. Nash equilibrium.

A Price-Based Approach for Controlling Networked Distributed Energy Resources

Long-Run versus Short-Run Player

Game Theory. Wolfgang Frimmel. Perfect Bayesian Equilibrium

6.207/14.15: Networks Lecture 10: Introduction to Game Theory 2

News. Good news. Bad news. Ugly news

Games with Perfect Information

Game Theory for Linguists

Other Equilibrium Notions

Normal-form games. Vincent Conitzer

Ph.D. Preliminary Examination MICROECONOMIC THEORY Applied Economics Graduate Program June 2016

Introduction to game theory LECTURE 1

Economics 209B Behavioral / Experimental Game Theory (Spring 2008) Lecture 3: Equilibrium refinements and selection

A Polynomial-time Nash Equilibrium Algorithm for Repeated Games

Game Theory Correlated equilibrium 1

Recap Social Choice Functions Fun Game Mechanism Design. Mechanism Design. Lecture 13. Mechanism Design Lecture 13, Slide 1

Government 2005: Formal Political Theory I

Computation of Efficient Nash Equilibria for experimental economic games

Confronting Theory with Experimental Data and vice versa. European University Institute. May-Jun Lectures 7-8: Equilibrium

Lecture December 2009 Fall 2009 Scribe: R. Ring In this lecture we will talk about

NASH IMPLEMENTATION USING SIMPLE MECHANISMS WITHOUT UNDESIRABLE MIXED-STRATEGY EQUILIBRIA

Game Theory and Algorithms Lecture 2: Nash Equilibria and Examples

Introduction to Game Theory Lecture Note 2: Strategic-Form Games and Nash Equilibrium (2)

EC319 Economic Theory and Its Applications, Part II: Lecture 2

Introduction to Game Theory

Nash Equilibrium: Existence & Computation

15 Real Analysis II Sequences and Limits , Math for Economists Fall 2004 Lecture Notes, 10/21/2004

Notes on Coursera s Game Theory

Near-Potential Games: Geometry and Dynamics

Equilibria in Games with Weak Payoff Externalities

A Few Games and Geometric Insights

EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium

General-sum games. I.e., pretend that the opponent is only trying to hurt you. If Column was trying to hurt Row, Column would play Left, so

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Msc Micro I exam. Lecturer: Todd Kaplan.

BELIEFS & EVOLUTIONARY GAME THEORY

Efficient Nash Equilibrium Computation in Two Player Rank-1 G

Higher Order Beliefs in Dynamic Environments

Game Theory and Control

WEAKLY DOMINATED STRATEGIES: A MYSTERY CRACKED

The effect of learning on membership and welfare in an International Environmental Agreement

CS364A: Algorithmic Game Theory Lecture #16: Best-Response Dynamics

Game Theory. Strategic Form Games with Incomplete Information. Levent Koçkesen. Koç University. Levent Koçkesen (Koç University) Bayesian Games 1 / 15

A (Brief) Introduction to Game Theory

We set up the basic model of two-sided, one-to-one matching

Solving Extensive Form Games

Area I: Contract Theory Question (Econ 206)

Variants of Nash Equilibrium. CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte

The Canonical Reputation Model and Reputation Effects

Weak Dominance and Never Best Responses

Bargaining, Contracts, and Theories of the Firm. Dr. Margaret Meyer Nuffield College

ON FORWARD INDUCTION

Extensive Form Games I

Moral hazard in teams

STOCHASTIC STABILITY OF GROUP FORMATION IN COLLECTIVE ACTION GAMES. Toshimasa Maruta 1 and Akira Okada 2

Bounded Rationality, Strategy Simplification, and Equilibrium

Title: The Castle on the Hill. Author: David K. Levine. Department of Economics UCLA. Los Angeles, CA phone/fax

A Note on the Existence of Ratifiable Acts

Advanced Microeconomics

Transcription:

MS&E 246: Lecture 4 Mixed strategies Ramesh Johari January 18, 2007

Outline Mixed strategies Mixed strategy Nash equilibrium Existence of Nash equilibrium Examples Discussion of Nash equilibrium

Mixed strategies Notation: Given a set X, we let Δ(X) denote the set of all probability distributions on X. Given a strategy space S i for player i, the mixed strategies for player i are Δ(S i ). Idea: a player can randomize over pure strategies.

Mixed strategies How do we interpret mixed strategies? Note that players only play once; so mixed strategies reflect uncertainty about what the other player might play.

Payoffs Suppose for each player i, p i is a mixed strategy for player i; i.e., it is a distribution on S i. We extend Π i by taking the expectation: Π i (p 1,...,p N )= X X p 1 (s 1 ) p N (s N )Π i (s 1,...,s N ) s 1 S 1 s N S N

Mixed strategy Nash equilibrium Given a game (N, S 1,, S N, Π 1,, Π N ): Create a new game with N players, strategy spaces Δ(S 1 ),, Δ(S N ), and expected payoffs Π 1,, Π N. A mixed strategy Nash equilibrium is a Nash equilibrium of this new game.

Mixed strategy Nash equilibrium Informally: All players can randomize over available strategies. In a mixed NE, player i s mixed strategy must maximize his expected payoff, given all other player s mixed strategies.

Mixed strategy Nash equilibrium Key observations: (1) All our definitions -- dominated strategies, iterated strict dominance, rationalizability -- extend to mixed strategies. Note: any dominant strategy must be a pure strategy.

Mixed strategy Nash equilibrium (2) We can extend the definition of best response set identically: R i (p -i ) is the set of mixed strategies for player i that maximize the expected payoff Π i (p i, p -i ).

Mixed strategy Nash equilibrium (2) Suppose p i R i (p -i ), and p i (s i ) > 0. Then s i R i (p -i ). (If not, player i could improve his payoff by not placing any weight on s i at all.)

Mixed strategy Nash equilibrium (3) It follows that R i (p -i ) can be constructed as follows: (a) First find all pure strategy best responses to p -i ; call this set T i (p -i ) S i. (b) Then R i (p -i ) is the set of all probability distributions over T i, i.e.: R i (p -i ) = Δ(T i (p -i ))

Mixed strategy Nash equilibrium Moral: A mixed strategy p i is a best response to p -i if and only if every s i with p i (s i ) > 0 is a best response to p -i

Example: coordination game We ll now apply this insight to the coordination game. Player 2 L R Player 1 l r (2,1) (0,0) (0,0) (1,2)

Example: coordination game Suppose player 1 puts probability p 1 on l and probability 1 - p 1 on r. Suppose player 2 puts probability p 2 on L and probability 1 - p 2 on R. We want to find all Nash equilibria (pure and mixed).

Example: coordination game Step 1: Find best response mapping of player 1. Given p 2 : Π 1 (l, p 2 ) = 2 p 2 Π 1 (r, p 2 ) = 1 - p 2

Example: coordination game Step 1: Find best response mapping of player 1. Then best If p 2 is: response is: < 1/3 r (p 1 = 0) > 1/3 l (p 1 = 1) = 1/3 anything (0 p 1 1)

Example: coordination game Best response of player 1: 1 R 1 (p 2 ) p 2 0 0 p 1 1

Example: coordination game Step 2: Find best response mapping of player 2. Then best If p 1 is: response is: < 2/3 R (p 2 = 0) > 2/3 L (p 2 = 1) = 2/3 anything (0 p 1 1)

Example: coordination game Best response of player 2: 1 R 2 (p 1 ) R 1 (p 2 ) p 2 0 0 p 1 1

Example: coordination game Step 3: Find Nash equilibria. As before, NE occur wherever the best response mappings cross.

Example: coordination game Nash equilibria: 1 R 2 (p 1 ) R 1 (p 2 ) p 2 0 0 p 1 1

Example: coordination game Nash equilibria: There are 3 NE: p 1 = 0, p 2 = 0 (r, R) p 1 = 1, p 2 = 1 (l, L) p 1 = 2/3, p 2 = 1/3 Note: In last NE, both players get expected payoff: 2/3 x 1/3 x 2 + 1/3 x 2/3 x 1 = 2/3.

The existence theorem Theorem: Any N-player game where all strategy spaces are finite has at least one Nash equilibrium. Notes: -The equilibrium may be mixed. -There is a generalization if strategy spaces are not finite.

The existence theorem: proof Let X = Δ(S 1 ) L Δ(S N ) be the product of all mixed strategy spaces. Define BR : X X by: BR i (p 1,, p N ) = R i (p -i )

The existence theorem: proof Key observations: -Δ(S i ) is a closed and bounded subset of R S i -Thus X is a closed and bounded subset of Euclidean space -Also, X is convex: If p, p are in X, then so is any point on the line segment between them.

The existence theorem: proof Key observations (continued): -BR is continuous (i.e., best responses don t change suddenly as we move through X) (Formal statement: BR has a closed graph, with convex and nonempty images)

The existence theorem: proof By Kakutani s fixed point theorem, there exists (p 1,, p N ) such that: (p 1,, p N ) BR(p 1,, p N ) From definition of BR, this implies: p i R i (p -i ) for all i Thus (p 1,, p N ) is a NE.

The existence theorem Notice that the existence theorem is not constructive: It tells you nothing about how players reach a Nash equilibrium, or an easy process to find one. Finding Nash equilibria in general can be computationally difficult.

Discussion of Nash equilibrium Nash equilibrium works best when it is unique: In this case, it is the only stable prediction of how rational players would play, assuming common knowledge of rationality and the structure of the game.

Discussion of Nash equilibrium How do we make predictions about play when there are multiple Nash equilibria?

1) Unilateral stability Any Nash equilibrium is unilaterally stable: If a regulator told players to play a given Nash equilibrium, they have no reason to deviate.

2) Focal equilibria In some settings, players may have prior preferences that focus attention on one equilibrium. Schelling s example (see MWG text): Coordination game to decide where to meet in New York City.

3) Focusing by prior agreement If players agree ahead of time on a given equilibrium, they have no reason to deviate in practice. This is a common justification, but can break down easily in practice: when a game is played only once, true enforcement is not possible.

4) Long run learning Another common defense is that if players play the game many (independent) times, they will naturally converge to some Nash equilibrium as a stable convention. Again, this is dangerous reasoning: it ignores a rationality model for dynamic play.

Problems with NE Nash equilibrium makes very strong assumptions: -complete information -rationality -common knowledge of rationality - focusing (if multiple NE exist)

Example Find all NE (pure and mixed) of the following game: Player 1 A B C D a (1,2) (0,1) (1,2) (0.5,1) b (4,0) (2,2) (0,3) (0,0) Player 2 c (0,3) (1,2) (3,0) (0,0) d (1,1) (0,3) (0,1) (2,0)