Evolution & Learning in Games

Similar documents
Evolutionary Game Theory

Computational Evolutionary Game Theory and why I m never using PowerPoint for another presentation involving maths ever again

Belief-based Learning

Evolution & Learning in Games

An Introduction to Evolutionary Game Theory

Other Equilibrium Notions

Costly Signals and Cooperation

Game Theory, Evolutionary Dynamics, and Multi-Agent Learning. Prof. Nicola Gatti

Game Theory -- Lecture 4. Patrick Loiseau EURECOM Fall 2016

April 29, 2010 CHAPTER 13: EVOLUTIONARY EQUILIBRIUM

Evolutionary Game Theory Notes

BELIEFS & EVOLUTIONARY GAME THEORY

Prisoner s Dilemma. Veronica Ciocanel. February 25, 2013

A (Brief) Introduction to Game Theory

Observations on Cooperation

Static Stability in Games Part I: Symmetric and Population Games

Problems on Evolutionary dynamics

SONDERFORSCHUNGSBEREICH 504

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

EVOLUTIONARY STABILITY FOR TWO-STAGE HAWK-DOVE GAMES

Summer School on Graphs in Computer Graphics, Image and Signal Analysis Bornholm, Denmark, August Outline

Outline for today. Stat155 Game Theory Lecture 16: Evolutionary game theory. Evolutionarily stable strategies. Nash equilibrium.

C31: Game Theory, Lecture 1

Game Theory and Evolution

UC Berkeley Haas School of Business Game Theory (EMBA 296 & EWMBA 211) Summer 2016

Variants of Nash Equilibrium. CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte

Evolutionary Game Theory

EVOLUTIONARY GAMES WITH GROUP SELECTION

Emergence of Cooperation and Evolutionary Stability in Finite Populations

Computation of Efficient Nash Equilibria for experimental economic games

Lectures Road Map

Lecture Notes on Game Theory

Introduction to game theory LECTURE 1

Imitation Processes with Small Mutations

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

First Prev Next Last Go Back Full Screen Close Quit. Game Theory. Giorgio Fagiolo

arxiv:math/ v1 [math.oc] 29 Jun 2004

Graph topology and the evolution of cooperation

Population Games and Evolutionary Dynamics

6.207/14.15: Networks Lecture 10: Introduction to Game Theory 2

Meaning, Evolution and the Structure of Society

6.207/14.15: Networks Lecture 11: Introduction to Game Theory 3

An Introduction to Evolutionary Game Theory: Lecture 2

Complexity in social dynamics : from the. micro to the macro. Lecture 4. Franco Bagnoli. Lecture 4. Namur 7-18/4/2008

Gains in evolutionary dynamics. Dai ZUSAI

Evolutionary Game Theory and Frequency Dependent Selection

Stable Matching Existence, Computation, Convergence Correlated Preferences. Stable Matching. Algorithmic Game Theory.

The coevolution of recognition and social behavior

EVOLUTIONARY GAME THEORY AND THE MODELING OF ECONOMIC BEHAVIOR**

Evolutionary game theory different from economic game theory

EVOLUTIONARY GAMES AND LOCAL DYNAMICS

Game Theory and Control

Bounded Rationality, Strategy Simplification, and Equilibrium

Complex networks and evolutionary games

Lecture December 2009 Fall 2009 Scribe: R. Ring In this lecture we will talk about

Weak Dominance and Never Best Responses

Acta Mathematica Academiae Paedagogicae Nyíregyháziensis 21 (2005), ISSN

The Stability of Strategic Plasticity

Irrational behavior in the Brown von Neumann Nash dynamics

Game Theory and its Applications to Networks - Part I: Strict Competition

Evolutionary Games and Periodic Fitness

Evolution of Cooperation in the Snowdrift Game with Incomplete Information and Heterogeneous Population

1. Introduction. 2. A Simple Model

Economics 204 Fall 2011 Problem Set 2 Suggested Solutions

arxiv: v2 [q-bio.pe] 18 Dec 2007

Game Theory, Population Dynamics, Social Aggregation. Daniele Vilone (CSDC - Firenze) Namur

Mohammad Hossein Manshaei 1394

Spatial Economics and Potential Games

Selfishness vs Altruism vs Balance

Some Analytical Properties of the Model for Stochastic Evolutionary Games in Finite Populations with Non-uniform Interaction Rate

Game theory Lecture 19. Dynamic games. Game theory

6.891 Games, Decision, and Computation February 5, Lecture 2

Matrix Theory, Math6304 Lecture Notes from March 22, 2016 taken by Kazem Safari

VII. Cooperation & Competition

On Hotelling s Stability in Competition

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Computing Minmax; Dominance

Conjectural Variations in Aggregative Games: An Evolutionary Perspective

Lindsell, E. A. B., & Wiesner, K. (2014). Entanglement gives rise to Pareto optimality in the generalised quantum Hawk-Dove Game. arxiv.

A Note on Discrete Convexity and Local Optimality

Basic Game Theory. Kate Larson. January 7, University of Waterloo. Kate Larson. What is Game Theory? Normal Form Games. Computing Equilibria

Evolutionary Stochastic Games

6.254 : Game Theory with Engineering Applications Lecture 13: Extensive Form Games

Game interactions and dynamics on networked populations

Understanding and Solving Societal Problems with Modeling and Simulation

MS&E 246: Lecture 4 Mixed strategies. Ramesh Johari January 18, 2007

FORWARD INDUCTION AND SUNK COSTS GIVE AVERAGE COST PRICING. Jean-Pierre Ponssard. Abstract

Communities and Populations

Game Theory. Wolfgang Frimmel. Perfect Bayesian Equilibrium

Extensive Form Games with Perfect Information

Kalle Parvinen. Department of Mathematics FIN University of Turku, Finland

Survival of Dominated Strategies under Evolutionary Dynamics. Josef Hofbauer University of Vienna. William H. Sandholm University of Wisconsin

Coevolutionary Modeling in Networks 1/39

Voluntarily Separable Repeated Prisoner s Dilemma

Online Appendices for Large Matching Markets: Risk, Unraveling, and Conflation

Introduction to Game Theory Lecture Note 2: Strategic-Form Games and Nash Equilibrium (2)

CS 573: Algorithmic Game Theory Lecture date: January 23rd, 2008

Chapter 7. Evolutionary Game Theory

The Evolution of Cooperation under Cheap Pseudonyms

arxiv: v1 [cs.sy] 13 Sep 2017

Transcription:

1 / 27 Evolution & Learning in Games Econ 243B Jean-Paul Carvalho Lecture 2. Foundations of Evolution & Learning in Games II

2 / 27 Outline In this lecture, we shall: Take a first look at local stability. In particular, we shall define an evolutionary stable state and explore its relationship to Nash equilibrium. Apply what we have learned so far to analyze iterated play of the prisoners dilemma.

3 / 27 Evolutionary Stable States (ESS) Maynard Smith and Price (1973) defined the notion of an evolutionary stable strategy: Their focus was on monomorphic populations: every member plays the same strategy, which can be a mixed strategy. We are concerned with a polymorphic population of agents each programmed with a pure strategy.

Evolutionary Stable States (ESS) Mathematically these problems are identical. For example: F 1 (x) = x 1 u(1, 1) + x 2 u(1, 2) = u(1, x). Being randomly matched with a population of agents proportion x 1 of which are programmed with pure strategy 1 and x 2 of which are programmed with pure strategy 2 is the same as playing a single individual who plays strategy 1 with probability x 1 and strategy 2 with probability x 2. 4 / 27

5 / 27 Evolutionary Stable States (ESS) Hence we can adapt the concept of an evolutionary stable strategy to a population setting: The term we shall use is evolutionary stable state (ESS). As we shall eventually see, ESS provides a sufficient condition for local stability under a wide range of evolutionary dynamics, including the replicator dynamic.

6 / 27 Invasion Let the state be x = x 1 x 2... x n. Consider a game F, where F(x) = F 1 (x) F 2 (x)... F n (x). Consider an invasion of mutants who make up a fraction ε of the post-entry population. The shares of each strategy in the mutant population are y 1 represented by y = y 2.... y n

7 / 27 Invasion Therefore, the post-entry population state is: x ε = (1 ε)x + εy = (1 ε)x 1 + εy 1 (1 ε)x 2 + εy 2... (1 ε)x n + εy n. The average payoff in the incumbent population in the post-entry state is x F ( (1 ε)x + εy ). The average payoff in the mutant population in the post-entry state is y F ( (1 ε)x + εy ).

8 / 27 Uniform Invasion Barrier The average payoff in the incumbent population is higher if: (y x) F ( (1 ε)x + εy ) < 0. (1) State x is said to admit a uniform invasion barrier if there exists an ε > 0 such that (1) holds for all y X {x} and ε (0, ε). That is, for all possible mutations y, as long as the mutant population is less than fraction ε of the postentry population, the incumbent population receives a higher average payoff.

9 / 27 Invasion Barriers Define the invasion barrier of x against y as: b x (y) = inf ( {ε (0, 1) : (y x) F ( (1 ε)x + εy ) 0} {1} ). (2) If x admits a uniform invasion barrier, then there exists an ε > 0 such that b x (y) ε > 0 for all y X {x}.

10 / 27 ESS DEFINITION. State x X is an evolutionary stable state (ESS) of F if there exists a neighborhood O of x such that: (y x) F(y) < 0 for all y O {x}. (3) In other words, if x is an ESS, then for any state y sufficiently close to x, a population playing x will receive a larger average payoff in state y than a population playing y (i.e. x is a better reply to y than y is to itself). Note that this considers invasions of other states y by x rather than invasions of x by other states. Hence it is not clear, at present, why this should be a stability condition.

11 / 27 ESS and Invasion Barriers Theorem 2.1. State x X is an evolutionary stable state (ESS) if and only if it admits a uniform invasion barrier. Thus if x is stable in the face of an arbitrarily large population of entrants who mutate to a nearby state, then it is stable in the face of a sufficiently small population of entrants who mutate to an arbitrary state.

12 / 27 ESS and NE What is the relationship between ESS and NE? DEFINITION. Suppose that x X is a NE. Then (y x) F(x) 0 for all y X. In addition, suppose there exists a neighborhood of x that does not contain any other NE. Then x is an isolated NE. Proposition 2.1. Every ESS is an isolated NE.

13 / 27 Proof Let x be an ESS of F, O be the nhd posited in (3) and y X {x} (not necessarily in O). Then for all ε > 0 sufficiently small, the postentry state x ε = εy + (1 ε)x is in O. This implies that: (x ε x) F(x ε ) < 0 (εy + (1 ε)x x) F(x ε ) < 0 ε(y x) F(x ε ) < 0 (y x) F(x ε ) < 0. (4)

14 / 27 Proof Taking ε 0 yields: (y x) F(x) 0. That is, x is a NE. To establish that x is isolated, note that if w O {x} were a NE then (w x) F(w) 0, contradicting the supposition that x is an ESS [by (3)]. The converse of Proposition 2.1 is not true. The mixed equilibrium of a two-strategy coordination game is a counterexample.

15 / 27 More on ESS and Nash Therefore, ESS is stronger than NE. In particular, an ESS satisfies the additional property: Suppose there exists a state y which is an alternative best reply to x, i.e. (y x) F(x) = 0. Then (y x) F(y) < 0, i.e. x is a better reply to y than y is to itself. Therefore: A strict NE is an ESS. A polymorphic population state (equivalent to a mixed NE) cannot be strict and hence must satisfy the additional property.

16 / 27 More on ESS and Nash In the case in which agents are matched uniformly at random to play a normal form game (the case we have been focussing on), then it is easy to see why the additional property is required. Suppose (y x) F(x) = 0, i.e. y is an alternative best reply to x. Then: (y x) F(εy + (1 ε)x) = ε(y x) F(y) + (1 ε) (y x) F(x) }{{} =0 = ε(y x) F(y). Therefore, (y x) F(y) must be negative for (1) to hold and hence, by Theorem 2.1, for x to be an ESS. (5)

17 / 27 Example: Hawk Dove Hawk Hawk 2 2 Dove 4 0 4 0 Dove 0 0 ESS: x = ( 2 3, 1 3 ). ESS payoff = 0.

18 / 27 Example: Hawk Dove Consider a mutation y such that y 1 > x 1 = 2 3. Check that (y x) F ( (1 ε)x + εy ) < 0 for all such y: (y 1 x 1 )[ 2((1 ε)x 1 + εy 1 ) + 4((1 ε)(1 x 1 ) + ε(1 y 1 )]. This equals: (y 1 x 1 )ε[ 2y 1 + 4(1 y 1 )] because 2 2 3 + 4 (1 2 3 ) = 0. This in turn equals: (y 1 x 1 )ε[4 6y 1 ] which is negative because y 1 > 2 3 by hypothesis. A similar argument can be applied to the case y 1 < x 1. Hence x is an ESS.

19 / 27 The Prisoners Dilemma C C 3 3 D 0 5 0 1 D 5 1 NE/ESS: x = (0, 1). Therefore, an ESS is not necessarily efficient.

20 / 27 Not Every Game has an ESS A B C A 1 0 1 2 0 B 2 1 0 1 2 C 0 2 2 0 1 2 0 1 x = ( 1 3, 1 3, 1 3 ) is the unique NE and therefore the only possible ESS.

21 / 27 Not Every Game has an ESS Note that x is a polymorphic population state (equivalent to a mixed strategy), so any basis vector (pure strategy) is an alternative best reply to x. Check that the additional property holds: (e 1 x) F(e 1 ) < 0, where e 1 = (1, 0, 0), i.e. the pure-strategy A. This is not the case: x F(e 1 ) = e 1 F(e 1) = 1.

22 / 27 The Iterated Prisoners Dilemma Two players engage in a series of PD games. The engagement ends after the current round with probability δ < 1 2. We call this the stopping probability. The expected number of rounds per engagement is: 1 + (1 δ) + (1 δ) 2 + (1 δ) 3 +... = 1 1 (1 δ) = 1 δ. Consider a population in which three strategies are present: C always cooperate, D always defect, T tit-for-tat, i.e. start by cooperating, thenceforth cooperate in period t if partner cooperated in t 1.

23 / 27 Expected Payoffs Within Each Pairing C D T 3 3 C δ 0 δ 5 1 D δ δ 4 + 1 δ T 3 δ 1 δ 1 3 δ Note: Payoff from playing T against D is 0 + (1 δ) 1 δ = 1 δ 1. Payoff from playing D against T is 5 + (1 δ) 1 δ = 4 + 1 δ.

24 / 27 Expected Payoffs Over All Pairings F C (x) = (x C + x T ) 3 δ 5 F D (x) = x C δ + x D 1 δ + x ( ) T 4 + 1 δ F T (x) = (x C + x T ) 3 δ + x D( 1 δ 1)

Replicator Dynamics d dt [ xt x C ] = x T x C ( FT (x) F C (x) ) = x T x C [ xd ( 1 δ 1)], which is positive because δ < 1 2. [ ] d xt dt x D = x T x D ( FT (x) F D (x) ) = x T x D [ xc 2 δ x D + x T ( 2 δ }{{ 4) ], } >0 which is positive for x T sufficiently large. 25 / 27

Vector Field D C T 26 / 27

All-T is not an ESS Let x = (x D, x C, x T ) = (0, 0, 1). Consider any alternative state y such that y D = 0. (y x) F(y) = (0 y C y T 1) F D (y) F C (y) F T (y) = y C F C (y) + (y T 1)F T (y) = [y C + (y T 1)] 3 δ (recall that y D = 0) = [y C + (1 y C 1)] 3 δ (because y T = 1 y C ) = 0. This violates (3). Hence all-t is not an ESS. This is a case of evolutionary drift. 27 / 27