Survival of Dominated Strategies under Evolutionary Dynamics. Josef Hofbauer University of Vienna. William H. Sandholm University of Wisconsin

Similar documents
Population Games and Evolutionary Dynamics

Evolutionary Game Theory: Overview and Recent Results

Survival of Dominated Strategies under Evolutionary Dynamics

Survival of dominated strategies under evolutionary dynamics

Deterministic Evolutionary Dynamics

Irrational behavior in the Brown von Neumann Nash dynamics

The projection dynamic and the replicator dynamic

Evolution & Learning in Games

Irrational behavior in the Brown von Neumann Nash dynamics

Pairwise Comparison Dynamics and Evolutionary Foundations for Nash Equilibrium

Stochastic Evolutionary Game Dynamics: Foundations, Deterministic Approximation, and Equilibrium Selection

Pairwise Comparison Dynamics for Games with Continuous Strategy Space

Stable Games and their Dynamics

Local Stability of Strict Equilibria under Evolutionary Game Dynamics

Distributional stability and equilibrium selection Evolutionary dynamics under payoff heterogeneity (II) Dai Zusai

Population Games and Evolutionary Dynamics

Gains in evolutionary dynamics: unifying rational framework for dynamic stability

Gains in evolutionary dynamics. Dai ZUSAI

Evolutionary Dynamics and Extensive Form Games by Ross Cressman. Reviewed by William H. Sandholm *

DETERMINISTIC AND STOCHASTIC SELECTION DYNAMICS

Population Dynamics Approach for Resource Allocation Problems. Ashkan Pashaie

DISCRETIZED BEST-RESPONSE DYNAMICS FOR THE ROCK-PAPER-SCISSORS GAME. Peter Bednarik. Josef Hofbauer. (Communicated by William H.

Deterministic Evolutionary Game Dynamics

Deterministic Evolutionary Game Dynamics

Stochastic Imitative Game Dynamics with Committed Agents

Distributed Learning based on Entropy-Driven Game Dynamics

Shapley Polygons in 4 4 Games

Spatial Economics and Potential Games

Sample Path Large Deviations for Stochastic Evolutionary Game Dynamics

Evolutionary Internalization of Congestion Externalities When Values of Time are Unknown

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

LEARNING IN CONCAVE GAMES

Nonaggregable evolutionary dynamics under payoff heterogeneity

Large Deviations and Stochastic Stability in the Small Noise Double Limit

epub WU Institutional Repository

Large deviations and stochastic stability in the small noise double limit

6.891 Games, Decision, and Computation February 5, Lecture 2

Near-Potential Games: Geometry and Dynamics

Sampling best response dynamics and deterministic equilibrium selection

Large Deviations and Stochastic Stability in the Small Noise Double Limit, I: Theory

arxiv: v1 [cs.sy] 13 Sep 2017

Learning in Games and the Topology of Dynamical Systems. Georgios Piliouras

Reinforcement Learning in Evolutionary Games

Evolutionary Games and Periodic Fitness

Equilibria in Games with Weak Payoff Externalities

Logit Dynamic for Continuous Strategy Games: Existence of Solutions.

Brown s Original Fictitious Play

An Introduction to Dynamo: Diagrams for Evolutionary Game Dynamics

Decompositions of normal form games: potential, zero-sum, and stable games

An Introduction to Evolutionary Game Theory: Lecture 2

Near-Potential Games: Geometry and Dynamics

MS&E 246: Lecture 17 Network routing. Ramesh Johari

Unified Convergence Proofs of Continuous-time Fictitious Play

WHEN ORDER MATTERS FOR ITERATED STRICT DOMINANCE *

The Dynamics of Generalized Reinforcement Learning

Sophisticated imitation in cyclic games

Static and dynamic stability conditions for structurally stable signaling games

Evolution & Learning in Games

Metric Spaces and Topology

p-dominance and Equilibrium Selection under Perfect Foresight Dynamics

Decompositions and Potentials for Normal Form Games

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

Strategic Properties of Heterogeneous Serial Cost Sharing

Computing Solution Concepts of Normal-Form Games. Song Chong EE, KAIST

April 29, 2010 CHAPTER 13: EVOLUTIONARY EQUILIBRIUM

EVOLUTIONARY GAMES WITH GROUP SELECTION

Evolutionary Stochastic Games

Imitation Processes with Small Mutations

Ex Post Cheap Talk : Value of Information and Value of Signals

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Influencing Social Evolutionary Dynamics

The Role of Population Games and Evolutionary Dynamics in Distributed Control Systems

Belief-based Learning

The Dynamic Instability of Dispersed Price Equilibria.

Robust Mechanism Design and Robust Implementation

Game Theory, Evolutionary Dynamics, and Multi-Agent Learning. Prof. Nicola Gatti

Evolutionary Game Theory

Best Experienced Payoff Dynamics and Cooperation in the Centipede Game

Smooth Calibration, Leaky Forecasts, Finite Recall, and Nash Dynamics

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

Stability and Long Run Equilibrium in Stochastic Fictitious Play

An Introduction to ABED: Agent-Based Simulation of Evolutionary Game Dynamics

Optimality, Duality, Complementarity for Constrained Optimization

Refined best-response correspondence and dynamics

Impulsive Stabilization and Application to a Population Growth Model*

An Introduction to Evolutionary Game Theory

epub WU Institutional Repository

Pure Strategy Equilibria in Symmetric Two-Player Zero-Sum Games

Entropic Selection of Nash Equilibrium

Local-Effect Games. Kevin Leyton-Brown. Moshe Tennenholtz. Stanford UBC. Technion

Long Run Equilibrium in Discounted Stochastic Fictitious Play

Mixed Nash Equilibria

Lecture Notes on Bargaining

Computational Evolutionary Game Theory and why I m never using PowerPoint for another presentation involving maths ever again

A Necessary and Sufficient Condition for a Unique Maximum with an Application to Potential Games

1 Basic Game Modelling

The Distribution of Optimal Strategies in Symmetric Zero-sum Games

1 Lyapunov theory of stability

Static Stability in Games Part I: Symmetric and Population Games

LECTURE 8: DYNAMICAL SYSTEMS 7

Transcription:

Survival of Dominated Strategies under Evolutionary Dynamics Josef Hofbauer University of Vienna William H. Sandholm University of Wisconsin a hypnodisk 1

Survival of Dominated Strategies under Evolutionary Dynamics One role of evolutionary game theory is to provide foundations for traditional, rationality-based solution concepts. example: rest points (of the dynamic) and Nash equilibria (of the underlying game) Typically, NE RP. For some general classes of dynamics, NE = RP. (Sandholm (2005, 2006)) Is it possible to guarantee that all dynamically stable sets consist of Nash equilibria? No. Hofbauer and Swinkels (1996) and Hart and Mas-Colell (2003) construct games in which most solutions of any reasonable evolutionary dynamic fail to converge to Nash equilibrium. 2

We therefore consider a more modest goal: elimination of strictly dominated strategies. Dominance is the mildest requirement employed in standard game-theoretic analysis, so it is natural to expect evolutionary dynamics to accord with it. And indeed: 1. Dominated strategies are eliminated by certain dynamics based on imitation, so long as play begins at an interior initial condition. replicator dynamic (Akin (1980)) families of imitative dynamics (Samuelson and Zhang (1992), Hofbauer and Weibull (1996)) 2. Clearly, dominated strategies are eliminated under the best response dynamic. 3

Still, there is no a priori reason to expect dominated strategies to be eliminated by evolutionary dynamics. Evolutionary dynamics are based on the idea that agents switch to strategies whose current payoffs are reasonably good. Even if a strategy is dominated, it can have reasonably good payoffs in many states. Put differently: decision making in evolutionary models is local ; domination is global. We show that any evolutionary dynamic satisfying three natural conditions continuity, positive correlation, and innovation must sometimes fail to eliminate strictly dominated strategies. Thus, under dynamics other than those noted above, we need not expect strictly dominated strategies to disappear. Moral: Evolutionary dynamics provide surprisingly little support for a basic rationality postulate. 4

Population Games We consider games played by a single unit-mass population of agents. S = {1,, n} strategies X = {x R + n : F i : X R = 1} population states/mixed strategies x i i payoffs to strategy i (Lipschitz continuous) F: X R n payoffs to all strategies = the payoff vector field 5

example: random matching in symmetric two-player normal form games A R nn A ij = e i Ae j payoff matrix ( bimatrix (A, A ) ) payoff for playing i S against j S F i (x) = e i Ax (expected) payoff for playing i against x X F(x) = Ax the payoff vector field note: {Nash equilibria of F} = {symmetric Nash equilibria of (A, A )} 6

example: congestion games (Beckmann, McGuire, and Winsten (1956)) Home and Work are connected by paths i S consisting of links. The payoff to choosing path i is (the delay on path i) = (the sum of the delays on the links on path i) formally: F i (x) = c (u (x)) payoff to path i x i u (x) = c (u ) i i: i x i mass of players choosing path i utilization of link cost of delay on link Examples: highway congestion c increasing positive externalities c decreasing 7

Microfoundations for evolutionary dynamics The choice procedure individual agents follow is called a revision protocol. : R n X R + nn ij (F(x), x) revision protocol conditional switch rate from strategy i to strategy j An all-purpose interpretation: Each agent is equipped with a rate Poisson alarm clock ( = n max i,j,x ij (F(x), x)). When an i player s clock rings, he receives a revision opportunity, during which (i) he selects a strategy j S at random; (ii) he switches to strategy j with probability proportional to ij (F(x), x). Often, simpler interpretations work for specific revision protocols. 8

The mean dynamic population game F revision protocol population size N Markov process {X t N } Over finite time spans, the process {X t N } is well-approximated by solutions to an ODE. (Benaim and Weibull (2003), Sandholm (2003)) This ODE, the mean dynamic, is defined by the expected increments of {X t N }. (MD) x i =V F i (x) = x j ji (F(x), x) x i ij (F(x), x) js js = inflow into i outflow from i 9

Examples Revision protocol Mean dynamic ij = x j [F j F i ] + x i = x i (F i (x) F(x)) the replicator dynamic (Taylor and Jonker (1978)) ij = B j F (x) x B F (x) x the best response dynamic (Gilboa and Matsui (1991)) ij = [F j F ] + x i = [F i (x) F(x)] + x i [F j (x) F(x)] js + the Brown-von Neumann-Nash (BNN) dynamic (B-vN (1950)) ij = [F j F i ] + x i = x j [F i (x) F j (x)] js + x i [F j (x) F i (x)] js + the pairwise difference (PD) dynamic (Smith (1984)) 10

Survival of dominated strategies I: Families of simple revision protocols Berger and Hofbauer (2006) show that strictly dominated strategies can survive under the Brown-von Neumann-Nash dynamic. We show that their construction also works for the pairwise difference dynamic families of dynamics that generalize the BNN and PD dynamics Step 1: Cycling in bad RPS Suppose agents are randomly matched to play bad Rock-Paper-Scissors. A = 0 2 1 1 0 2 2 1 0 Unless the initial state is the unique NE x* = ( 1 3, 1 3, 1 3 ), play approaches a limit cycle. Proof: Use a Lyapunov function. (The function depends on the dynamic at issue). 11

the BNN dynamic in bad RPS the PD dynamic in bad RPS colors represent speeds (red = fastest, blue = slowest) 12

Step 2: Add a twin to bad RPS Add a fourth strategy to bad RPS that duplicates Scissors. A = 0 2 1 1 1 0 2 2 2 1 0 0 2 1 0 0 Except from the line of NE {x X: x = (( 1 3, 1 3, 1 6 c, 1 6 + c)}, play converges to a cycle on the plane where x 3 = x 4. (Why? (PD) x i = x j [F i (x) F j (x)] + x i [F j (x) F i (x)] + js js x i x j = (x j x i ) [F k (x) F i (x)] +.) ks 13

the PD dynamic in bad RPS with a twin 14

Step 3: A feeble twin Uniformly reduce the payoff of the twin by : A = 0 2 1 1 1 0 2 2 2 1 0 0 2 1 Since the dynamic depends continuously on the game s payoffs, the attracting interior limit cycle persists when is slightly greater than zero. The feeble twin, a strictly dominated strategy, survives. 15

the PD dynamic in bad RPS + twin the PD dynamic in bad RPS + feeble twin 16

Survival of dominated strategies II: The general result Three conditions for evolutionary dynamics: (C) Continuity: is Lipschitz continuous. Small changes in aggregate behavior do not lead to large changes in agents responses. Rules out protocols that are extremely sensitive to the exact values of F(x) or x (for example, the best response protocol). (PC) Positive correlation: V F (x) F(x) 0, with equality if and only if x NE(F). A weak payoff monotonicity condition for disequilibrium states (see Friedman (1991), Swinkels (1993), Sandholm (2001)). Implies that RP(V F ) = NE(F), ruling out purely imitative dynamics. (IN) Innovation: If x NE(F), x i = 0, and i BR F (x), then V i F (x) > 0. If a disequilibrium state has an unused best response, some agents switch to it. Also rules out purely imitative dynamics. 17

Theorem: If the dynamic V satisfies (C), (PC), and (IN), there are games F in which a strictly dominated strategy survives from most initial conditions under V F. Analysis The bad RPS construction only works under specific assumptions about the functional form of the revision protocol. Since we now impose little structure on, we require a more elaborate construction. 18

Step 0: Potential games (Monderer and Shapley (1996), Sandholm (2001)) F is a potential game if there exists an f: X R s.t. f (x) = F(x) for all x X externality symmetry: DF(x) is symmetric for all x X. examples: common interest games, congestion games, games generated by variable externality pricing schemes In potential games, evolutionary dynamics are very well-behaved. Lemma: If F is a potential game, and the dynamic V F satisfies (PC), then nonstationary solutions of V F ascend the potential function f. (That is, f is a strict Lyapunov function for V F.) Proof: d f (x ) = f (x dt t t ) x t = F(x t ) V F (x t ) 0. 19

example: coordination C = 1 0 0 0 1 0 0 0 1 f(x) = 1 2 (x 1 2 + x 2 2 + x 3 2 ) F(x) = Cx (projected onto TX) 20

example: anti-coordination C = 1 0 0 0 1 0 0 0 1 f(x) = 1 (x 2 2 1 + x 2 2 + x 2 3 ) F(x) = Cx 21

Step 1: Construction of the hypnodisk game Begin with the coordination game F(x) = Cx above. Start twisting the portion of the vector field outside the inner circle in a clockwise direction. Exclude larger and larger inner circles as you twist, so that the outer circle is reached when the total twist is 180. 22

The twist the coordination game the hypnodisk game 23

the hypnodisk game 24

Analytically: x 1 1 3 F(x) = cos((x)) x 2 1 3 + 3 x 2 x 3 3 sin((x)) x 3 x 1, x 3 1 3 x 1 x 2 where (x) = 0 if x x* r; 1 if x x* R; smooth in between. By construction, F(x) = Cx inside the inner circle, and F(x) = Cx outside the outer circle. Therefore, the annulus between the circles attracts almost all solutions to V F. 25

Step 2: Add a twin Let ˆF be F augmented by a twin. In an inner cylinder I, ˆF (x) = Ĉx = 1 0 0 0 0 1 0 0 x is a potential game. 0 0 1 1 0 0 1 1 Outside an outer cylinder O, ˆF (x) = Ĉx is a potential game. By (PC ), all solutions to V ˆF that do not start on the line of NE ( 1 3, 1 3, 1 6 c, 1 6 + c) converge to D = O I, the difference of the two cylinders. But unlike in the previous analysis, the plane where x 3 = x 4 typically is not attracting. 26

the cylinders O and I, and their difference D = O I 27

Step 3: The attractor A We describe the possible limit behaviors of the dynamic using the notion of attractor-repeller pairs (Conley (1978)). Define the flow from set U X under the dynamic V ˆF by t (U) = { X: there is a solution {x s } to x = V ˆF(x) with x 0 U and x t = }. The set NE is a repeller under V ˆF : there is a neighborhood U of NE such that NE = t (cl(u)). t<0 Let A be the dual attractor of the repeller NE: A = t (X cl(u)). t>0 Fact: A is compact, and it is (forward & backward) invariant under V ˆF. Also: By construction, A D. 28

Let Z = {x X: x 4 = 0} be the face of X on which the twin is unused. Lemma: A and Z are disjoint. Proof: Since strategies cannot become extinct in finite time, X Z is forward invariant. a backward solution trajectory from Z cannot enter int(x). Therefore, if A Z, the entire backward orbit from is in A Z, and so in D Z (since A D). Since D Z contains no rest points (by (PC)), the Poincaré-Bendixson Theorem implies that the backward orbit from converges to a closed orbit in D Z that circumnavigates I Z. But this closed orbit must pass through a region in which strategies 3 and 4 are optimal, contradicting innovation (IN). (See the picture.) 29

The best response correspondence for the hypnodisk game. 30

Step 4: The feeble twin The Lemma shows that in game ˆF, the attractor A is a compact subset of int(x), and hence is bounded away from Z. Therefore, if we make the twin feeble, a continuity argument shows that it will survive. 31

Discussion 1. The hypnodisk game ensures the existence of a cycle under any evolutionary dynamic that satisfies (PC). As we have seen, survival results can be proved using simpler games if one fixes the evolutionary dynamic in advance. 2. Some purely imitative dynamics (e.g., the replicator dynamic) eliminate dominated strategies along interior solution trajectories. The results in this paper imply that if agents usually imitate but occasionally choose strategies directly, then dominated strategies can survive. 32

Inflow-outflow symmetry What is special about imitative dynamics? A revision protocol generates a mean dynamic V F for any population game F. (MD) x i = V i F (x) = x j ji x i ij js js = inflow into i outflow from i Notice that the outflow term is proportional to x i, but that the inflow term is not. Observation: If the functional form of ij eliminates inflow-outflow asymmetry, the dynamic V F is one step closer to eliminating dominated strategies. 33

Inflow-outflow symmetry under imitative dynamics ex. the replicator dynamic ij = x j [F j F i ] + ( ij is proportional to x j ) x i = x j ji x i ij js js = x j x i [F i F j ] js + x i x j [F j F i ] js + = x i x j (F i F j ) js = x i ˆFi (x) (x i is proportional to x i ) d dt x i x j = x i x j ( F i F j ) If strategies i and j are twins, x i x j is constant along interior solution trajectories. 34

the replicator dynamic in RPS + twin the replicator dynamic in RPS + feeble twin 35

Flows with a continuum of invariant planes are not robust. Indeed, the main theorem tells us that if agents only imitate most of the time, dominated strategies can survive. (.9 replicator) + (.1 pairwise difference) (.9 replicator) + (.1 pairwise difference) in slightly bad RPS in slightly bad RPS + feeble twin 36

Inflow-outflow symmetry under an abandonment dynamic ex. the projection dynamic (on int(x)) (Nagurney and Zhang (1997), Sandholm, Dokumaci, and Lahkar (2006)). ij = [F j F i ] + nx i on int(x) ( ij is inversely proportional to x i ) x i = x j js ji x i js ij = x j [F i F j ] + [F i F j ] + x nx i j nx i js js = 1 n js (F i F j ) = F i 1 F n j (x i does not depend explicitly on x i ) js d dt ( x x i j )= F i F j If strategies i and j are twins, x i x j is constant along interior solution trajectories. 37

the projection dynamic in RPS + twin the projection dynamic in RPS + feeble twin (But: The projection dynamic has unusual boundary behavior: solution trajectories can enter and exit bd(x). Because of this, strictly dominated strategies can survive (Sandholm, Dokumaci, and Lahkar (2006)).) 38

In general, ij is neither proportional to x j nor inversely proportional to x i, so inflow-outflow symmetry does not hold. Without this symmetry, dominance arguments have no bite: the global character of dominance is not captured by the local criteria used by agents in evolutionary models. 39