Computational Evolutionary Game Theory and why I m never using PowerPoint for another presentation involving maths ever again

Similar documents
Evolution & Learning in Games

Game Theory and Evolution

Evolutionary Game Theory and Frequency Dependent Selection

Evolutionary Game Theory

EVOLUTIONARY GAMES WITH GROUP SELECTION

Game Theory -- Lecture 4. Patrick Loiseau EURECOM Fall 2016

Game Theory, Evolutionary Dynamics, and Multi-Agent Learning. Prof. Nicola Gatti

Population Games and Evolutionary Dynamics

Evolutionary Game Theory Notes

Problems on Evolutionary dynamics

April 29, 2010 CHAPTER 13: EVOLUTIONARY EQUILIBRIUM

Mohammad Hossein Manshaei 1394

CS364A: Algorithmic Game Theory Lecture #13: Potential Games; A Hierarchy of Equilibria

EVOLUTIONARY STABILITY FOR TWO-STAGE HAWK-DOVE GAMES

An Introduction to Evolutionary Game Theory: Lecture 2

Algorithmic Game Theory. Alexander Skopalik

UC Berkeley Haas School of Business Game Theory (EMBA 296 & EWMBA 211) Summer 2016

An Introduction to Evolutionary Game Theory

Computation of Efficient Nash Equilibria for experimental economic games

Summer School on Graphs in Computer Graphics, Image and Signal Analysis Bornholm, Denmark, August Outline

Meaning, Evolution and the Structure of Society

First Prev Next Last Go Back Full Screen Close Quit. Game Theory. Giorgio Fagiolo

Chapter 7. Evolutionary Game Theory

Variants of Nash Equilibrium. CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte

BELIEFS & EVOLUTIONARY GAME THEORY

Natural selection, Game Theory and Genetic Diversity

General-sum games. I.e., pretend that the opponent is only trying to hurt you. If Column was trying to hurt Row, Column would play Left, so

Introduction to Game Theory

Basic Game Theory. Kate Larson. January 7, University of Waterloo. Kate Larson. What is Game Theory? Normal Form Games. Computing Equilibria

Gains in evolutionary dynamics. Dai ZUSAI

Game interactions and dynamics on networked populations

Exercise 11. Solution. EVOLUTION AND THE THEORY OF GAMES (Spring 2009) EXERCISES Show that every 2 2 matrix game with payoff matrix

Complexity in social dynamics : from the. micro to the macro. Lecture 4. Franco Bagnoli. Lecture 4. Namur 7-18/4/2008

Other Equilibrium Notions

Graph topology and the evolution of cooperation

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

Evolutionary Game Theory

Outline for today. Stat155 Game Theory Lecture 16: Evolutionary game theory. Evolutionarily stable strategies. Nash equilibrium.

A (Brief) Introduction to Game Theory

a i,1 a i,j a i,m be vectors with positive entries. The linear programming problem associated to A, b and c is to find all vectors sa b

Area I: Contract Theory Question (Econ 206)

Complex networks and evolutionary games

Belief-based Learning

Industrial Organization Lecture 3: Game Theory

Application of Evolutionary Game theory to Social Networks: A Survey

Lecture December 2009 Fall 2009 Scribe: R. Ring In this lecture we will talk about

SF2972 Game Theory Exam with Solutions March 15, 2013

Other-Regarding Preferences: Theory and Evidence

SONDERFORSCHUNGSBEREICH 504

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

Prisoner s Dilemma. Veronica Ciocanel. February 25, 2013

RANDOM SIMULATIONS OF BRAESS S PARADOX

Introduction to Game Theory

Chapter 7. Evolutionary Game Theory

Satisfaction Equilibrium: Achieving Cooperation in Incomplete Information Games

The price of anarchy of finite congestion games

Game Theory for Linguists

AGlimpseofAGT: Selfish Routing

Evolutionary stability of games with costly signaling

6.891 Games, Decision, and Computation February 5, Lecture 2

Costly Signals and Cooperation

New Perspectives and Challenges in Routing Games: Query models & Signaling. Chaitanya Swamy University of Waterloo

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms

1 Extensive Form Games

EVOLUTIONARILY STABLE STRATEGIES AND GROUP VERSUS INDIVIDUAL SELECTION

Exact and Approximate Equilibria for Optimal Group Network Formation

Game Theory: Spring 2017

The Evolutionary Hawk-Dove Model on Graphs

Game Theory: introduction and applications to computer networks

Iterated Strict Dominance in Pure Strategies

Normal-form games. Vincent Conitzer

MS&E 246: Lecture 18 Network routing. Ramesh Johari

Traffic Games Econ / CS166b Feb 28, 2012

Distributed Learning based on Entropy-Driven Game Dynamics

Computational Game Theory Spring Semester, 2005/6. Lecturer: Yishay Mansour Scribe: Ilan Cohen, Natan Rubin, Ophir Bleiberg*

Static and dynamic stability conditions for structurally stable signaling games

Bayesian Learning in Social Networks

Evolutionary Dynamics and Extensive Form Games by Ross Cressman. Reviewed by William H. Sandholm *

Chapter 9. Mixed Extensions. 9.1 Mixed strategies

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Assignment 2 : Probabilistic Methods

CS364A: Algorithmic Game Theory Lecture #16: Best-Response Dynamics

Lindsell, E. A. B., & Wiesner, K. (2014). Entanglement gives rise to Pareto optimality in the generalised quantum Hawk-Dove Game. arxiv.

Population Games and Evolutionary Dynamics

The Role of Population Games and Evolutionary Dynamics in Distributed Control Systems

Game Theory and Control

Evolutionary game theory different from economic game theory

Computing Solution Concepts of Normal-Form Games. Song Chong EE, KAIST

A finite population ESS and a long run equilibrium in an n players coordination game

The Multi-Commodity Source Location Problems and the Price of Greed

Perforce there perished many a stock, unable: By propagation to forge a progeny.

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

C31: Game Theory, Lecture 1

Game Theory: Lecture 3

Potential Games. Krzysztof R. Apt. CWI, Amsterdam, the Netherlands, University of Amsterdam. Potential Games p. 1/3

Microeconomics. 2. Game Theory

Models of Strategic Reasoning Lecture 2

Single parameter FPT-algorithms for non-trivial games

6.254 : Game Theory with Engineering Applications Lecture 8: Supermodular and Potential Games

Game Theory. Summary: History:

Transcription:

Computational Evolutionary Game Theory and why I m never using PowerPoint for another presentation involving maths ever again Enoch Lau 5 September 2007

Outline What is evolutionary game theory? Why evolutionary game theory? Evolutionary game theory concepts Computational complexity of evolutionary stable strategies Evolutionary game theory and selfish routing Evolutionary game theory over graphs Selection strategies Finite populations 2

What is evolutionary game theory? Not creationism game theory 3

Evolutionary game theory (EGT) An infinite number of agents in 2-player symmetric games Payoffs calculate a fitness used for replication or imitation Similarities with conventional game theory Both concerned with the decisions made by agents in a game Equilibria are important concepts for both Differences from conventional game theory Rationality of agents not assumed Strategies selected by some force (evolution, cultural factors) Higher fitness means more (asexual) reproduction Other assumptions: complete mixing 4

Approaches to evolutionary game theory Two approaches 1. Evolutionary stable strategy: derives from work of Maynard Smith and Price 2. Properties of evolutionary dynamics by looking at frequencies of change in strategies 5

Evolutionary stable strategy (ESS) Incumbents and mutants in the population ESS is a strategy that cannot be invaded by a mutant population In an ESS, mutants have lower fitness (reproductive success) compared with the incumbent population ESS is more restrictive than a Nash equilibrium Not all 2-player, symmetric games have an ESS Assumptions very important: If we have a finite number of players, instead of an infinite number, different ESS 6

Evolutionary stable strategy (ESS) Finite population simulations on the Hawk-Dove game 7

History First developed by R.A. Fisher in The Genetic Theory of Natural Selection (1930) Attempted to explain the sex ratio in mammals Why is there gender balance in animals where most males don t reproduce? R.C. Lewontin explicitly applied game theory in Evolution and the Theory of Games (1961) Widespread use since The Logic of Animal Conflict (1973) by Maynard Smith and Price Seminal text: Evolution and the Theory of Games (1984) by Maynard Smith 8

Example: hawks & doves Two organisms fighting over a resource, worth V Hawks: will fight for the resource, fighting costs C Doves: will retreat from aggressive hawks, share resource with other doves Example payoff matrix: H D H -25 50 D 0 15 Nash equilibrium and ESS given by mixed strategy of (7/12, 5/12) 9

Why evolutionary game theory? Why not? 10

Equilibrium selection problem Problems with using Nash equilibria: Not all games have pure Nash equilibria Prisoner s Dilemma: sub-optimality of equilibria Multiple Nash equilibria How to choose between different Nash equilibria? Introduce refinements to the concept of Nash equilibria Then how to choose between refinements? 11

Hyper-rational agents Humans sometimes prefer A to B, B to C, and C to A EGT can predict behaviour of animals, where strong rationality assumptions fail EGT better able to handle weaker rationality assumptions? 12

Lack of dynamical theory Traditional game theory, which is static, lacks the dynamics of rational deliberation Could use extensive form (tree form) instead of normal form Quickly becomes unmanageable Presupposes hyper-rational agents Will not learn from observing opponent s behaviour 13

Philosophical problems Objections to EGT, mainly from application to human subjects Measure of fitness in cultural evolutionary interpretations Explanatory irrelevance of evolutionary game theory Does EGT simply reinforce existing values and biases? EGT does not provide sufficient evidence for the origin of phenomena Historical records more useful? 14

Evolutionary game theory concepts This is where your head is meant to start hurting 15

Classical model Infinite population of organisms Each organism assumed equally likely to interact with each other organism Fixed, 2-player, symmetric game Fitness function F A is set of actions (A) is set of probability distributions F: (A) x (A) R F(s t) = fitness of s playing t ε proportion are mutants, 1 ε are incumbents 16

Evolutionary stable strategy s is an incumbent, t is a mutant Expected fitness of an incumbent: (1 - ε) F(s s) + ε F(s t) Expected fitness of mutant: (1 - ε) F(t s) + ε F(t t) s is an ESS if there exists an ε t such that for all 0 < ε < ε t, fitness of incumbent > fitness of mutant Implies: 1. F(s s) > F(t s), or 2. F(s s) = F(t s) and F(s t) > F(t t) A strategy s is an ESS for a 2-player, symmetric game given by a fitness function F, iff (s, s) is a Nash equilibrium of F, and for every best response t to s, t s, F(s t) > F(t t) 17

Example: hawks & doves Generalised payoff matrix: H D Note that (D, D) is not a Nash equilibrium Cannot be an ESS either If V > C: H is an ESS If V C: H (V C) / 2 V D 0 V / 2 Mixed strategy: H with prob V/C, D with prob 1 V/C is ESS 18

Example: hawks & doves Map of proportions for Hawk-Dove game. Note that where the curve meets the straight line at a gradient of less than 1 (the middle point), that is a stable equilibrium. Where it meets it at a gradient greater than 1, it is an unstable equilibrium. 19

Replicator dynamics Continuous dynamics for EGT Find differential equations for the change in the proportion of each strategy over time In some cases, such as the Prisoner s Dilemma, stable states of replicator dynamics occur when everyone in the population follows the ESS Roughly, true when only two pure strategies exist Can fail to be true with more than two pure strategies 20

Example: Prisoner s Dilemma Generalised payoff matrix C D C (R, R ) (S, T ) D (T, S ) (P, P ) with T > R > P > S and T > R > P > S Fitness functions 21

Example: Prisoner s Dilemma Proportion of C and D in next generation: where W is the overall fitness of population (weighted by proportion) Leads to differential equations: Use payoff matrix to show that p d > 0 and p c < 0 22

Computational complexity of evolutionary stable strategies No good news here 23

Results and proof outline Finding an ESS is both NP-hard and conp-hard Reduction from the problem of checking if a graph has a maximum clique of size exactly k Recognising whether a given strategy is an ESS is also conp-hard Transform a graph G into a payoff matrix F, which will have an ESS iff the size of the largest clique in G is not equal to k Transform adjacency matrix: replace all diagonal entries with the value ½, inserting 0 th row and 0 th column with entries 1 1/(2k) 24

Proof idea For a mixed strategy s to be an ESS, incumbents should receive a relatively high payoff when playing other incumbents When s plays itself, it must guarantee that the pure strategies chosen will correspond to two adjacent vertices Mixed strategy with support over a clique will achieve this When max clique is greater than k, uniform mixed strategy corresponding to clique will be an ESS When max clique is less than k, get pure strategy ESS No ESS in the case where max clique is exactly k 25

Technical lemma If s is a strategy with s 0 = 0, then F(s s) 1 1/(2k ), where k is the size of the maximum clique in G. This holds with equality iff s is the uniform distribution over a k -clique. Proof idea By induction over the number of non-edges between the vertices in G Inductive step: Find two non-adjacent vertices u and v, and construct a new strategy s by moving the probability in s from v to u 26

Lemmas 1. If C is a maximal clique in G of size k > k, and s is the uniform distribution on C, then s is an ESS 2. If the maximum size clique in G is of size k < k, then the pure strategy 0 is an ESS 3. If the maximum size clique of G is at least k, then the pure strategy 0 is not an ESS 4. If the maximum size clique of G is at most k, then any strategy for F that is not equal to the pure strategy 0, is not an ESS for F 27

Proof of Lemma 1 By technical lemma, F(s s) = 1 1/(2k ) Any best response to s must have support over only C F(0 s) = 1 1/(2k) < F(s s) by construction Take a u not in C: u is connected to at most k 1 vertices in C (since max clique size is k ) F(u s) 1 1/k (sum up the entries in the payoff matrix) F(u s) < F(s s) Also by technical lemma, payoff of s is maximised when s is uniform distribution over C Hence, s is a best response to itself 28

Proof of Lemma 1 Now, need to show that for all best responses t to s, t s, F(s t) > F(t t) (note: t has support over C) By technical lemma, F(t t) < 1 1/(2k ) (note: no equality here since t s) Using F, we can show that F(s t) = 1 1/(2k ) (C is a clique, s and t are distributions with support over C) You can get this by summing up the values in the payoff matrix (k ½)/k = 1 1/(2k ) Hence, F(s t) > F(t t) 29

Proof of Lemma 2 Mutant strategy t F(t 0) = 1 1/(2k) = F(0 0) 0 is a best response to itself So need to show F(0 t) > F(t t) Form t* by setting the probability of strategy 0 in t to zero and then renormalising Applying the technical lemma: F(t* t*) 1 1/(2k ) < 1 1(2k) = F(0 t) 30

Proof of Lemma 2 Expression for F(t t): By expanding out expressions for F(t t) and F(t* t*): F(0 t) > F(t t) iff F(0 t) > F(t* t*) 31

Evolutionary game theory and selfish routing Ah, something related to my thesis topic 32

The model Each agent assumed to play an arbitrary pure strategy Imitative dynamics switch to lower latency path with probability proportional to difference in latencies Recall: at a Nash flow, all s-t paths have the same latency If we restrict the latency functions to be strictly increasing, then Nash flows are essentially ESS Paths with below average latency will have more agents switching to them than from them Paths with above average latency will have more agents switching from them than to them 33

Convergence to Nash flow As t, any initial flow with support over all paths in P will eventually converge to a Nash flow Use Lyapunov s direct method to show that imitative dynamics converge to a Nash flow General framework for proving that a system of differential equations converges to a stable point Define a potential function that is defined in the neighbourhood of the stable point and vanishes at the stable point itself Then show that the potential function decreases with time System will not get stuck in any local minima 34

Convergence to approximate equilibria ε-approximate equilibrium: Let P ε be the paths that have latency at least (1 + ε)l*, and let x ε be the fraction of agents using these paths. A population is at ε-approximate equilibrium iff x ε < ε Only a small fraction of agents experience latency significantly worse than the average latency Potential function Measures the total latency the agents experience Integral: sums latency if agents were inserted one at a time 35

Convergence to approximate equilibria Theorem: the replicator dynamics converge to an ε- approximate equilibrium time O(ε -3 ln(l max /l * )) Proof: see handout 36

Evolutionary game theory over graphs Did you know? I am my neighbour s neighbour. 37

The model No longer assume that two organisms are chosen uniformly at random to interact Organisms only interact with those in their local neighbourhood, as defined by an undirected graph or network Use: Depending on the topology, not every mutant is affected equally Groups of mutants with lots of internal attraction may be able to survive Fitness given by the average of playing all neighbours 38

Mutant sets to contract We consider an infinite family G = {G n } (where G n is a graph with n vertices) Examine asymptotic (large n) properties When will mutant vertex sets contract? Let M n be the mutant subset of vertices M n εn for some constant ε > 0 M n contracts if, for sufficiently large n, for all but o(n) of the j in M n, j has an incumbent neighbour i such that F(j) < F(i) ε-linear mutant population: smaller than invasion threshold ε n but remain some constant fraction of the population (isn t a vanishing population) 39

Results A strategy s is ESS if given a mutant strategy t, the set of mutant strategies M n all playing t, for n sufficiently large, M n contracts Random graphs: pairs of vertices jointed by probability p If s is classical ESS of game F, if p = Ω(1/n c ), 0 c < 1, s is an ESS with probability 1 with respect to F and G Adversarial mutations: At an ESS, at most o(n) mutants can be of abnormal fitness (i.e. outside of a additive factor τ) 40

Selection methods The art of diplomacy 41

Role of selection Dynamics of EGT not solely determined by payoff matrix Let the column vector p represent strategy proportions F(p) is a fitness function S(f, p) is the selection function Returns the state of the population for the next generation, given fitness values and current proportions p t + 1 = S(F(p t ), p t ) Different selection strategies result in different dynamics Any S that maintains stable fixed points must obey p fix = S(c 1, p fix ), and show convergence around p fix 42

Selection methods Some selection methods commonly used in evolutionary algorithms: Truncation (μ, ƛ)-ES Linear rank Boltzmann selection 43

Example: Truncation selection Population size n, selection pressure k Sort population according to fitness Replace worst k percent of the population with variations of the best k percent 44

Example: Linear rank selection Often used in genetic algorithms Agents sorted according to fitness, assigned new fitness values according to rank Create roulette wheel based on new fitness values, create next generation Useful for ensuring that even small differences in fitness levels are captured 45

References Just to prove I didn t make the whole talk up. 46

References (not in any proper format!) Suri S. Computational Evolutionary Game Theory, Chapter 29 of Algorithmic Game Theory, edited by Nisan N, Roughgarden T, Tardos E, and Vazirani V. Ficici S, and Pollack J. Effects of Finite Populations on Evolutionary Stable Strategies Ficici S, Melnik O, and Pollack J. A Game-Theoretic Investigation of Selection Methods Used in Evolutionary Algorithms 47