Mean-field-game model for Botnet defense in Cyber-security

Similar documents
warwick.ac.uk/lib-publications

Large-population, dynamical, multi-agent, competitive, and cooperative phenomena occur in a wide range of designed

Mean Field Games on networks

Mean-Field-Game Model of Corruption

Variational approach to mean field games with density constraints

Explicit solutions of some Linear-Quadratic Mean Field Games

MEAN FIELD GAMES WITH MAJOR AND MINOR PLAYERS

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio

HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS. CIME Courses-Cetraro August 29-September COURSES

Cyber security based on mean field game model of the defender: Attacker strategies

arxiv: v4 [math.oc] 24 Sep 2016

Learning in Mean Field Games

LAW OF LARGE NUMBERS FOR THE SIRS EPIDEMIC

September Math Course: First Order Derivative

Noncooperative continuous-time Markov games

Multi-Robotic Systems

Maximum Process Problems in Optimal Control Theory

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games

UNIQUENESS OF SELF-SIMILAR VERY SINGULAR SOLUTION FOR NON-NEWTONIAN POLYTROPIC FILTRATION EQUATIONS WITH GRADIENT ABSORPTION

On Equilibria of Distributed Message-Passing Games

ON MEAN FIELD GAMES. Pierre-Louis LIONS. Collège de France, Paris. (joint project with Jean-Michel LASRY)

Dynamic Games with Asymmetric Information: Common Information Based Perfect Bayesian Equilibria and Sequential Decomposition

Interval-Valued Cores and Interval-Valued Dominance Cores of Cooperative Games Endowed with Interval-Valued Payoffs

Dynamic Bertrand and Cournot Competition

Numerical methods for Mean Field Games: additional material

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Lecture December 2009 Fall 2009 Scribe: R. Ring In this lecture we will talk about

arxiv: v1 [math.pr] 18 Oct 2016

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Distributed Learning based on Entropy-Driven Game Dynamics

k-protected VERTICES IN BINARY SEARCH TREES

This is a repository copy of Density flow over networks: A mean-field game theoretic approach.

Computation of Efficient Nash Equilibria for experimental economic games

THE RADIO NUMBERS OF ALL GRAPHS OF ORDER n AND DIAMETER n 2

Preliminary Results on Social Learning with Partial Observations

Global Analysis of an Epidemic Model with Nonmonotone Incidence Rate

A MODEL FOR THE LONG-TERM OPTIMAL CAPACITY LEVEL OF AN INVESTMENT PROJECT

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Mean-Field Games with non-convex Hamiltonian

FAST AND HETEROCLINIC SOLUTIONS FOR A SECOND ORDER ODE

Mean Field Control: Selected Topics and Applications

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

arxiv: v1 [math.ds] 17 Mar 2017

Cyber-Awareness and Games of Incomplete Information

Quitting games - An Example

MEAN FIELD GAMES AND SYSTEMIC RISK

Unimodality of Steady Size Distributions of Growing Cell Populations

Generalized Hypothesis Testing and Maximizing the Success Probability in Financial Markets

1 Lyapunov theory of stability

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q)

arxiv: v2 [math.ap] 28 Nov 2016

Unitary Dynamics and Quantum Circuits

Linearization of Differential Equation Models

WITH SWITCHING ARMS EXACT SOLUTION OF THE BELLMAN EQUATION FOR A / -DISCOUNTED REWARD IN A TWO-ARMED BANDIT

A NEW SOLUTION OF SIR MODEL BY USING THE DIFFERENTIAL FRACTIONAL TRANSFORMATION METHOD

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

Game theory Lecture 19. Dynamic games. Game theory

Uncertainty and Disagreement in Equilibrium Models

Risk-Sensitive and Robust Mean Field Games

Mathematics-I Prof. S.K. Ray Department of Mathematics and Statistics Indian Institute of Technology, Kanpur. Lecture 1 Real Numbers

The SIS and SIR stochastic epidemic models revisited

Stochastic Volatility and Correction to the Heat Equation

Game Theory. Solutions to Problem Set 4

Mean field equilibria of multiarmed bandit games

Notes on Measure Theory and Markov Processes

Definitions and Proofs

NONTRIVIAL SOLUTIONS TO INTEGRAL AND DIFFERENTIAL EQUATIONS

NONLINEAR FREDHOLM ALTERNATIVE FOR THE p-laplacian UNDER NONHOMOGENEOUS NEUMANN BOUNDARY CONDITION

c 2004 Society for Industrial and Applied Mathematics

On Quadratic Stochastic Operators Having Three Fixed Points

Stochastic Processes

1. Introduction. 2. A Simple Model

Viscosity Solutions for Dummies (including Economists)

For general queries, contact

Mean Field Consumption-Accumulation Optimization with HAR

HOPFIELD neural networks (HNNs) are a class of nonlinear

A-PRIORI ESTIMATES FOR STATIONARY MEAN-FIELD GAMES

Principles of Real Analysis I Fall I. The Real Number System

Spectral theory for magnetic Schrödinger operators and applicatio. (after Bauman-Calderer-Liu-Phillips, Pan, Helffer-Pan)

Could Nash equilibria exist if the payoff functions are not quasi-concave?

Time-delay feedback control in a delayed dynamical chaos system and its applications

Population Games and Evolutionary Dynamics

An Introduction to Noncooperative Games

MODELING AND ANALYSIS OF THE SPREAD OF CARRIER DEPENDENT INFECTIOUS DISEASES WITH ENVIRONMENTAL EFFECTS

1 The Observability Canonical Form

From SLLN to MFG and HFT. Xin Guo UC Berkeley. IPAM WFM 04/28/17 Based on joint works with J. Lee, Z. Ruan, and R. Xu (UC Berkeley), and L.

Nonlinear dynamics in the Cournot duopoly game with heterogeneous players

Production and Relative Consumption

MCMC 2: Lecture 3 SIR models - more topics. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

Propagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R

1 Large Deviations. Korea Lectures June 2017 Joel Spencer Tuesday Lecture III

DETERMINISTIC AND STOCHASTIC SELECTION DYNAMICS

An infinite dimensional pursuit evasion differential game with finite number of players

Gärtner-Ellis Theorem and applications.

Estimating the Region of Attraction of Ordinary Differential Equations by Quantified Constraint Solving

Discrete quantum random walks

THE NEARLY ADDITIVE MAPS

Impulsive Stabilization and Application to a Population Growth Model*

A Rothschild-Stiglitz approach to Bayesian persuasion

Transcription:

1 Mean-field-game model for Botnet defense in Cyber-security arxiv:1511.06642v1 [math.oc] 20 Nov 2015 V. N. Kolokoltsov and A. Bensoussan November 23, 2015 Abstract We initiate the analysis of the response of computer owners to various offers of defence systems against a cyber-hacker (for instance, a botnet attack), as a stochastic game of a large number of interacting agents. We introduce a simple mean-field game that models their behavior. It takes into account both the random process of the propagation of the infection (controlled by the botner herder) and the decision making process of customers. Its stationary version turns out to be exactly solvable (but not at all trivial) under an additional natural assumption that the execution time of the decisions of the customers (say, switch on or out the defence system) is much faster that the infection rates. Mathematics Subject Classification (2010): Key words: botnet defence, mean-field game, stable equilibrium, phase transitions 1 Introduction A botnet, or zombie network, is a network of computers infected with a malicious program that allows cybercriminals to control the infected machine remotely without the user s knowledge. Botnets have become a source of income for entire groups of cybercriminals since the cost of running botnets is cheap and the risk of getting caught is relatively small due to the fact that other people s assets are used to launch attacks. The interactive process of the attackers and defenders can be modeled asagame. The use of game theory in modeling attacker-defender has been extensively adopted in the computer security domain recently; see [5], [22] and [24] and bibliography there for more details. Two aspects are important. The first one is the contamination effect. The second one is the large number of computers. So, in fact, one deals with a stochastic game of a large number of interacting agents. This is amenable to Mean Field theory. To investigate this approach represents the main objective of this paper. Our model takes into account Department of Statistics, niversity of Warwick, Coventry CV4 7AL K, Email: v.kolokoltsov@warwick.ac.uk and associate member of Institute of Informatics Problems, FRC CSC RAS School of Management, The niversity of Texas at Dallas, P.O. Box 830688, SM 30 Richardson S, Graduate School of Business, The Hong Kong Polytechnic niversity, Hung Hom, Kowloon, Hong Kong, and Ajou niversity, Korea

2 both the random process of the propagation of the infection (controlled by the botnet herder) and the decision making process of customers. We develop a stationary version which turns out to be exactly solvable (but not at all trivial) under an additional natural assumption that the execution time of the decisions of the customers (say, switch on or out the defense system) is much faster that the infection rates. Similar models can be applied to the analysis of defense against a biological weapon, for instance by adding the active agent (principal interested in spreading the disease), into the general mean-field epidemic model of [23] that extends the well established SIS (susceptible-infectious-susceptible) and SIR (susceptible-infectious-recovered) models. Mean-field games present a quickly developing area of the game theory. It was initiated by Lasry-Lions [21] and Huang-Malhame-Caines [14], [15], see [1], [4], [13], [12], [6] for recent surveys, as well as [8], [7], [9], [2], [20], [25], [3] and references therein. The papers [10] and [11] initiated the study of finite-state space mean-field games that are the objects of our analysis here. The paper is organized as follows. In the next section we introduce our model, formulate the basic mean-field game (MFG) consistency problem in its dynamic and stationary versions leading to precise formulation of our main problem of characterizing the stable solutions (equilibria) of the stationary problem. This problem is a consistency problem between an HJB equation for a stochastic control of individual players and a fixed point problem for an evolutionary dynamics. These two preliminary problems are fully analyzed in Sections 3 and 4 respectively. Section 5 is devoted to the final synthesis of the stationary MFG problem from the solutions to these two preliminary problems. In particular, the phase transitions and the bifurcation points changing the number of solutions are explicitly found. In the last section further perspectives are discussed. 2 The model Assume that any computer can be in 4 states: DI,DS,I,S, where the first letter, D or, refers to the state of a defended (by some system, which effectiveness we are trying to analyze) or an unprotected computer, and the second letter, S and I, to susceptible or infected state. The change between D and is subject to the decisions of computer owners (though the precise time of the execution of her intent is noisy) and the changes between S and I are random with distributions depending on the level of efforts v H of the Herder and the state D or of the computer. Let n DI,n DS,n I,n S denote the numbers of computers in the corresponding states with N = n DS + n DI + n I + n S the total number of computers. By a state of the system we shall mean either the 4-vector n = (n DI,n DS,n I,n S ) or its normalized version x = (x DI,x DS,x I,x S ) = n/n. The fraction of defended computers x DI +x DS represents the analogue of the control parameter v D from [5], the level of defense of the system, though here it results as a compound effect of individual decisions of all players. The control parameter u of each player may have two values, 0 and 1, meaning that the player is happy with the level of defense (D or I) or she prefers to switch one to another. When the updating decision 1 is made, the updating effectively occurs after some exponential time with the parameter λ (measuring the speed of the response of the defense system). The limit λ corresponds to the immediate execution. The recovery rates (the rates of change from I to S) are given constants q D rec and q rec

3 for defended and unprotected computers respectively, and the rates of infection from the direct attacks arev H qinf D and v Hqinf respectively with constants qd inf andq inf. The rates of infectionspreadingfrominfectedtosusceptiblecomputersareβ /N,β D /N,β D /N,β DD /N, with numbers β,β D,β D,β DD, where the first (resp second) letter in the index refers to the state of the infected (resp. susceptible) computer (the scaling 1/N is necessary to make the rates of unilateral changes and binary interactions comparable in the N limit). Thus if all computers use the strategy u DS,u DI,u S,u I, u {0,1} and the level of attack is v H, the evolution of the frequencies x in the limit N can be described by the following system of ODE: ẋ DI = x DS qinf D v H x DI qrec D +x DIx DS β DD +x I x DS β D +λ(x I u I x DI u DI ), ẋ DS = x DS qinfv D H +x DI qrec D x DI x DS β DD x I x DS β D +λ(x S u S x DS u DS ), ẋ I = x S qinfv H x I qrec +x DI x S β D +x I x S β λ(x I u I x DI u DI ), ẋ S = x S qinf v H +x I qrec x DIx S β D x I x S β λ(x S u S x DS u DS ). (1) Remark 1. If all β D,β,β D,β are equal to some β, q D inf = q inf = qinf and q D rec = q D rec = v D, where v D is interpreted as the defender group s combined defense effort, then summing up the first and the third equations in (1) leads to the equation ẋ = q inf v H (1 x)+βx(1 x) v D x, (2) for the total fraction of infected computers x = x DI +x I. This equation coincides (up to some constants) with equation (2) from [5], which is the starting point of the analysis of paper [5]. It is instructive to see, how evolution (1) can be deduced rigorously as the limit of the Markov processes specifying the random dynamics of N players. The generator of this Markov evolution on the states n is (where the unchanged values in the arguments of F on the r.h.s are omitted) L N F(n DI,n DS,n I,n S ) = n DS q D inf v HF(n DS 1,n DI +1)+n S q inf v HF(n S 1,n I +1) +n DI q D rec F(n DI 1,n DS +1)+n I q rec F(n I 1,n S +1) +n DI n DS β DD F(n DS 1,n DI +1)/N +n DI n S β D F(n S 1,n I +1)/N +n I n DS β D F(n DS 1,n DI +1)/N +n I n S β F(n S 1,n I +1)/N or in terms of x as +λn DS u DS F(n DS 1,n S +1)+λn S u S F(n S 1,n DS +1) +λn DI u DI F(n DI 1,n I +1)+λn I u I F(n I 1,n DI +1), L N F(x DI,x DS,x I,x S ) = Nx DS q D infv H F(x e DS /N+e DI /N)+Nx S q infv H F((x e S /N+e I /N) +Nx DI q D rec F(x e DI/N +e DS /N)+Nx I q rec F(x e I/N +e S /N) +Nx DI x DS β DD F(x e DS /N +e DI /N)+Nx DI x S β D F(x e S /N +e I /N)

4 +Nx I x DS β D F(x e DS /N +e DI /N)+Nx I x S β F(x e S /N +e I /N) +Nλx DS u DS F(x e DS /N +e S /N)+Nλx S u S F(x e S /N +e DS /N) +Nλx DI u DI F(x e DI /N +e I /N)+Nλx I u I F(x e I /N +e DI /N), (3) where {e j } is the standard basis in R 4. If F is a differentiable function, the generator L N turns to the generator ( F LF(x DI,x DS,x I,x S ) = x DS qinfv D H F ) ( F +x S q x infv H F DI x I +x DI q D rec ( F x DS F x DI +x DI x DS β DD ( F x DI F x DS +x I x DS β D ( F x DI F x DS +λx DS u DS ( F x S F x DS +λx DI u DI ( F x I F x DI x DS ) +x I qrec ) ( F x S F x I ) +x DI x S β D ( F x I F x S ) ( F +x I x S β F ) x I x S ) ( F +λx S u S F ) x DS x S ) ( F +λx I u I F ) x DI x I ) x S in the limit N. This is a first order partial differential operator. Its characteristics are given precisely by the ODE (1). A rigorous derivation showing the solutions to (1) describe the limit of the Markov chain generated by (3) can be found e.g. in [17]. We shall now use the Markov model above to assess the actions of individual players. If x(t) and v H (t) are given, the dynamics of each individual player is the Markov chain on 4 states with the generator ) (4) L ind g(di) = λu ind (DI)(g(I) g(di))+qrec(g(ds) g(di)), D L ind g(ds) = λu ind (DS)(g(S) g(ds))+qinf D v H(g(DI) g(ds)) +x DI β DD (g(di) g(ds))+x I β D (g(di) g(ds)), L ind g(i) = λu ind (I)(g(DI) g(i))+qrec (g(s) g(i)), L ind g(s) = λu ind (S)(g(DS) g(s))+qinfv H (g(i) g(s)) +x DI β D (g(i) g(s))+x I β (g(i) g(s)) (5) depending on the individual control u ind. Assuming that an individual pays a fee k D per unit of time for the defense system and k I per unit time for losses resulting from being infected, her cost during a period of time T, that she tries to minimize, is T 0 (k D 1 D +k I 1 I )ds, (6) where 1 D (resp. 1 I ) is the indicator function of the states DI,DS (resp. of the states DI, I). Assuming that the Herder has to pay k H v H per unit of time using efforts v H and

5 receive the income f(x) depending on the distribution x of the states of the computers, her payoff, that she tries to maximize, is T Therefore, starting with some control u com = (u com t 0 (f H (x) k H v H )ds. (7) (DI),u com (DS),u com (I),u com (S)) t the Herder can find his optimal strategy v H (t) solving the deterministic optimal control problem with dynamics (1) and payoff (7) finding both optimal v H and the trajectory x(t). Once x(t) and v H (t) are known, each individual should solve the Markov control problem (5) with costs (6) thus finding the individual optimal strategy u ind t = (u ind t (DI),u ind t (DS),u ind t (I),u ind t (S)). The basic MFG consistency equation can now be explicitly written as u ind t t = u com t. Instead of analyzing this rather complicated dynamic problem, we shall look for a simpler problem of consistent stationary strategies. There are two standard stationary problems naturally linked with a dynamic one, one being the search for the average payoff 1 g = lim T T T 0 (k D 1 D +k I 1 I )dt for long period game, and another the search for discounted optimal payoff. The first is governed by the solutions of HJB of the form (T t)µ+g, linear in t (then µ describing the optimal average payoff), so that g satisfies the stationary HJB equation: λminu(g(i) g(di))+q u rec(g(ds) g(di))+k D I +k D = µ, λminu(g(s) g(ds))+qinf D v u H(g(DI) g(ds)) +x DI β DD (g(di) g(ds))+x I β D (g(di) g(ds))+k D = µ, λminu(g(di) g(i))+qrec (g(s) g(i))+k I = µ, u λminu(g(ds) g(s))+q u infv H (g(i) g(s)) +x DI β D (g(i) g(s))+x I β (g(i) g(s)) = µ where min is over two values {0,1}. We shall denote u = (u DI,u I,u DS,u S ) the argmax in this solution. The discounted optimal payoff (with the discounting coefficient ) satisfies the stationary HJB t (8)

6 λminu(g(i) g(di)+qrec D (g(ds) g(di))+k I +k D = g(di), u λminu(g(s) g(ds))+q u infv D H (g(di) g(ds)) +x DI β DD (g(di) g(ds))+x I β D (g(di) g(ds))+k D = g(ds), λminu(g(di) g(i))+q u rec(g(s) g(i))+k I = g(i), λminu(g(ds) g(s))+qinf v u H(g(I) g(s)) +x DI β D (g(i) g(s))+x I β (g(i) g(s)) = g(s) (9) The analysis of these two settings is mostly analogous. We shall concentrate on the first one. Introducing the coefficients α = q D inf v H +x DI β DD +x I β D, β = q inf v H +x DI β D +x I β, (10) the stationary HJB equation (8) rewrites as λmin(g(i) g(di),0)+qrec D (g(ds) g(di))+k I +k D = µ, λmin(g(s) g(ds),0)+α(g(di) g(ds))+k D = µ, λmin(g(di) g(i),0)+qrec (g(s) g(i))+k I = µ, λmin(g(ds) g(s),0)+β(g(i) g(s)) = µ, where the choice of the first term as the infimum in these equations corresponds to the choice of control u = 1. The stationary MFG consistency problem is in finding x = (x DI,x DS,x I,x S ) and u = (u DI,u DS,u I,u S ), where x is the stationary point of evolution (1), that is x DS α x DI qrec D +λ(x Iu I x DI u DI ) = 0 x DS α+x DI qrec D +λ(x S u S x DS u DS ) = 0 (12) x S β x I qrec λ(x I u I x DI u DI ) = 0 x S β +x I qrec λ(x S u S x DS u DS ) = 0, with u = (u DI,u DS,u I,u S ) giving minimum in the solution to (8) or (11). Thus x is a fixed point of the limiting dynamics of the distribution of large number of agents such that the corresponding stationary control is individually optimal subject to this distribution. Yet in other words, x = (x DI,x DS,x I,x S ) and u = (u DI,u DS,u I,u S ) solve (8), (12) simultaneously. Fixed points can practically model a stationary behavior only if they are stable. Thus we are interested in stable solutions (x, u) to the stationary MFG consistency problem (12),(8), where a solution is stable if the corresponding stationary distribution x is a stable equilibrium to (1) (with u fixed by this solution). Apart from stability, the fixed points can be classified via its efficiency. Namely, let us say that a solution to the stationary MFG is efficient (or globally optimal) if the corresponding average cost µ is minimal among all other solutions. Talking about strategies, let us reduce the discussion to non-degenerate situations, where the minima in (11) are achieved on a single value of u only. In principle, there are (11)

7 16 possible pure stationary strategies (functions from the state space to {0, 1}). But not all of them can be realized as solutions to (11). In fact if u DI = 1, then g(i) < g(di) (can be equal in degenerate case) and thus u I = 0. This argument forbids all but four strategies as possible solutions to (11), namely (i) g(i) g(di), g(s) g(ds) u I = u S = 0, u DI = u DS = 1, (ii) g(i) g(di), g(s) g(ds) u DI = u DS = 0, u I = u S = 1, (iii) g(i) g(di), g(s) g(ds) u I = u DS = 0, u DI = u S = 1, (iv) g(i) g(di), g(s) g(ds) u DI = u S = 0, u I = u DS = 1. (13) The first two strategies, either always choose or always choose D, are acyclic, that is the corresponding Markov processes are acyclic in the sense that there does not exist a cycle in a motion subject to these strategies. Other two strategies choose between and D differently if infected or not. Of course, allowing degenerate strategies, more possibilities arise. To complete the model, let us observe that the natural assumptions on the parameters of the model arising directly from their interpretation are as follows: qrec D qrec, qinf D < qinf, β D β, β DD β D, (14) k D k I. We shall always assume (14) hold. Two additional natural simplifying assumptions that we shall use sometimes are the following: the infection rate does not depend on the level of defense of the computer transferring the infection, but only on the level of defence of the susceptible computer, that is, instead of four coefficients β one has only 2 of them β = β D = β, β D = β D = β DD, (15) and the recovery rate do not depend on whether a computer is protected against the infection or not: q rec = q D rec = q rec. (16) As we shall see, a convenient assumption, which is weaker than (16), turns out to be q D rec q rec < (q inf qd inf )v H. (17) Finally, it is reasonable to assume that customers can switch rather quickly their regime of defence (once they are willing to) meaning that we are effectively interested in the asymptotic regime of large λ. As we shall show, in this regime the stationary MFG problem above can be completely solved analytically. In this sense the present model is more complicated than a related mean-field game model of corruption with three basic states developed in [19], where a transparent analytic classification of stable solutions is available already for arbitrary finite λ. 3 Analysis of the stationary HJB equation Let us start by solving HJB equation (11).

8 Consider strategy (i) of (13), so that being unprotected is always optimal. Then (11) becomes λ(g(i) g(di))+qrec(g(ds) g(di))+k D I +k D = µ, λ(g(s) g(ds))+α(g(di) g(ds))+k D = µ, qrec (g(s) g(i))+k (18) I = µ, β(g(i) g(s)) = µ. As the solution g is defined up to an additive constant we can set g(s) = 0. Then (18) becomes λ(g(i) g(di))+qrec(g(ds) g(di))+k D I +k D = µ, λg(ds)+α(g(di) g(ds))+k D = µ, (19) qrecg(i)+k I = µ, βg(i) = µ. From the third and fourth equations we find g(i) = k I β +q rec, µ = βg(i) = βk I. (20) β +qrec Substituting these values in the first and second equations we obtain g(ds) = k D µ α(β +λ+q +k rec) I λ λ(β +qrec)(α+λ+q rec), D g(di) = k D µ (α+λ)(β +λ+qrec +k ) I λ λ(β +qrec )(α+λ+qd rec ), and the conditions g(i) g(di),g(s) = 0 g(ds) become (21) k D (β +q rec )(α+λ+qd rec ) k I[(β +λ)q D rec (α+λ)q rec ], k D (β +q rec )(α+λ+qd rec ) k I[β(λ+q D rec ) α(λ+q rec )] (22) respectively. Consider strategy (ii) of (13), so that being defended is optimal. Then (11) becomes qrec D (g(ds) g(di))+k I +k D = µ, α(g(di) g(ds))+k D = µ, λ(g(di) g(i))+qrec (g(s) g(i))+k (23) I = µ, λ(g(ds) g(s))+β(g(i) g(s)) = µ. Setting g(ds) = 0 yields qrec D g(di))+k I +k D = µ, αg(di)+k D = µ, λ(g(di) g(i))+qrec (g(s) g(i))+k I = µ, λg(s))+β(g(i) g(s)) = µ. (24) From the first and second equations we find g(di) = k I α+q D rec, µ = k D +αg(di) = α(k D +k I )+k D qrec D. (25) α+qrec D

9 Substituting these values in the third and fourth equations we obtain g(s) = k D λ +k β(λ+qrec D ) α(λ+q rec ) I λ(α+qrec)(β D +λ+qrec), g(i) = k D λ +k (β +λ)(λ+qrec D ) αq rec I λ(α+qrec D )(β +λ+q rec ). (26) and the conditions g(i) g(di), g(s) g(ds) = 0 turn to k D (α+q D rec )(β +λ+q rec ) k I[(β +λ)q D rec (α+λ)q rec ], k D (α+q D rec )(β +λ+q rec ) k I[β(λ+q D rec ) α(λ+q rec )] (27) respectively. Consider strategy (iii) of (13). Then (11) becomes λ(g(i) g(di))+qrec(g(ds) g(di))+k D I +k D = µ, α(g(di) g(ds))+k D = µ, qrec(g(s) g(i))+k I = µ, λ(g(ds) g(s))+β(g(i) g(s)) = µ. (28) Setting g(ds) = 0 yields from the second equation, then µ = αg(di)+k D from the first equation and λg(i) = g(di)(α+λ+q D rec) k I g(s) = g(di) [αλ+q λqrec rec (α+λ+qd rec )]+ λk D (λ+qrec )k I λqrec from the third one. Plugging these expressions in the fourth equation of (28) we find (after many cancelations) g(di) and then the other values of g: (β +λ+qrec g(di) = )(k I k D ) α(β +λ+qrec)+q rec(α+λ+q rec), D g(s) = 1 k I [β(λ+qrec) α(λ+q D rec)] k D (β +qrec)(α+λ+q rec) D λ α(β +λ+qrec )+q rec (α+λ+qd rec ), g(i) = 1 k I [(λ+qrec D )(λ+β) αq rec ] k D(β +λ+qrec )(α+λ+qd rec ) λ α(β +λ+qrec)+q rec(α+λ+q rec) D (29) Hence µ = k Iα(β +λ+qrec )+k Dqrec (α+λ+qd rec ). α(β +λ+qrec)+q rec(α+λ+q rec) D The conditions g(i) g(di), g(s) g(ds) rewrite as { kd (α+q D rec )(β +λ+q rec ) k I[(β +λ)q D rec (α+λ)q rec ], k D (α+λ+q D rec)(β +q rec) k I [β(λ+q D rec) α(λ+q rec)]. (30)

10 Consider strategy (iv) of (13). Then (11) becomes qrec D (g(ds) g(di))+k I +k D = µ, λ(g(s) g(ds))+α(g(di) g(ds))+k D = µ, λ(g(di) g(i))+qrec (g(s) g(i))+k I = µ, β(g(i) g(s)) = µ. (31) Setting g(s) = 0 yields µ = βg(i) from the fourth equation, then from the third equation and λg(di) = g(i)(β +λ+q rec ) k I λq D rec g(ds) = g(i)[βλ+qd rec (β +λ+q rec )] k Dλ k I (λ+q D rec ) from the first one. Plugging these expressions in the second equation of (31) we find g(i) and then the other values of g: g(i) = g(ds) = 1 λ (k D +k I )(α+λ+q D rec ) β(α+λ+q D rec)+q D rec(β +λ+q rec), k D (β +λ+q rec )(α+qd rec )+k I[α(λ+q rec ) β(λ+qd rec )] β(α+λ+q D rec )+qd rec (β +λ+q rec ), g(di) = 1 k D (β +λ+qrec)(α+λ+q rec)+k D I [(α+λ)(λ+qrec) βq rec] D λ β(α+λ+qrec D )+qd rec (β +λ+q rec ). Hence the conditions g(i) g(di), g(ds) g(s) = 0 rewrite as { kd (α+λ+q D rec)(β +q rec) k I [(β +λ)q D rec (α+λ)q rec], k D (α+q D rec )(β +λ+q rec ) k I[β(λ+q D rec ) α(λ+q rec )]. (33) We are now interested in finding out how many solutions equation (11) may have for a given x. The first observation in this direction is that the interior of the domain defined by (22) (that is, with a solution of case (i)) and the interior of the domain defined by (30) (that is, with a solution of case (iii)) do not intersect, because the first inequality in (22) contradicts the second inequality in (30) (apart from the boundary). Similarly, the interior of the domain defined by (22) (that is with a solution of case (i)) and the interior of the domain defined by (33) (that is, with a solution of case (iv)) do not intersect, and the interior of the domain defined by (27) (that is, with a solution of case (ii)) does not intersect with the domains having solutions in cases (iii) or (iv). Next we find that one can distinguish two natural domains of x classifying the solutions to HJB equation (11): More explicitly, D 1 = {x : β +q rec > α+qd rec }, D 2 = {x : β +q rec < α+qd rec }. D 1 = {x : x DI (β D β DD )+x I (β β D ) > (q D inf q inf )v H +q D rec q rec }. (32)

11 By (14) it is seen that under a natural additional simplifying assumptions (16) or even (17), all positive x belong to D 1 (or its boundary), so that D 2 is empty. nder additional assumption (15) the condition x D 1 gets a simpler form x > x = (qd inf q inf )v H +q D rec q rec β β D. (34) To link with the conditions for cases (i)-(iv) one observes the following equivalent forms of the main condition of being in D 1 : β +q rec > α+qd rec (β +q rec )(α+qd rec +λ) > (α+qd rec )(β +q rec +λ) β(λ+q D rec) α(λ+q rec) > (β +λ)q D rec (α+λ)q rec. (35) From here it is seen that if x belongs simultaneously to the interiors of the domains specified by (22) and (27) (that is, with solutions in cases (i) and (ii) simultaneously), then necessarily x D 1 (that is, for x D 2 the conditions specifying cases (i) and (ii) are incompatible). On the other hand, if x belongs simultaneously to the interiors of the domains specified by (30) and (33) (that is, with solutions in cases (iii) and (iv) simultaneously), then necessarily x D 2 (that is, for x D 1 the conditions specifying cases (iii) and (iv) are incompatible). Denoting κ = k D /k i, we can summarize the properties of HJB equation (11) as follows (uniqueness is always understood up to the shifts in g). Proposition 3.1. Suppose x D 1. (1) If (β +λ)qrec D (α+λ)qrec (β +qrec +λ)(α+qd rec ) < κ < β(λ+qd rec) α(λ+qrec), (36) (β +qrec )(α+qd rec +λ) then there exists a unique solution to (11) belonging to case (iii) and there are no other solutions to (11). (2) If β(λ+qrec D ) α(λ+q rec ) (β +qrec +λ)(α+qd rec ) < κ < (β +λ)qd rec (α+λ)q rec, (37) (β +qrec )(α+qd rec +λ) then there exists a unique solution to (11) belonging to case (iv) and there are no other solutions to (11). (3) A solution belonging to case (i) exists if and only if κ β(λ+qd rec) α(λ+qrec), (38) (β +qrec )(α+qd rec +λ) and is unique if this holds. A solution belonging to case (ii) exists if and only if κ (β +λ)qd rec (α+λ)q rec (β +q rec +λ)(α+q D rec), (39) and is unique if this holds. Either of conditions (38) or (39) is incompatible with either (36) or (37). In particular, equation (11) may have at most two solutions (if both (38) and (39) hold).

12 (4) nder (16), one has always β(λ+qrec D ) α(λ+q rec ) (β +qrec +λ)(α+qd rec ) (β +λ)qd rec (α+λ)q rec, (40) (β +qrec )(α+qd rec +λ) and (β +λ)qrec D (α+λ)q rec (β +qrec +λ)(α+qd rec ) β(λ+qd rec ) α(λ+q rec ). (41) (β +qrec )(α+qd rec +λ) Hence (37) becomes impossible and conditions (38) and (39) become incompatible implying the uniqueness of the solution to (11) for any x D 1. This unique solution belongs to cases (ii), (iii) and (i) respectively for κ satisfying (39), (36), (38) (when equality holds in (38) or (39), two solutions from different cases become coinciding). Proof. Statements (1)-(3) follow from the arguments given above. (iv) nder (16), conditions (41) and (40) rewrite as and which obviously hold. q rec (β α) 2 (β α)(β +λ+q rec )(α+q rec ) 0 q rec (β α) 2 +(β α)(β +q rec )(α+λ+q rec ) 0, Remark 2. (1) When (16) does not hold one can find situations when solutions from cases (i) and (ii) exist simultaneously. To get simple examples one can assume κ = 1. (2) When two solutions exist simultaneously one can discriminate them by the values of the average payoff µ. One sees from (20) and (25), that µ arising from cases (i) and (ii) are different (apart from a single value of κ). (3) The uniqueness result under (16) is quite remarkable, as it does not seem to follow a priori from any intuitive arguments. Again directly from the argument above one can conclude the following. Proposition 3.2. Suppose x D 2. (1) If κ > (β +λ)qd rec (α+λ)qrec, (42) (β +qrec )(α+qd rec +λ) then there exists a unique solution to (11) belonging to case (i) and there are no other solutions to (11). (2) If κ < β(λ+qd rec ) α(λ+q rec ) (β +qrec +λ)(α+qrec), (43) D then there exists a unique solution to (11) belonging to case (ii) and there are no other solutions to (11). (3) A solution belonging to case (iii) exists if and only if (β +λ)q D rec (α+λ)q rec (β +q rec +λ)(α+q D rec) κ β(λ+qd rec ) α(λ+q rec ) (β +q rec)(α+q D rec +λ), (44) and is unique if this holds. A solution belonging to case (iv) exists if and only if β(λ+qrec D ) α(λ+q rec ) (β +qrec +λ)(α+qd rec ) κ (β +λ)qd rec (α+λ)q rec, (45) (β +qrec )(α+qd rec +λ)

13 and is unique if this holds. Either of conditions (44) or (45) is incompatible with either (42) or (43). In particular, equation (11) may have at most two solutions (if (44) and (45) hold simultaneously). Essential simplifications that allow eventually for a full classification of the stationary MFG consistency problem occur in the limit of large λ. For a precise formulation in case = q D rec q rec > 0 (46) one needs further decomposition of the domains D 1,D 2. Namely, for j = 1,2, let D j1 = {x D j : α+q D rec < β α }. β +qrec Proposition 3.3. The following hold for large λ outside an interval of κ of size of order λ 1 : (1) nder (16) conditions (39), (36), (38) classifying the solutions to the HJB equation rewrite as (β α) (β α) κ 0, 0 < κ <, κ β +q β +q, (47) respectively. In particular, solutions of case (ii) become impossible. (2) Suppose x D 1 and (46) holds. If x D 11, there exists a unique solution to (11), which belongs to cases (ii), (iii), (i) for κ < α+q D rec, α+q D rec < κ < β α, κ > β α, (48) β +qrec β +qrec respectively. If x D 12, solutions from case (iii) do not exist and there exist two solutions to (11) for β α < κ <, (49) β +qrec α+qrec D belonging to cases (i) and (ii), and only one solution otherwise. (3) Suppose x D 2. If x D 22, solutions from case (iii) do not exist and there is always a unique solution to (11) belonging to case (ii), (iv) or (i), for κ < β α, α+qrec D β α α+q D rec < κ <, κ > β +qrec, (50) β +qrec respectively. If x D 21, then there are two solutions to (11) for α+q D rec < κ < β α, (51) β +qrec which belong to cases (iii) and (iv), and one solution otherwise. This unique solution belongs to case (ii) or (i) for respectively and to case (iv) otherwise. κ < β α, κ > α+qrec D β +q rec

14 Proof. Statement (ii) follows from the observation that, for > 0 and large λ, conditions (36) - (37) turn to α+q D rec < κ < β α, β +qrec β α α+q D rec < κ < β +q rec (52) respectively, and conditions (38) - (39) turn to κ β α, κ β +qrec α+q D rec (53) respectively. Other statements are similar. 4 Analysis of the fixed points Next we are solving the fixed point system (12). In case (i), that is with u I = u S = 0,u DI = u DS = 1, equation (12) takes the form x DS α x DI qrec D λx DI = 0 x DS α+x DI qrec D λx DS = 0 (54) x S β x I qrec +λx DI = 0 x S β +x I qrec +λx DS = 0. Adding the first two equations we get x DI = x DS = 0, and the system reduces to the single equation x S β x I q rec = 0. Substituting the value of β yields x S (q infv H +x I β ) x I q rec = 0. Denoting y = x I it follows that x S = 1 y and thus Q (y) = β y 2 +y(q rec β +q infv H ) q infv H = 0. This equation has a unique solution on the interval (0,1): x = x I = 1 [ ] β qrec 2β q inf v H + (β +qinf v H) 2 +(qrec )2 2qrec (β qinf v H). (55) The stability of the fixed point x = (0,0,x,1 x ) means its stability as a fixed point of the dynamics ẋ DI = x DS α x DI qrec D λx DI ẋ DS = x DS α+x DI q D rec λx DS ẋ I = x S β x I qrec +λx DI ẋ S = x S β +x I qrec +λx DS. (56)

15 We rewrite it by shifting the variables by the value of the stationary point, that is, in terms of x DI,x DS,y = x I x,z = x S (1 x ). Since the sum of these variables is one, we have effectively the system of three equations on the variables x DI,x DS,y: ẋ DI = [qinfv D H +x DI β DD +(x +y)β D ]x DS (λ+qrec)x D DI ẋ DS = [qinf D v H +x DI β DD +(x +y)β D ]x DS +qrec D x DI λx DS ẏ = (1 x y x DI x DS )[qinf v H +β D x DI +β (y +x )] (y +x )qrec +λx DS. Its linear approximation around the fixed point (0,0,0) is ẋ DI = (λ+qrec D )x DI +(qinf D v H +x β D )x DS ẋ DS = qrec D x DI (qinf D v H +x β D +λ)x DS ẏ = (1 x )[β D x DI +β y] (y +x DI +x DS )(qinf v H +x β ) qrec y +λx DS, and the corresponding characteristic equation for the eigenvalues ξ is [(1 x )β (q inf v H +x β ) q rec ξ] [(ξ +λ+qinfv H +x β )(ξ +λ+qrec) q D rec(q D infv H +x β )] = 0. The free term cancels in the second multiplier and we get the eigenvalues ξ 1 = (1 x )β qinf v H x β qrec ξ 2 = λ (qrec D +q inf v H +x β ) ξ 3 = λ. (57) The second and the third eigenvalues being negative, the condition of stability is reduced to the negativity of the first eigenvalue, that is, to the condition 2x > 1 q rec +q inf v H β. (58) But it always holds for x of form (55). Thus we proved the first part of the following statement and the second is analogous. Proposition 4.1. (1) There exists a unique solution to system (12) with the strategy being individually optimal (that is, with the first acyclic stationary strategy u I = u S = 0,u DI = u DS = 1) and it is stable. It equals x = (0,0,x I,1 x I ) with x I given by (55). (2) There exists a unique solution to system (12) with the strategy D being individually optimal (that is, with the second acyclic stationary strategy) and it is stable. It equals x = (x DI,1 x DI,0,0) with x DI being the unique solution of equation on the interval (0,1), that is Q D (y) = β DD y 2 +y(q D rec β DD +q D inf v H) q D inf v H = 0 (59) x DI = 1 [ ] β DD qrec D q D 2β infv H + (β DD +qinf D v H) 2 +(qrec) D 2 2qrec(β D DD qinf D v H). DD

16 Let us consider case (iii): u I = u DS = 0,u DI = u S = 1. Then (12) takes the form x DS α x DI qrec D λx DI = 0 x DS α+x DI qrec D +λx S = 0 x S β x I qrec +λx (60) DI = 0 x S β +x I qrec λx S = 0. By adding the first two equations we get x DI = x S with two independent equations left: { xds α (q D rec +λ)x DI = 0 x DI (β +λ) x I q rec = 0. (61) This rewrites as two equations on the two independent variables x DI,x I as { (1 xi 2x DI )(q D inf v H +β DD x DI +β D x I ) (q D rec +λ)x DI = 0 x DI (q infv H +β D x DI +β x I +λ) x I q rec = 0. Solving the second equation with respect to x I, (62) x I = x DI(q inf v H +x DI β D +λ) q rec β x DI, (63) and substituting in the first one, leads to a fourth order equation on y = x DI. This equation does not seem to be much revealing in general. Of course it can be fully analyzed by numeric methods, but we shall turn now to the large λ asymptotics that yields more manageable results. For large λ we get directly from (63) that x I = x DI λ q rec β x DI (1+O(λ 1 )). But this implies that x DI is small of order O(λ 1 ), so that x I = x DIλ q rec Substituting this in the first equation of (62) yields (1+O(λ 1 )), x DI = x Iqrec (1+O(λ 1 )). (64) λ β D x 2 I +x I (q rec β D +q D infv H ) q D infv H = O(λ 1 ), (65) which is of the same type as equations (59) up to terms of order λ 1 (and coincides with it under (15), (16)). Therefore, for large λ, there exists a unique solution to (65) from the interval (0, 1): x I = O(λ 1 ) + 1 [ ] β D qrec q D 2β infv H + (β D +qinf D v H) 2 +(qrec) 2 2qrec(β D qinf D v H). D (66)

17 The stability of the fixed point x = (x DI,x DS = 1 x I 2x DI, x I,x S = x DI ) means its stability as a fixed point of the dynamics ẋ DI = x DS α x DI qrec D λx DI ẋ DS = x DS α+x DI qrec D +λx S ẋ I = x S β x I qrec +λx (67) DI ẋ S = x S β +x I qrec λx S. In terms of independent variables x DI = x DI x DI, x S = x S x S, y = x I = x I x I this rewrites as dt x d DI = (1 y x I x DI x S )α x DI qrec D λ( x DI +x DI)+O(λ 1 ) ẏ = x S β (y + x I)qrec +λ( x DI +x DI)+O(λ 1 ) dt x d S = x S β +(y + x I)qrec λ( x S +x S)+O(λ 1 ). with α = q D infv H + x DI β DD +(y + x I)β D +O(λ 1 ), β = q infv H + x DI β D +(y + x I)β +O(λ 1 ). Linearized around the fixed point (0,0,0) system (68) takes the form dt x d DI = (y + x DI + x S )(qinf D v H + x I β D)+(1 x I )( x DIβ DD + x I β D ) x DI (qrec D +λ) ẏ = x S (qinf v H + x I β ) x I qrec +λ x DI dt x d S = x S (qinfv H + x Iβ )+ x I qrec λ x S up to terms of order O(λ 1 ). Thus the matrix of the linear approximation divided by λ is O(λ 1 ) 1 [ qrec D v H +β D (1 2 x I )]/λ+o(λ 2 ) O(λ 1 ) M(λ) = O(λ 1 )+1 qrec/λ+o(λ 2 ) O(λ 1 ). O(λ 1 ) O(λ 1 ) O(λ 1 ) 1 The first order approximation of this matrix in λ 1 is 1 0 0 M 0 = 1 0 0. 0 0 1 and has eigenvalue 1 of double multiplicity and a zero eigenvalue. Hence all eigenvalues of M(λ) are negative if and only if its determinant det(m(λ)) is negative. As seen directly det(m(λ)) = [ q rec qd rec v H +β D (1 2 x I )]/λ+o(λ 2 ), (68)

18 and is negative for large λ if and only if β D (2 x I 1) > q rec +q D recv H, which always holds by (66). Thus we proved the first part of the following statement and the second part is analogous. Proposition 4.2. (1) For large λ there exists a unique solution to system (12) in case (iii), that is with u I = u DS = 0,u DI = u S = 1, and it is stable. It has the form x = (0,1 x I, x I,0) up to corrections of order λ 1, with x I being the unique solution of equation (65) on (0,1) given by (66). (2) For large λ there exists a unique solution to system (12) in case (iv), that is with u I = u DS = 1,u DI = u S = 0, and it is stable. It has the form x = ( x DI,0,0,1 x I ) up to corrections of order λ 1, with x DI being the unique solution of equation on (0,1). β D x 2 DI +x DI (q rec β D +q infv H ) q infv H = O(λ 1 ), (69) 5 Solutions to the stationary MFG problem Combining Propositions 4.1, 4.2 and 3.3 allows one to fully characterize the solutions to our stationary MFG consistency problem for large λ. The most straightforward general conclusion is the following. Theorem 5.1. For large λ there may exist up to 4 solutions to the stationary MFG problem, with only one in each of the cases (i) -(iv). All these solutions are stable. Remark 3. Notice that already this statement is not at all obvious a priori, and may not be true for finite λ, where solutions to case (iii) or (iv) are found from an equation of fourth order. As an example of more precise classification, let us present it under assumption (17) that ensures that all solutions lie in the domain D 1. Let us introduce the function κ(z) = (q inf qd inf )v H +z(β β D ) qinf v. H +zβ +q First let (16) hold. It is seen from Propositions 4.1 and 4.2 that for large λ, (and apart from κ from negligible intervals of size of order λ 1 that we shall ignore), a solution of the stationary MFG problem exists in case (i) if κ > κ = κ(x I), and a solution of the stationary MFG problem exists in case (iii) if κ < κ = κ( x I ), where x I and x I are given by (55) and (66) respectively. Thus one can have up to two (automatically stable) solutions to the stationary MFG problem. Let us make this number precise.

19 Differentiating κ(z) we find directly that it is increasing if β (q D infv H +q) > β D (q infv H +q), (70) and decreasing otherwise. Hence the relation κ > κ is equivalent to the same or the opposite relation for x I and x I. Thus we are led to the following conclusion. Theorem 5.2. Let (16) hold. (1) If (70) holds and x I > x I, or the opposite to (70) holds and x I < x I, then κ > κ. Consequently, for κ < κ there exists a unique solution to the stationary MFG problem, which is stable and belongs to case (iii); for κ ( κ,κ ) there are no solutions to the stationary MFG problem; for κ > κ there exists a unique solution to the stationary MFG problem, which is stable and belongs to case (i). (2) If (70) holds and x I < x I, or the opposite to (70) holds and x I > x I, then κ < κ. Consequently, for κ < κ there exists a unique solution to the stationary MFG problem, which is stable and belongs to case (iii); for κ (κ, κ ) there exist two (stable) solutions to the stationary MFG problem; for κ > κ there exists a unique solution to the stationary MFG problem, which is stable and belongs to case (i). Thus if one considers the system for all parameters fixed except for κ (essentially specifying the price of the defence service), points κ and κ are the bifurcation points, where the phase transitions occur. To deal with case when (16) does not hold let us introduce the numbers κ 1 = β α (x β +qrec I ), κ 2 = α+q D rec (x DI ), κ 3 = α+q D rec ( x I ), κ 4 = β α ( x β +qrec I ), where x I,x DI, x I in brackets mean that α,β defined in (10) are evaluated at the corresponding solutions given by Propositions 4.1 and 4.2. Since x I,x DI, x I are expressed in terms of different parameters, any order relation between them are to be expected in general. Of course, restrictions appear under additional assumptions, for instance x DI = x I under (15). From Proposition 3.3 we deduce the following. Theorem 5.3. Let (17) and (46) hold. Depending on the order relation between x I,x DI, x I, one can have up to 3 solutions to the stationary MFG problem for large λ, the characterization in each case being fully explicit, since for κ > κ 1, there exists a unique solution in case (i), for κ < κ 2, there exists a unique solution in case (ii), for κ 3 < κ < κ 4, there exists a unique solution in case (iii). Thus in this case the points κ 1,κ 2,κ 3,κ 4 are the bifurcation points, where the phase transitions occur. The situation when(17) does not hold is analogous, though there appears an additional bifurcation relating to x crossing the border between D 1 and D 2, and the possibility of having four solutions arises. 6 Discussion Our model of four basic states is of course the simplest one that takes effective account of both interaction (infection) and rational decision making. It suggests extensions in

20 various directions. For instance, it is practically important to allow for the choice of various competing protection systems, leading to a model with 2d basic states: ii and is, where i {1,,d} denotes the ith defense system available (which can be alternatively interpreted as the levels of protection provided by a single or different firms), while S and I denote again susceptible or infected state, with all other parameters depending on i. On the other hand, in the spirit of papers [18], [17] that concentrate on modeling myopic behavior (rather than rational optimization) of players one can consider the set of computer owners consisting of two groups, rational optimizers and those changing their strategies by copying their neighbors. The main theoretical question arising from our results concerns the rigorous relation between stationary and dynamic MFG solutions, which in general is in front of research in the mean-field game literature. We hope that working with our simple model with fully solved stationary version can help to get new insights in this direction. In the present context the question can be formulated as follows. Suppose that, if at some moment of time N players are distributed according certain frequency vector x among the four basic state, each player chooses the optimal strategy u arising from the solution of the stationary problem for fixed x (fully described in Section 3), and the Markov evolution continues according to the generator L. When two solutions are available, players may be supposed to choose the one with the lowest µ (see see Remark 2 (2)). The resulting changes in x induce the corresponding changes of u specifying a well-defined Markov process on the states of N agents. Intuitively, we would expect this evolution stay near our stationary MFG solutions for large N and t. Can one prove something like that? References [1] M. Bardi, P. Caines and I. Capuzzo Dolcetta. Preface: DGAA special issue on mean field games. Dyn. Games Appl. 3:4 (2013), 443 445. [2] R. Basna, A. Hilbert and V. Kolokoltsov. An epsilon-nash equilibrium for non-linear Markov games of mean-field-type on finite spaces. Commun. Stoch. Anal. 8:4 (2014), 449-468. [3] D. Bauso, H. Tembine, and T. Basar. Robust mean-field games. Dynamic Games and Applications, online, 6 June 2015, 10.1007/s13235-015-0160-4 [4] A. Bensoussan, J. Frehse, P. Yam. Mean Field Games and Mean Field Type Control Theory. Springer, 2013. [5] A Bensoussan, S. Hoe and M. Kantarcioglu. A Game-Theoretical Approach for Finding Optimal Strategies in a Botnet Defense Model. Decision and Game Theory for Security First International Conference, GameSec 2010, Berlin, Germany, November 22-23, 2010. Proceeding, T. Alpcan, L. Buttyan, and J. Baras (Eds.), Vol. 6442 pp. 135-148. [6] P. E. Caines, Mean Field Games, Encyclopedia of Systems and Control, Eds. T. Samad and J. Ballieul. Springer Reference 364780; DOI 10.1007/978-1-4471-5102-9 30-1, Springer-Verlag, London, 2014.

21 [7] P. Cardaliaguet, J-M. Lasry, P-L. Lions and A. Porretta. Long time average of mean field games with a nonlocal coupling. SIAM J. Control Optim. 51:5 (2013), 3558 3591. [8] R. Carmona and D. Lacker. A probabilistic weak formulation of mean field games and applications. Ann. Appl. Probab. 25:3 (2015), 1189 1231. [9] R. Carmona and F. Delarue. Probabilistic analysis of mean-field games. SIAM J. Control Optim. 514 (2013), 2705 2734. [10] D. A. Gomes, J. Mohr and R. R. Souza. Discrete time, finite state space mean feld games. J. Math. Pures Appl. 93:3 (2010), 308-328. [11] D. A. Gomes, J. Mohr and R. R. Souza. Continuous time finite state space mean feld games. Appl. Math. Optim. 68:1 (2013), 99-143. [12] D. A. Gomes and J. Saude. Mean field games models a brief survey. Dyn. Games Appl. 4:2 (2014), 110 154. [13] O. Guéant O, J-M. Lasry and P-L. Lions. Mean Field Games and Applications. Paris-Princeton Lectures on Mathematical Finance 2010. Lecture Notes in Math. 2003, Springer, Berlin, p. 205-266. [14] M. Huang, R. Malhamé, P. Caines. Large population stochastic dynamic games: closed-loop Mckean-Vlasov systems and the Nash certainty equivalence principle. Communications in information and systems 6 (2006), 221 252. [15] M. Huang, P. Caines and R. Malhamé. Large-Population Cost-Coupled LQG Problems With Nonuniform Agents: Individual-Mass Behavior and Decentralized ǫ-nash Equilibria. IEEE Trans Automat Control 52:9 (2007), 1560 1571. [16] V. N. Kolokoltsov. Nonlinear Markov processes and kinetic equations. Cambridge Tracks in Mathematics 182, Cambridge niv. Press, 2010. See the review of Prof. D. Applebaum in Bull. London Math. Soc. (2011) 43(6): 1245-1247. [17] V. N. Kolokoltsov. Nonlinear Markov games on a finite state space (mean-field and binary interactions). International Journal of Statistics and Probability 1:1 (2012), 77-91. http://www.ccsenet.org/journal/index.php/ijsp/article/view/16682 [18] V. N. Kolokoltsov. The evolutionary game of pressure (or interference), resistance and collaboration (2014). http://arxiv.org/abs/1412.1269 [19] V. N. Kolokoltsov and O. A. Malafeyev. Mean field game model of corruption (2015). http://arxiv.org/abs/1507.03240 [20] V. Kolokoltsov, M. Troeva and W. Yang. On the rate of convergence for the meanfield approximation of controlled diffusions with large number of players. Dyn. Games Appl. 4:2 (2014), 208 230. [21] J-M. Lasry and P-L. Lions. Jeux à champ moyen. I. Le cas stationnaire (French). C.R. Math. Acad. Sci. Paris 343:9 (2006) 619-625.

22 [22] Zh. Li, Q. Liao and A. Striegel. Botnet Economics: ncertainty Matters. http://weis2008.econinfosec.org/papers/liao.pdf [23] J. Liu, Y. Tang and Z. R. Yang. The spread of disease with birth and death on networks. arxiv:q-bio/0402042v3. Published in Journal of Statistical Mechanics: Theory and Experiment, 2004. [24] K-W. Lye, J. M. Wing. Game strategies in network security. Int J Inf Secur 4 (2005), 71-86. [25] H. Tembine, Q. Zhu and T. Basar. Risk-sensitive mean-field games. IEEE Trans. Automat. Control 59:4 (2014), 835 850.