1 AUTOCRATIC STRATEGIES

Size: px
Start display at page:

Download "1 AUTOCRATIC STRATEGIES"

Transcription

1 AUTOCRATIC STRATEGIES. ORIGINAL DISCOVERY Recall that the transition matrix M for two interacting players X and Y with memory-one strategies p and q, respectively, is given by p R q R p R ( q R ) ( p R )q R ( p R )( q R ) M = p S q T p S ( q T ) ( p S )q T ( p S )( q T ) p T q S p T ( q S ) ( p T )q S ( p T )( q S ). (.) p P q P p P ( q P ) ( p P )q P ( p P )( q P ) M is a stochastic matrix (all rows sum up to ) and hence has an eigenvalue of. The stationary vector v of M satisfies v M = v or, equivalently v M = 0 (.2) with M = M I where I denotes the identity matrix. Using v we can immediately derive the average payoffs per round, π X and π Y, in an infinitely iterated game for players X and Y. All we need is to rewrite the payoff matrix for each player as a vector: P X = (R, S, T, P ) and P Y = (R, T, S, P ), respectively, to obtain π X = v P X v π Y = v P Y v. (.4) Note that the division by v, with = (,,, ), is only necessary in case v was not normalized, i.e. its elements v i did not sum up to. In the following we now set out to establish an interesting relationship between determinants derived from M and its stationary distribution v with rather unexpected consequences. For the adjugate of the matrix M, we have adj(m )M = det(m )I = 0 (.5) where the first equality reflects a property of the adjugate matrix, whereas the second equality follows because M is singular. Note that each element a ij of adj(a) is given by a ij = ( ) i+j m ji where m ij is the (i, j)- minor of A, i.e. the determinant of A after removing row i and column j. Because of Eq. (.2), every row of adj(m ) must be proportional to v and hence we can find v based on determinants of M. Since det(a) is invariant to adding one column of A to another one, we can rewrite M to obtain p R q R p R q R f M = p S q T p S q T f 2 p T q S p T q S f 3. (.6) p P q P p P q P f 4 (.3)

2 2 CHAPTER. AUTOCRATIC STRATEGIES by adding the first column to the second and third one, as well as replacing the last column by an arbitrary vector f, without affecting the derivation of v based on the last row of adj(m ). More specifically, the entries a 4j in the last row of adj(m ) are the determinants of the 3 3 matrices after deleting the last column and row j from M. Although every row of adj(m ) is proportional to v, the last row does not depend on the last column of M and hence is the same for adj(m ) regardless of the choice of f. With this, we have established a link between the stationary distribution v and the determinant of M via the adjugate adj(m ): det(m ) = a 4, f + a 4,2 f 2 + a 4,3 f 3 + a 4,4 f 4 = v f. (.7) The reason why this exercise is worthwhile and will turn out to be rather rewarding is due to the particular form of M. More specifically, note that (i) the second column of M is under the sole control of player X; (ii) the third column is similarly under the sole control of player Y ; and, finally, (iii) the last column is an arbitrary vector f that ends up in the dot-product with the stationary vector v. For convenience let us define a new function D(p, q, f) := det(m ), which also highlights the fact that M is a function of the strategies, p and q, of the two players X and Y, as well as the arbitrary vector f. Since f is arbitrary, we can actually set it to the payoff vectors P X and P Y to obtain the average payoffs for each player: π X = v P X v = D(p, q, P X) D(p, q, ) π Y = v P Y v = D(p, q, P Y) D(p, q, ). (.9) Because π X and π Y are linear in P X and P Y, respectively, we can also form a linear combination of the payoffs: απ X + βπ Y + γ = D(p, q, αp X + βp Y + γ) D(p, q, ) (.8) (.0) for some α, β, γ R. Interestingly and notably after over 50 years of research in game theory and the prisoner s dilemma in particular, the rather technical Eq. (.0) caused quite a stir in the scientific community because it enables players to unilaterally exert an unexpected level of control over iterated interactions. This happens because either player can unilaterally set the right-hand-side of Eq. (.0) to zero. Player X achieves this by choosing a strategy p R = αr + βr + γ + p S = αs + βt + γ + p T = αt + βs + γ p P = αp + βp + γ (.) (.2) (.3) (.4) for suitable α, β, γ R such that p i [0, ] and player Y can accomplish the same feat by choosing his strategy q analogously. Either case yields απ X + βπ Y + γ = 0 (.5) and hence either player can unilaterally enforce a linear relationship between his or her payoff and the one of the opponent. Because these strategies set a determinant in Eq. (.0) to zero, this class of strategies was termed zero-determinant strategies or ZD-strategies, for short. The level of control that these zero-determinant strategies offer was previously thought to be impossible. The essential feature of zero-determinant strategies is not so much the technical aspect that they render a particular determinant in Eq. (.0) zero but rather that they unilaterally enforce the linear payoff relation in Eq. (.5). In order to emphasize the latter, we name those strategies in the following autocratic strategies because players adopting such strategies gain an unprecedented degree of control over interactions.

3 .2. MORE TRADITIONAL DERIVATION 3.. EXAMPLES Let us now explore different particularly interesting scenarios where player X adopts an autocratic strategy. SET OPPONENT S PAYOFF The first interesting case arises when setting α = 0, β 0 and hence π Y = γ/β. The corresponding strategy enables player X to fix the payoff of player Y completely independent of the strategy of player Y. As we will see below, this even remains true if player Y has an arbitrarily long memory and employs more sophisticated strategies than just based on the previous interaction as for memory-one strategies. However, player X is not completely free in setting her opponent s payoff because p needs to be a probabilistic strategy, i.e. p i [0, ]. As a result, P π Y R must hold. These strategies are traditionally termed equalizers because they result in equal payoffs regardless of player Y s actions. Exercise: Show that P π Y R. In order to see this, note that from p R follows βr + γ 0 and similarly from p P 0 requires βp + γ 0. Since R P, without loss of generality, we establish β 0 and hence, using the same two inequalities, we find R γ/β and P γ/β or P π Y R, as required. SET OWN PAYOFF Choosing α 0, β = 0 allows player X to set her own score to π X = γ/α, or, does it? Exercise: () Find range of possible π X. (2) Does it depend on the game, i.e. ranking of R, S, T, P? (3) Discuss features of the resulting strategies. EXTORTIONERS.2 MORE TRADITIONAL DERIVATION The discovery of autocratic strategies by? makes an impressive example of how the unbiased view of outsiders can result in original discoveries and spark novel lines of research. This also explains their ingenious but rather unusual approach to a game theoretical problem. For this reason the more traditional and direct approach to autocratic strategies followed only afterwards by?. Alternate and more direct way of showing that autocratic strategies enforce a linear relation between the payoffs of player X and Y Definition.. An autocratic strategy p for player X in infinitely iterated 2 2 games is given by p R = αr + βr + γ + p S = αs + βt + γ + p T = αt + βs + γ p P = αp + βp + γ (.6) (.7) (.8) (.9) for some α, β, γ R such that p i [0, ]. Theorem.. If player X uses an autocratic strategy then απ X + βπ Y + γ = 0 (.20) where π X and π Y denote the average payoff per round to players X and Y, respectively, in the limit of infinitely many rounds.

4 4 CHAPTER. AUTOCRATIC STRATEGIES Proof. For the proof we need to first introduce some notation. Let π X (n) and π Y (n) denote the payoffs to player X and Y in round n; s i (n) the probability that player X experiences outcome i {R, S, T, P }; and q i (n) the conditional probability that player Y plays C in round n + given outcome i in round n. With this we can write the probability that player X cooperates in round n + as p C (n + ) = s R (n + ) + s S (n + ) = s(n) p, (.2) with s(n) = (s R (n), s S (n), s T (n), s P (n)) and p the strategy of player X. The second equality in Eq. (.2) follows from s R (n + ) = s R (n)p R q R (n) + s S (n)p S q T (n) + s T (n)p T q S (n) + s P (n)p P q P (n) (.22) s S (n + ) = s R (n)p R ( q R (n)) + s S (n)p S ( q T (n)) + s T (n)p T ( q S (n)) + s P (n)p P ( q P (n)). (.23) Note that when summing s R (n + ) + s S (n + ) all terms involving q i (n) cancel and a simple dot-product remains. This is the reason that no assumptions regarding the strategy of player Y were necessary. If player X uses an autocratic strategy, see Eq. (.6), we obtain p C (n + ) = s(n) (αr + βr + γ +, αs + βt + γ +, αt + βs + γ, αp + βp + γ) (.24) = s(n) (αp X + βp Y + γ + g), (.25) with = (,,, ) and g = (,, 0, 0). Note that g represents a stubborn strategy that ignores the opponent s moves and continues with whatever it started with. Let us now consider w(n) := p C (n + ) p C (n) = s(n) p s R (n) s S (n) (.26) = s(n) (αp X + βp Y + γ) (.27) and determine the average of w(n) in the limit of infinitely many rounds. The left-hand-side is straight forward and we obtain LHS: lim N N w(n) = lim N N (p C(N) p C (0)) = 0 (.28) because p C (n) is bounded. For the right-hand-side we get RHS: lim N N s(n) (αp X + βp Y + γ) = (αp X + βp Y + γ) lim N N s(n) (.29) but the limit is just the (normalized) stationary probability distribution v for the four outcomes R, S, T, P with v =. Thus, we find that (αp X + βp Y + γ) v = απ X + βπ Y + γ, (.30) which equals zero to match the left-hand-side. Note that if player Y also adopts a memory-one strategy then v is simply the stationary state of the Markov chain given by the transition matrix, Eq. (.).

5 .3. GENERALIZATIONS 5.3 GENERALIZATIONS.3. DISCOUNTED INTERACTIONS.3.2 ARBITRARY NUMBER OF PLAYERS.3.3 ARBITRARY STRATEGIES

16.1 L.P. Duality Applied to the Minimax Theorem

16.1 L.P. Duality Applied to the Minimax Theorem CS787: Advanced Algorithms Scribe: David Malec and Xiaoyong Chai Lecturer: Shuchi Chawla Topic: Minimax Theorem and Semi-Definite Programming Date: October 22 2007 In this lecture, we first conclude our

More information

Markov Processes Hamid R. Rabiee

Markov Processes Hamid R. Rabiee Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete

More information

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Mostafa D. Awheda Department of Systems and Computer Engineering Carleton University Ottawa, Canada KS 5B6 Email: mawheda@sce.carleton.ca

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

3 Matrix Algebra. 3.1 Operations on matrices

3 Matrix Algebra. 3.1 Operations on matrices 3 Matrix Algebra A matrix is a rectangular array of numbers; it is of size m n if it has m rows and n columns. A 1 n matrix is a row vector; an m 1 matrix is a column vector. For example: 1 5 3 5 3 5 8

More information

Prisoner s Dilemma. Veronica Ciocanel. February 25, 2013

Prisoner s Dilemma. Veronica Ciocanel. February 25, 2013 n-person February 25, 2013 n-person Table of contents 1 Equations 5.4, 5.6 2 3 Types of dilemmas 4 n-person n-person GRIM, GRIM, ALLD Useful to think of equations 5.4 and 5.6 in terms of cooperation and

More information

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem Algorithmic Game Theory and Applications Lecture 4: 2-player zero-sum games, and the Minimax Theorem Kousha Etessami 2-person zero-sum games A finite 2-person zero-sum (2p-zs) strategic game Γ, is a strategic

More information

Extortion and Evolution in the Iterated Prisoner s Dilemma

Extortion and Evolution in the Iterated Prisoner s Dilemma Extortion and Evolution in the Iterated Prisoner s Dilemma Michael Earnest Nicholas Pippenger, Advisor Francis Su, Reader May, 2013 Department of Mathematics Copyright c 2013 Michael Earnest. The author

More information

arxiv: v1 [cs.gt] 18 Dec 2017

arxiv: v1 [cs.gt] 18 Dec 2017 Invincible Strategies of Iterated Prisoner s Dilemma Shiheng Wang and Fangzhen Lin Department of Computer Science The Hong Kong University of Science and Technology Clear Water Bay,Kowloon,Hong Kong arxiv:1712.06488v1

More information

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true? . Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in

More information

Chapter 3. Determinants and Eigenvalues

Chapter 3. Determinants and Eigenvalues Chapter 3. Determinants and Eigenvalues 3.1. Determinants With each square matrix we can associate a real number called the determinant of the matrix. Determinants have important applications to the theory

More information

Linear Programming Methods

Linear Programming Methods Chapter 11 Linear Programming Methods 1 In this chapter we consider the linear programming approach to dynamic programming. First, Bellman s equation can be reformulated as a linear program whose solution

More information

, some directions, namely, the directions of the 1. and

, some directions, namely, the directions of the 1. and 11. Eigenvalues and eigenvectors We have seen in the last chapter: for the centroaffine mapping, some directions, namely, the directions of the 1 coordinate axes: and 0 0, are distinguished 1 among all

More information

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006. Markov Chains As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006 1 Introduction A (finite) Markov chain is a process with a finite number of states (or outcomes, or

More information

1 Matrices and Systems of Linear Equations. a 1n a 2n

1 Matrices and Systems of Linear Equations. a 1n a 2n March 31, 2013 16-1 16. Systems of Linear Equations 1 Matrices and Systems of Linear Equations An m n matrix is an array A = (a ij ) of the form a 11 a 21 a m1 a 1n a 2n... a mn where each a ij is a real

More information

ENGI 9420 Lecture Notes 2 - Matrix Algebra Page Matrix operations can render the solution of a linear system much more efficient.

ENGI 9420 Lecture Notes 2 - Matrix Algebra Page Matrix operations can render the solution of a linear system much more efficient. ENGI 940 Lecture Notes - Matrix Algebra Page.0. Matrix Algebra A linear system of m equations in n unknowns, a x + a x + + a x b (where the a ij and i n n a x + a x + + a x b n n a x + a x + + a x b m

More information

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University SEQUENTIAL DATA So far, when thinking

More information

Example I: Capital Accumulation

Example I: Capital Accumulation 1 Example I: Capital Accumulation Time t = 0, 1,..., T < Output y, initial output y 0 Fraction of output invested a, capital k = ay Transition (production function) y = g(k) = g(ay) Reward (utility of

More information

Computational Problems Related to Graph Structures in Evolution

Computational Problems Related to Graph Structures in Evolution BACHELOR THESIS Štěpán Šimsa Computational Problems Related to Graph Structures in Evolution Department of Applied Mathematics Supervisor of the bachelor thesis: Study programme: Study branch: Prof. Krishnendu

More information

1 Matrices and Systems of Linear Equations

1 Matrices and Systems of Linear Equations March 3, 203 6-6. Systems of Linear Equations Matrices and Systems of Linear Equations An m n matrix is an array A = a ij of the form a a n a 2 a 2n... a m a mn where each a ij is a real or complex number.

More information

Elementary Linear Algebra

Elementary Linear Algebra Matrices J MUSCAT Elementary Linear Algebra Matrices Definition Dr J Muscat 2002 A matrix is a rectangular array of numbers, arranged in rows and columns a a 2 a 3 a n a 2 a 22 a 23 a 2n A = a m a mn We

More information

Linear Algebra: Lecture Notes. Dr Rachel Quinlan School of Mathematics, Statistics and Applied Mathematics NUI Galway

Linear Algebra: Lecture Notes. Dr Rachel Quinlan School of Mathematics, Statistics and Applied Mathematics NUI Galway Linear Algebra: Lecture Notes Dr Rachel Quinlan School of Mathematics, Statistics and Applied Mathematics NUI Galway November 6, 23 Contents Systems of Linear Equations 2 Introduction 2 2 Elementary Row

More information

a i,1 a i,j a i,m be vectors with positive entries. The linear programming problem associated to A, b and c is to find all vectors sa b

a i,1 a i,j a i,m be vectors with positive entries. The linear programming problem associated to A, b and c is to find all vectors sa b LINEAR PROGRAMMING PROBLEMS MATH 210 NOTES 1. Statement of linear programming problems Suppose n, m 1 are integers and that A = a 1,1... a 1,m a i,1 a i,j a i,m a n,1... a n,m is an n m matrix of positive

More information

Markov Model. Model representing the different resident states of a system, and the transitions between the different states

Markov Model. Model representing the different resident states of a system, and the transitions between the different states Markov Model Model representing the different resident states of a system, and the transitions between the different states (applicable to repairable, as well as non-repairable systems) System behavior

More information

Asymmetric Social Norms

Asymmetric Social Norms Asymmetric Social Norms Gabriele Camera Chapman University University of Basel Alessandro Gioffré Goethe University December 8, 2016 Abstract Studies of cooperation in infinitely repeated matching games

More information

Game Theory, Evolutionary Dynamics, and Multi-Agent Learning. Prof. Nicola Gatti

Game Theory, Evolutionary Dynamics, and Multi-Agent Learning. Prof. Nicola Gatti Game Theory, Evolutionary Dynamics, and Multi-Agent Learning Prof. Nicola Gatti (nicola.gatti@polimi.it) Game theory Game theory: basics Normal form Players Actions Outcomes Utilities Strategies Solutions

More information

Algebraic Methods in Combinatorics

Algebraic Methods in Combinatorics Algebraic Methods in Combinatorics Po-Shen Loh 27 June 2008 1 Warm-up 1. (A result of Bourbaki on finite geometries, from Răzvan) Let X be a finite set, and let F be a family of distinct proper subsets

More information

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms

Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Artificial Intelligence Review manuscript No. (will be inserted by the editor) Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Mostafa D. Awheda Howard M. Schwartz Received:

More information

Selfishness vs Altruism vs Balance

Selfishness vs Altruism vs Balance Selfishness vs Altruism vs Balance Pradeep Dubey and Yair Tauman 18 April 2017 Abstract We give examples of strategic interaction which are beneficial for players who follow a "middle path" of balance

More information

ECON 186 Class Notes: Linear Algebra

ECON 186 Class Notes: Linear Algebra ECON 86 Class Notes: Linear Algebra Jijian Fan Jijian Fan ECON 86 / 27 Singularity and Rank As discussed previously, squareness is a necessary condition for a matrix to be nonsingular (have an inverse).

More information

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0. Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the

More information

Google PageRank. Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano

Google PageRank. Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano Google PageRank Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano fricci@unibz.it 1 Content p Linear Algebra p Matrices p Eigenvalues and eigenvectors p Markov chains p Google

More information

MATH36001 Perron Frobenius Theory 2015

MATH36001 Perron Frobenius Theory 2015 MATH361 Perron Frobenius Theory 215 In addition to saying something useful, the Perron Frobenius theory is elegant. It is a testament to the fact that beautiful mathematics eventually tends to be useful,

More information

Section 3.9. Matrix Norm

Section 3.9. Matrix Norm 3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

Game Theory and Algorithms Lecture 2: Nash Equilibria and Examples

Game Theory and Algorithms Lecture 2: Nash Equilibria and Examples Game Theory and Algorithms Lecture 2: Nash Equilibria and Examples February 24, 2011 Summary: We introduce the Nash Equilibrium: an outcome (action profile) which is stable in the sense that no player

More information

3.3.3 Illustration: Infinitely repeated Cournot duopoly.

3.3.3 Illustration: Infinitely repeated Cournot duopoly. will begin next period less effective in deterring a deviation this period. Nonetheless, players can do better than just repeat the Nash equilibrium of the constituent game. 3.3.3 Illustration: Infinitely

More information

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property Chapter 1: and Markov chains Stochastic processes We study stochastic processes, which are families of random variables describing the evolution of a quantity with time. In some situations, we can treat

More information

MDP Preliminaries. Nan Jiang. February 10, 2019

MDP Preliminaries. Nan Jiang. February 10, 2019 MDP Preliminaries Nan Jiang February 10, 2019 1 Markov Decision Processes In reinforcement learning, the interactions between the agent and the environment are often described by a Markov Decision Process

More information

Notes on Linear Algebra and Matrix Theory

Notes on Linear Algebra and Matrix Theory Massimo Franceschet featuring Enrico Bozzo Scalar product The scalar product (a.k.a. dot product or inner product) of two real vectors x = (x 1,..., x n ) and y = (y 1,..., y n ) is not a vector but a

More information

0.1 Naive formulation of PageRank

0.1 Naive formulation of PageRank PageRank is a ranking system designed to find the best pages on the web. A webpage is considered good if it is endorsed (i.e. linked to) by other good webpages. The more webpages link to it, and the more

More information

Economics 209A Theory and Application of Non-Cooperative Games (Fall 2013) Extensive games with perfect information OR6and7,FT3,4and11

Economics 209A Theory and Application of Non-Cooperative Games (Fall 2013) Extensive games with perfect information OR6and7,FT3,4and11 Economics 209A Theory and Application of Non-Cooperative Games (Fall 2013) Extensive games with perfect information OR6and7,FT3,4and11 Perfect information A finite extensive game with perfect information

More information

Optimism in the Face of Uncertainty Should be Refutable

Optimism in the Face of Uncertainty Should be Refutable Optimism in the Face of Uncertainty Should be Refutable Ronald ORTNER Montanuniversität Leoben Department Mathematik und Informationstechnolgie Franz-Josef-Strasse 18, 8700 Leoben, Austria, Phone number:

More information

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02)

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02) Linear Algebra (part ) : Matrices and Systems of Linear Equations (by Evan Dummit, 206, v 202) Contents 2 Matrices and Systems of Linear Equations 2 Systems of Linear Equations 2 Elimination, Matrix Formulation

More information

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i := 2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]

More information

MATH 56A SPRING 2008 STOCHASTIC PROCESSES

MATH 56A SPRING 2008 STOCHASTIC PROCESSES MATH 56A SPRING 008 STOCHASTIC PROCESSES KIYOSHI IGUSA Contents 4. Optimal Stopping Time 95 4.1. Definitions 95 4.. The basic problem 95 4.3. Solutions to basic problem 97 4.4. Cost functions 101 4.5.

More information

arxiv: v2 [q-bio.pe] 23 Nov 2017

arxiv: v2 [q-bio.pe] 23 Nov 2017 Zero-determinant strategies in finitely repeated games Genki Ichinose and Naoki Masuda 2 arxiv:707.03097v2 [q-bio.pe] 23 Nov 207 Department of Mathematical and Systems Engineering, Shizuoka University,

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning March May, 2013 Schedule Update Introduction 03/13/2015 (10:15-12:15) Sala conferenze MDPs 03/18/2015 (10:15-12:15) Sala conferenze Solving MDPs 03/20/2015 (10:15-12:15) Aula Alpha

More information

A Polynomial-time Nash Equilibrium Algorithm for Repeated Games

A Polynomial-time Nash Equilibrium Algorithm for Repeated Games A Polynomial-time Nash Equilibrium Algorithm for Repeated Games Michael L. Littman mlittman@cs.rutgers.edu Rutgers University Peter Stone pstone@cs.utexas.edu The University of Texas at Austin Main Result

More information

Daily Update. Math 290: Elementary Linear Algebra Fall 2018

Daily Update. Math 290: Elementary Linear Algebra Fall 2018 Daily Update Math 90: Elementary Linear Algebra Fall 08 Lecture 7: Tuesday, December 4 After reviewing the definitions of a linear transformation, and the kernel and range of a linear transformation, we

More information

Zero sum games Proving the vn theorem. Zero sum games. Roberto Lucchetti. Politecnico di Milano

Zero sum games Proving the vn theorem. Zero sum games. Roberto Lucchetti. Politecnico di Milano Politecnico di Milano General form Definition A two player zero sum game in strategic form is the triplet (X, Y, f : X Y R) f (x, y) is what Pl1 gets from Pl2, when they play x, y respectively. Thus g

More information

Designing Information Devices and Systems I Spring 2016 Elad Alon, Babak Ayazifar Homework 12

Designing Information Devices and Systems I Spring 2016 Elad Alon, Babak Ayazifar Homework 12 EECS 6A Designing Information Devices and Systems I Spring 06 Elad Alon, Babak Ayazifar Homework This homework is due April 6, 06, at Noon Homework process and study group Who else did you work with on

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

Homework 4 Solutions

Homework 4 Solutions CS 174: Combinatorics and Discrete Probability Fall 01 Homework 4 Solutions Problem 1. (Exercise 3.4 from MU 5 points) Recall the randomized algorithm discussed in class for finding the median of a set

More information

Stochastic Shortest Path Problems

Stochastic Shortest Path Problems Chapter 8 Stochastic Shortest Path Problems 1 In this chapter, we study a stochastic version of the shortest path problem of chapter 2, where only probabilities of transitions along different arcs can

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

DIMACS Technical Report March Game Seki 1

DIMACS Technical Report March Game Seki 1 DIMACS Technical Report 2007-05 March 2007 Game Seki 1 by Diogo V. Andrade RUTCOR, Rutgers University 640 Bartholomew Road Piscataway, NJ 08854-8003 dandrade@rutcor.rutgers.edu Vladimir A. Gurvich RUTCOR,

More information

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer. Lecture 5, January 27, 2006

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer. Lecture 5, January 27, 2006 Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer Lecture 5, January 27, 2006 Solved Homework (Homework for grading is also due today) We are told

More information

THE EIGENVALUE PROBLEM

THE EIGENVALUE PROBLEM THE EIGENVALUE PROBLEM Let A be an n n square matrix. If there is a number λ and a column vector v 0 for which Av = λv then we say λ is an eigenvalue of A and v is an associated eigenvector. Note that

More information

Multiagent Learning Using a Variable Learning Rate

Multiagent Learning Using a Variable Learning Rate Multiagent Learning Using a Variable Learning Rate Michael Bowling, Manuela Veloso Computer Science Department, Carnegie Mellon University, Pittsburgh, PA 15213-3890 Abstract Learning to act in a multiagent

More information

Game theory Lecture 19. Dynamic games. Game theory

Game theory Lecture 19. Dynamic games. Game theory Lecture 9. Dynamic games . Introduction Definition. A dynamic game is a game Γ =< N, x, {U i } n i=, {H i } n i= >, where N = {, 2,..., n} denotes the set of players, x (t) = f (x, u,..., u n, t), x(0)

More information

Lecture Notes in Linear Algebra

Lecture Notes in Linear Algebra Lecture Notes in Linear Algebra Dr. Abdullah Al-Azemi Mathematics Department Kuwait University February 4, 2017 Contents 1 Linear Equations and Matrices 1 1.2 Matrices............................................

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Mixed Nash Equilibria

Mixed Nash Equilibria lgorithmic Game Theory, Summer 2017 Mixed Nash Equilibria Lecture 2 (5 pages) Instructor: Thomas Kesselheim In this lecture, we introduce the general framework of games. Congestion games, as introduced

More information

18.600: Lecture 32 Markov Chains

18.600: Lecture 32 Markov Chains 18.600: Lecture 32 Markov Chains Scott Sheffield MIT Outline Markov chains Examples Ergodicity and stationarity Outline Markov chains Examples Ergodicity and stationarity Markov chains Consider a sequence

More information

13 More on free abelian groups

13 More on free abelian groups 13 More on free abelian groups Recall. G is a free abelian group if G = i I Z for some set I. 13.1 Definition. Let G be an abelian group. A set B G is a basis of G if B generates G if for some x 1,...x

More information

Belief-based Learning

Belief-based Learning Belief-based Learning Algorithmic Game Theory Marcello Restelli Lecture Outline Introdutcion to multi-agent learning Belief-based learning Cournot adjustment Fictitious play Bayesian learning Equilibrium

More information

Markov Processes. Stochastic process. Markov process

Markov Processes. Stochastic process. Markov process Markov Processes Stochastic process movement through a series of well-defined states in a way that involves some element of randomness for our purposes, states are microstates in the governing ensemble

More information

M340(921) Solutions Practice Problems (c) 2013, Philip D Loewen

M340(921) Solutions Practice Problems (c) 2013, Philip D Loewen M340(921) Solutions Practice Problems (c) 2013, Philip D Loewen 1. Consider a zero-sum game between Claude and Rachel in which the following matrix shows Claude s winnings: C 1 C 2 C 3 R 1 4 2 5 R G =

More information

Lecture Notes on Game Theory

Lecture Notes on Game Theory Lecture Notes on Game Theory Levent Koçkesen Strategic Form Games In this part we will analyze games in which the players choose their actions simultaneously (or without the knowledge of other players

More information

Stochastic Processes

Stochastic Processes Stochastic Processes 8.445 MIT, fall 20 Mid Term Exam Solutions October 27, 20 Your Name: Alberto De Sole Exercise Max Grade Grade 5 5 2 5 5 3 5 5 4 5 5 5 5 5 6 5 5 Total 30 30 Problem :. True / False

More information

Effi ciency in Repeated Games with Local Monitoring

Effi ciency in Repeated Games with Local Monitoring Effi ciency in Repeated Games with Local Monitoring Francesco Nava and Michele Piccione London School of Economics April 2013 Nava & Piccione (LSE) Local Monitoring April 2013 1 / 68 General Motivation

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

Problems on Evolutionary dynamics

Problems on Evolutionary dynamics Problems on Evolutionary dynamics Doctoral Programme in Physics José A. Cuesta Lausanne, June 10 13, 2014 Replication 1. Consider the Galton-Watson process defined by the offspring distribution p 0 =

More information

Reinforcement Learning

Reinforcement Learning 1 Reinforcement Learning Chris Watkins Department of Computer Science Royal Holloway, University of London July 27, 2015 2 Plan 1 Why reinforcement learning? Where does this theory come from? Markov decision

More information

STOCHASTIC PROCESSES Basic notions

STOCHASTIC PROCESSES Basic notions J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving

More information

Complexity in social dynamics : from the. micro to the macro. Lecture 4. Franco Bagnoli. Lecture 4. Namur 7-18/4/2008

Complexity in social dynamics : from the. micro to the macro. Lecture 4. Franco Bagnoli. Lecture 4. Namur 7-18/4/2008 Complexity in Namur 7-18/4/2008 Outline 1 Evolutionary models. 2 Fitness landscapes. 3 Game theory. 4 Iterated games. Prisoner dilemma. 5 Finite populations. Evolutionary dynamics The word evolution is

More information

Game Theory and Rationality

Game Theory and Rationality April 6, 2015 Notation for Strategic Form Games Definition A strategic form game (or normal form game) is defined by 1 The set of players i = {1,..., N} 2 The (usually finite) set of actions A i for each

More information

Designing Information Devices and Systems I Spring 2018 Homework 5

Designing Information Devices and Systems I Spring 2018 Homework 5 EECS 16A Designing Information Devices and Systems I Spring 2018 Homework All problems on this homework are practice, however all concepts covered here are fair game for the exam. 1. Sports Rank Every

More information

1 Markov decision processes

1 Markov decision processes 2.997 Decision-Making in Large-Scale Systems February 4 MI, Spring 2004 Handout #1 Lecture Note 1 1 Markov decision processes In this class we will study discrete-time stochastic systems. We can describe

More information

The Robustness of Zero-Determinant Strategies in Iterated Prisoner s Dilemma Games

The Robustness of Zero-Determinant Strategies in Iterated Prisoner s Dilemma Games The Robustness of Zero-Determinant Strategies in Iterated Prisoner s Dilemma Games Jing Chen,a and Aleksey Zinger b a Department of Computer Science, Stony Brook University, Stony Brook, NY 11794, USA

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Jim Lambers MAT 610 Summer Session Lecture 2 Notes Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the

More information

Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University)

Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University) Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University) The authors explain how the NCut algorithm for graph bisection

More information

1 Determinants. 1.1 Determinant

1 Determinants. 1.1 Determinant 1 Determinants [SB], Chapter 9, p.188-196. [SB], Chapter 26, p.719-739. Bellow w ll study the central question: which additional conditions must satisfy a quadratic matrix A to be invertible, that is to

More information

arxiv:math/ v1 [math.oc] 29 Jun 2004

arxiv:math/ v1 [math.oc] 29 Jun 2004 Putting the Prisoner s Dilemma in Context L. A. Khodarinova and J. N. Webb Magnetic Resonance Centre, School of Physics and Astronomy, University of Nottingham, Nottingham, England NG7 RD, e-mail: LarisaKhodarinova@hotmail.com

More information

Notes on Row Reduction

Notes on Row Reduction Notes on Row Reduction Francis J. Narcowich Department of Mathematics Texas A&M University September The Row-Reduction Algorithm The row-reduced form of a matrix contains a great deal of information, both

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Introduction to Game Theory

Introduction to Game Theory COMP323 Introduction to Computational Game Theory Introduction to Game Theory Paul G. Spirakis Department of Computer Science University of Liverpool Paul G. Spirakis (U. Liverpool) Introduction to Game

More information

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued) 1 A linear system of equations of the form Sections 75, 78 & 81 a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2 a m1 x 1 + a m2 x 2 + + a mn x n = b m can be written in matrix

More information

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T. Notes on singular value decomposition for Math 54 Recall that if A is a symmetric n n matrix, then A has real eigenvalues λ 1,, λ n (possibly repeated), and R n has an orthonormal basis v 1,, v n, where

More information

Newtonian Mechanics. Chapter Classical space-time

Newtonian Mechanics. Chapter Classical space-time Chapter 1 Newtonian Mechanics In these notes classical mechanics will be viewed as a mathematical model for the description of physical systems consisting of a certain (generally finite) number of particles

More information

The Iterated Prisoner s Dilemma: Good Strategies and Their Dynamics

The Iterated Prisoner s Dilemma: Good Strategies and Their Dynamics arxiv:1211.0969v3 [math.ds] 25 Apr 2017 The Iterated Prisoner s Dilemma: Good Strategies and Their Dynamics Ethan Akin Mathematics Department The City College 137 Street and Convent Avenue New York City,

More information

1111: Linear Algebra I

1111: Linear Algebra I 1111: Linear Algebra I Dr. Vladimir Dotsenko (Vlad) Michaelmas Term 2015 Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Michaelmas Term 2015 1 / 10 Row expansion of the determinant Our next goal is

More information

MATRICES AND MATRIX OPERATIONS

MATRICES AND MATRIX OPERATIONS SIZE OF THE MATRIX is defined by number of rows and columns in the matrix. For the matrix that have m rows and n columns we say the size of the matrix is m x n. If matrix have the same number of rows (n)

More information

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov Chains, Stochastic Processes, and Matrix Decompositions Markov Chains, Stochastic Processes, and Matrix Decompositions 5 May 2014 Outline 1 Markov Chains Outline 1 Markov Chains 2 Introduction Perron-Frobenius Matrix Decompositions and Markov Chains Spectral

More information

Linear Algebra Review

Linear Algebra Review January 29, 2013 Table of contents Metrics Metric Given a space X, then d : X X R + 0 and z in X if: d(x, y) = 0 is equivalent to x = y d(x, y) = d(y, x) d(x, y) d(x, z) + d(z, y) is a metric is for all

More information

Topics on Game Theory

Topics on Game Theory THESIS FOR THE DEGREE OF LICENTIATE OF PHILOSOPHY Topics on Game Theory Emilio Bergroth Department of Mathematical Sciences Division of Mathematical Statistics Chalmers University of Technology and University

More information

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < + Random Walks: WEEK 2 Recurrence and transience Consider the event {X n = i for some n > 0} by which we mean {X = i}or{x 2 = i,x i}or{x 3 = i,x 2 i,x i},. Definition.. A state i S is recurrent if P(X n

More information

Two-player games : a reduction

Two-player games : a reduction Two-player games : a reduction Nicolas Vieille December 12, 2001 The goal of this chapter, together with the next one, is to give a detailed overview of the proof of the following result. Theorem 1 Let

More information