Oblivious Equilibrium for Large-Scale Stochastic Games with Unbounded Costs
|
|
- Willa Melton
- 5 years ago
- Views:
Transcription
1 Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 008 Oblivious Equilibrium for Large-Scale Stochastic Games with Unbounded Costs Sachin Adlakha, Ramesh Johari, Gabriel Weintraub, and Andrea Goldsmith Abstract We stud stochastic dnamic games with a large number of plaers, where plaers are coupled via their cost functions. A standard solution concept for stochastic games is Markov perfect equilibrium MPE. In MPE, each plaer s strateg is a function of its own state as well as the state of the other plaers. This makes MPE computationall prohibitive as the number of plaers becomes large. An approximate solution concept called oblivious equilibrium OE was introduced in 1, where each plaer s decision depends onl on its own state and the long-run average state of other plaers. This makes OE computationall more tractable than MPE. It was shown in 1 that, under a set of assumptions, as the number of plaers become large, OE closel approximates MPE. In this paper we relax those assumptions and generalize that result to cases where the cost functions are unbounded. Furthermore, we show that under these relaxed set of assumptions, the OE approximation result can be applied to large population linear quadratic Gaussian LQG games. I. INTRODUCTION In this paper, we stud stochastic games with a large number of plaers. Such games are used to model complex dnamical sstems such as wireless networks 3, 4, industr dnamics with man firms 5, etc. In such games, the plaers tpicall have competing objectives. A common equilibrium notion for such games is Markov perfect equilibrium. In MPE, each plaer minimizes its individual cost b choosing a strateg that is a function of the current state of all the plaers. MPE suffers from at least two drawbacks. First, as an equilibrium concept, it is implausible in settings where man agents interact with each other; MPE requires the agents to be aware of the state evolution of all other agents. Second, MPE can be computationall intractable. MPE is tpicall obtained numericall using dnamic programming DP algorithms 7. Thus, as the number of plaers increases, the cost of computing the strateg and maintaining the state of all the plaers becomes prohibitivel large 8. Several techniques have been proposed in the literature to deal with the complexit of large scale sstems 9, 10, 15. Recentl, a scheme for approximating MPE for such large scale games was proposed in 1, via a solution concept called oblivious equilibrium. In oblivious equilibrium, a plaer optimizes given onl the long-run average statistics of other plaers, rather than the entire instantaneous vector of its S. Adlakha and A. Goldsmith are with the Department of Electrical Engineering, Stanford Universit. adlakha@stanford.edu, andrea@wsl.stanford.edu R. Johari is with the Department of Management Science and Engineering, Stanford Universit. ramesh.johari@stanford.edu G. Weintraub is with the Columbia Business School, Columbia Universit. gweintraub@columbia.edu competitors state. OE resolves both of the difficulties raised above: in OE, a plaer is reacting to far simpler aggregate statistics of the behavior of other plaers. Further, OE computation is significantl simpler than MPE computation, since each plaer onl needs to solve a one-dimensional dnamic program. Under what conditions will OE approximate MPE? When there are a large number of plaers in a game, an individual plaer can optimize based on the average behavior of it competitors provided a the actual behavior of the competitors is close to the average behavior, and b an deviation from the average behavior has small impact on the cost of the individual plaer. It is intuitive to believe that as the number of plaers in a game becomes large, the actual behavior would approach its mean b a law of large numbers effect. To measure the impact of an deviation from the average behavior, the authors in 1 defined a light-tail condition. Informall, this condition implies that the effect of a small perturbation in the instantaneous state of the competitors has a small effect on the cost of a plaer. It is reasonable to expect that under such a condition, if plaers make decisions based onl on the long-run average, the should achieve near-optimal performance. Indeed, it is established in 1 that under a reasonable set of technical conditions including the light-tail condition, OE is a good approximation to MPE for industr dnamic models with man firms; formall, this is called the asmptotic Markov equilibrium AME propert. As presented in 1, the main approximation result is tailored to the class of firm competition models presented there. In 6, using the methods of 1, the authors isolated a set of parsimonious assumptions for a general class of stochastic games, under which OE is a good approximation to MPE. The also stud the case of non-uniform plaers with heterogeneous cost functions. However, the results in 1, 6 are based on the assumption that the cost functions are uniforml bounded over states and actions. This is a restrictive assumption; for example, tpical cost functions used in decentralized control are unbounded e.g., quadratic cost. In this paper, we remove this boundedness assumption on the cost functions; this change necessitates an alternate approach to establish that OE is a good approximation to MPE. We show a general result under this assumption, then appl our result to a class of games including games with quadratic cost and linear dnamics. The latter result is a generalization of a similar result derived b. The rest of the paper is organized as follows. In section II, we outline our model of stochastic game, describe our /08/$ IEEE 5531
2 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 notation and define the oblivious equilibrium. In section III, we introduce the asmptotic Markov equilibrium AME propert, formall define the light-tail condition and introduce our assumptions on the cost function. Section IV details the proof of the main theorem. In section V, we show that linear-quadratic games can be reduced to a special case of our model. Section VI concludes the paper. II. MODEL, DEFINITIONS AND NOTATION We consider an m-plaer stochastic game evolving over discrete time periods with an infinite horizon. The discrete time periods are indexed b non-negative integers t N. The state of a plaer i at time t is denoted b x i,t X, where X is a discrete subset of R. We assume that the state evolution of a plaer i depends onl on its own current state and the action it takes. This can be represented b a conditional probabilit mass function pmf x i,t+1 h x x i,t,a i,t, 1 where a i,t is the action taken b the plaer i at time t. We denote the set of actions available to a plaer b A; we assume this is a discrete subset of Euclidean space R. The single period cost to a plaer i is given as cx i,t,a i,t,x,t. Here x,t is the state of all plaers except plaer i at time t. Note that the cost to plaer i does not depend on the actions taken b other plaers. Furthermore, we assume that the cost function is independent of the identit of other plaers. That is, it onl depends on the current state x i,t of plaer i, the total number m of plaers at an time, and the fraction f m,t, which is the fraction of plaers excluding plaer i, that have their state as. In other words, we can write the cost function as c x i,t,a i,t,f m,t,m, where f m,t can be expressed as f m 1,t m 1 1 {x}. From equation 1 and the definition of the cost function, we note that the plaers are coupled via their cost functions onl. Each plaer i chooses an action a i,t µ m i x i,t,f m,t to minimize its expected present value. Note that the polic µ m i depends on the total number of plaers m because of the underling dependence of the cost function on m. Let µ m be the vector of policies of all plaers, and µ m be the vector of policies of all plaers except plaer i. We define V x,f,m µ m i,µ m to be the expected net present value for plaer i with current state x, if the current aggregate state of plaers other than i is f, given that i follows the polic and the polic vector of plaers other than i is given µ m i b µ m. In particular, we have V x,f,m µ m i,µ m E β t cx i,t,a i,t,f m,t,m x i,0 x,f m,0 f;µm i,µ m, 3 where 0 < β < 1 is the discount factor. Note that the random variables x i,t,f m i,t depend on the polic vector µ m and the state evolution function h. We focus on smmetric Markov perfect equilibrium, where all plaers use the same polic µ m. We thus drop the subscript i in the polic of a plaer i. Let M be the set of all policies available to a plaer. Note that this set also depends on the total number of plaers m. Definition 1 Markov Perfect Equilibrium: The vector of policies µ m is a Markov perfect equilibrium if for all i,x, and f we have inf V x,f,m µ,µ m µ V x,f,m µ m,µ m. M As the number of plaers becomes large, the MPE becomes computationall intractable. This is because the set of all policies grows exponentiall in the number of plaers. However, if the coupling between the plaers is weak, it is possible that the plaers can choose their optimal action based solel on their own state and the average state of the other plaers. We expect that as the number of plaers becomes large, the changes in the plaers states average out such that the state vector f m,t is well approximated b its long run average. Thus, each plaer can find its optimal polic based solel on its own state and the long-run average aggregate state of the other plaers. We therefore restrict attention to policies that are onl a function of the plaer s own state, and an underling constant aggregate distribution of the competitors. Such strategies are referred to as oblivious strategies since the do not take into account the complete state of the competitors at an time. Let us denote µ m as an oblivious polic of a plaer i; we let M denote the set of all oblivious policies available to a plaer. This set also depends on the number of plaers m. Note that if all plaers use oblivious strategies, their states evolve as independent Markov chains. We make the following assumption regarding the Markov chain of each plaer plaing an oblivious polic. Assumption 1: The Markov chain associated with the state evolution of each plaer i plaing an oblivious polic µ m is positive recurrent, and reaches a stationar distribution q m. The stationar distribution depends on the number of plaers m because the oblivious polic depends on m. Let µ m be the vector of oblivious policies for all plaers, µ m i be the oblivious polic for a plaer i, and µ m be the vector of oblivious policies of all plaer except the plaer i. For simplification of analsis, we assume that the initial state of a plaer i is sampled from the stationar distribution q m of its state Markov chain; without this assumption, 553
3 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 the OE approximation holds onl after sufficient mixing of the individual plaers state evolution Markov chains. Given µ m, for a particular plaer i, the long-run average aggregate m state of its competitors is denoted b f, and is defined as f m E f m,t q m 4 Note that, fm is completel determined b the state evolution function h and the oblivious polic µ m. As with the case of smmetric MPE defined above, we assume that plaers use the same oblivious polic denoted b µ m. We thus drop the subscript i in the oblivious polic of a plaer i. We define the oblivious value function Ṽ x,m µ m, µ m to be the expected net present value for a plaer i with current state x, if plaer i follows the oblivious polic µ m, and plaers other than i follow the oblivious polic vector µ m. Specificall, we have Ṽ x,m µ m, µ m E β t m c x i,t,a i,t, f,m x i,0 x; µ m, µ m. Note that the expectation does not depend explicitl on the policies used b plaers other than i; this dependence onl m enters through the long-run average aggregate state f. In particular, the state evolution is onl due to the polic of plaer i. Using the oblivious value function, we define oblivious equilibrium as follows. Definition Oblivious Equilibrium: The vector of policies µ represents an oblivious equilibrium if for all i, we have inf Ṽ x,m µ, µ m µ M Ṽ x,m µ m, µ m, x. In this paper, we do not show the existence of Markov perfect equilibrium or of oblivious equilibrium. We assume that both the equilibrium points exist for the stochastic game under consideration 14. Assumption : Markov perfect equilibrium and oblivious equilibrium exist for the stochastic game under consideration. III. ASYMPTOTIC MARKOV EQUILIBRIUM AND THE LIGHT TAIL As mentioned before, we would like to approximate MPE using OE. To formalize the notion under which OE approximates MPE, we define the asmptotic Markov equilibrium AME propert. Intuitivel, this propert sas that an oblivious polic is approximatel optimal even when compared against Markov policies. Formall, the AME ensures that as number of plaers in the game becomes large, the approximation error between the expected net present value obtained b deviating from the oblivious polic µ m and instead following the optimal non-oblivious polic goes to zero for each state x of the plaer. Definition 3 Asmptotic Markov Equilibrium: We sa that a sequence of oblivious policies µ m possesses the 5 asmptotic Markov equilibrium AME propert if for all x and i, we have lim E V x,f,m µ m, µ m m inf V x,f,m µ, µ m µ 0. M Notice that the expectation here is over f, which denotes the aggregate state of all plaers other than i. MPE requires the error to be zero for all x,f, rather than in expectation; of course, in general, it will not be possible to find a single oblivious polic that satisfies the AME propert for an f. In particular, in OE, actions taken b a plaer will perform poorl if the other plaers state is far from the long-run average aggregate state. Thus, AME implies that the OE polic performs nearl as well as the non-oblivious best polic for those aggregate states of other plaers that occur with high probabilit. In order to establish the AME propert, we make some assumptions on the cost functions. For notational convenience, we drop the subscripts i,t whenever it does not lead to an ambiguit. Note that in oblivious equilibrium, we replace the actual distribution of the opponents state f m,t b its mean m distribution f. In order to measure the distance between the actual distribution of the state and its mean, we define a notion of norm on a distribution. Definition 4 1-g Norm: Given a function g : X 0,, we define the 1 g norm of the distribution f as f f g 6 The 1 g norm is a weighted norm where g is the weight function. Note that this function depends on the actual form of the cost function. For a given function g, the distance between the actual distribution f m m,t and its mean f is given b f m m,t f f m m,t f g. We now formall define the light tail condition. Assumption 3 Light Tail: Given an ǫ > 0, there exists a state value z > 0, such that E E gũm 1 Ũ m >z Ũm g Ũm 1 Ũ m >z Ũm m f f m ǫ, m, 7 ǫ, m. 8 Here Ũm is a random variable distributed according to f m. As mentioned before, g is a weight function for the state. Thus, the light tail assumption requires that the weighted tail probabilit of the competitors goes to zero uniforml over m. Also, note that onl the second condition needs to be checked; a straightforward application of Jensen s inequalit then shows the first condition must hold. However, we retain both inequalities for clarit. Furthermore, if sup g <, then the light tail condition is just an m assumption on the tail of the mean distribution f. In order that the AME propert holds, we would like that the cost functions are close to each other when the actual 5533
4 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 distribution f m m,t is close to its mean distribution f. For this to hold, we impose a uniformit condition on the growth of the cost function. Specificall, we make the following assumption on the cost function. Assumption 4: There exists a constant C such that, for all x,a,f 1,f and m, we have cx,a,f 1,m cx,a,f,m C cx,a,f 1,m f 1 f + H f 1 f Note that here H f denotes a function that is linear in moments of f up to the second moment; i.e., H f cannot include an terms that depend on moments higher than f. Note that the function g in Definition 4 must be chosen such that the assumptions 3 and 4 are simultaneousl satisfied. As we will show in section V, the linear quadratic Gaussian LQG tracking problem discussed in is a special case where the difference in the cost functions can be expressed in the form given above. Furthermore,we will show that for the LQG problem, the function g is a polnomial in the state. The next assumption is on the set of policies M. We will restrict attention to sequences µ m, µ m that satisf the following assumption. Assumption 5: Given a sequence µ m, µ m, we assume that: sup E β t c x i,t,a i,t, m m f,m x i,0 x,f m,0 f;µm, µ m < In contrast to earlier work on oblivious equilibrium 1 6, our assumptions are significantl different. Specificall, we do not assume that the cost functions are uniforml bounded in state x, distribution f, or actions a as was done in the previous work. This lack of uniform bound on the cost function necessitates a significantl different proof technique. However, some of the ideas in the proofs are borrowed from 6. IV. ASYMPTOTIC RESULTS FOR OBLIVIOUS EQUILIBRIUM In this section, we prove the AME propert using a series of technical lemmas. Assumptions 1-5 are kept throughout the remainder of this section. The first lemma shows that under the 1 g norm, the variance of the distribution f m,t goes to zero as the number of plaers become large. Lemma 1: Under the light-tail assumption and if all the plaers use oblivious polic µ m, we have f m m E,t f 0 as m f m,t Proof: We can write m f g f m,t m f. Now, let a small ǫ > 0 be given and let z be such that the light tail condition in equation 7 and 8 is satisfied for the given ǫ. Then, f m m,t f f z max g m m,t f 1,g z + gf m,t + m. g f >z >z Using the identit a + b + c 4a + 4b + 4c, we get that Ef m m,t f 4z E max 1,g z g f m m,t f }{{} + 4E gf m >z,t } {{ } B z m +4E A m z >z m g f } {{ } C z m. Note that term in parenthesis in C z m is independent of t and hence a constant. B the light tail assumption, for sufficientl small ǫ > 0 and sufficientl large z, we have C z m 4ǫ < 4ǫ for all m. Let us now consider the term A z m. We have E f m m,t f 1 m 1 E 1 {x} E 1 {x}, 1 m 1 Var 1 {x}, 1 0 as m. 4m 1 The random variable 1 {x} is a Bernoulli random variable with E 1 {x} q m and Var 1 {x} q m 1 q m 1 4. Thus, for each z, there exists an m such that for m m we have f m m E,t f ǫ 4z g Define m a max z {m }. Then, for m > m a we have A z m < ǫ. Let us now consider the term B z m we have of f m,t B z m 4 m 1 E >z 4 m 1 E >z 9. Using the definition g1 {x} g1 {x} where the interchange of summation is justified since for an given m, the summations are finite. Let us denote 5534
5 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 >z g1 {x } W m. We can thus write B z m as B z m 4 m 1 E W m 4 m 1 Var W m + EW m 10 From equation 4, we know that E 1 {x} fm. From the light tail condition, for chosen z, we have E W m < ǫ for all m. Thus, E < m 1ǫ 11 W m Let us now consider the first term in equation 10. We have Var Var W m W m m 1Var m 1 E < m 1E W m W m W m E W m + m 1ǫ 1 where the first equalit follows from the fact that f m,t is an m m 1 fold convolution of f, since all the opponents are using oblivious polic µ and hence are evolving independentl. The last inequalit is because of the chosen value of z, which gives E W m < ǫ for all m. Note that W m as defined is a random variable that takes the value g with m probabilit f. Thus, E W m g m f < ǫ >z where the last inequalit follows from the light tail condition. Substituting the above equation in equation 1, we get that Var < m 1ǫ + m 1ǫ < m 1ǫ W m 13 Substituting equations 11 and 13 in equation 10, we get that for some m > m b we have B z m 4 < m 1ǫ + m 1 m 1 ǫ 8 < m 1 ǫ + 4ǫ 1ǫ From bounds on A z m and C z m and the above equation, we get that for sufficientl large z and for m > max{m a,m b } and we have Ef m m f 17ǫ. 1,g,t Since ǫ is arbitrar, this proves the lemma. Lemma : For all x and µ m, µ m satisfing assumptions 1-5, we have lim E β t c x i,t,a i,t,f m m,t,m m c x i,t,a i,t, f,m xi,0 x;µ m, µ m Proof: Let us define m i,t c x i,t,a i,t,f m,t,m c 0. m. x i,t,a i,t, f,m m Also denote c m,t c x i,t,a i,t, f,m and Fm,t f m m f. Using assumption 4 on the cost function,t we have E β t m i,t µ m, µ m CE β t c m,t F m,t µ m, µ m + E β t H Fm,t C β t E c m,t F m,t µ m, µ m + β t H E Fm,t where the last equalit follows from monotone convergence theorem. Let us denote the first term of the above equation as T 1 and the second term as T. Using Cauch-Schwarz inequalit we get that T 1 C β t E c 1/ m,t β t E Fm,t 1/ C β t E c 1/ m,t β t E Fm,t 1/ where the last inequalit is due to Hölder s inequalit. Note that we have dropped the polic vector in the conditioning field for notational compactness. B assumption 5, the first term in above equation is bounded. Also, note that E Fm,t is independent of t and hence β t E Fm,t E Fm,t 1 β Substituting the above equation in second term of T 1 and also in T and using lemma 1 we get the desired result. Theorem 3 Main Theorem: Consider a sequence of oblivious equilibrium policies µ m that satisfies assumption 5 with µ m equal to either the oblivious or nonoblivious best response to µ m. Then the AME propert holds. That is, for all i,x, we have lim E V m x,f,m µ m, µ m inf V µ M x,f,m µ, µ m 0. Proof: The proof of the theorem is similar to one given in 1 and is omitted due to space constraints. 5535
6 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 V. LINEAR QUADRATIC SYSTEM In this section, we consider linear quadratic LQ games with man plaers. We show that these games are a special case of the model considered in the previous section. We keep assumptions 1, and 5 throughout the remainder of this section. Our objective in this section is to prove that assumptions 3 and 4 hold for the LQ model, and thus the AME propert holds for the LQ model. Compared to, we consider a discrete time version of the LQ games. Furthermore, for simplicit we assume that all plaers have the same form of cost function, i.e., the are uniform. This assumption can be relaxed, as we discuss in section VI. The state evolution of a plaer 1 i m is given as x i,t+1 Ax i,t + Ba i,t + w i,t, 14 where x i,t Z and a i,t Z. Here A,B > 0. The noise process w i,t is assumed to be independent across time as well as plaers. We also assume that the noise process has zero mean and all its moments are finite. We assume that the single period cost function for a plaer i is separable in its action. That is c x i,t,a i,t,f m,t c 1 x i,t,f m,t + c a i,t where we assume that c a i,t is a quadratic function of a i,t. Note that the cost function does not depend upon the number of plaers plaing the game. Here f m,t is the distribution of m plaers over the state space. Let us define the kth moment of f m,t as γ k f m,t k. We assume that c 1 is a jointl strictl convex, quadratic function of x and γ k, k 1,...,K. Also, for given x and, we have inf f m,t c 1 x,f ǫ 0 > 0. x,f This model is a generalization of the LQG games considered in, where the cost function depends onl on the first moment γ 1 of the distribution f m,t. Furthermore, that result was derived for a specific form of cost function. Each plaer choses an action a i,t µ m x i,t,f m,t to minimize its total expected cost. The expected net present cost for a plaer i is given as V x,f µ m,µ E β t c x i,t,a i,t,f m,t x i,0 x,f m,0 f;µm,µ We assume that each plaer chooses its polic µ m to make its expected net present value V x,f µ m,µ finite. In oblivious equilibrium, each plaer s polic is onl a function of its current state and average aggregate distribution of its competitors. The next lemma shows that for the LQ model described above, the oblivious equilibrium is independent of the number of plaers m Lemma 4: Under assumption, for the LQ model described above, there exists an oblivious equilibrium that is independent of the number of plaers m plaing the game. Proof: Let ˆµ, ˆf be an oblivious equilibrium for the LQ model with m 1 plaers plaing the game. This means that ˆµ is an optimal polic under the cost function cx,a, ˆf and that ˆf is the stationar distribution obtained from the dnamic equation x i,t+1 Ax i,t + Bˆµx i,t, ˆf + w i,t. Now let us assume that the number of plaers changes to m. Consider a plaer i and assume that ever one of its m 1 opponent uses the polic ˆµ. Then f ˆf, and since the cost function does not depend upon the total number of plaers, the optimal polic for plaer i is ˆµ. Since the dnamics are independent of the number of plaers, the stationar distribution for plaer i is ˆf. Thus, ˆµ, ˆf is an oblivious equilibrium for the game with m plaers. To prove that there is no optimalit loss if a plaer uses oblivious equilibrium polic, we first define the weight function g for the LQ model. Definition 5 LQ 1-g norm: For the LQ model described above, let g K. With this weight function, we define f The following lemma verifies the light-tail condition for the LQ model. Lemma 5: For the LQ model, the light tail condition holds for the weight function g K. Proof: Note that f q where q is the stationar probabilit of the Markov chain to be in state and is independent of the number of plaers m. For linear sstems with quadratic cost, we know that the optimal polic is a linear function of the state 11. That is, µ Lx+l 0, where L > 0 and l 0 are constants that ma depend on f. The closed loop sstem thus evolves as x t+1 A BLx t +Bl 0 +w t. From assumption 1, we know that the closed loop sstem is stable; thus we have A BL < 1. We use the Foster-Lapunov stabilit criterion 1 to establish the light tail condition. Using the Lapunov function V x x and the fact that the noise process has finite mean, we can show that E q x is finite. We then use induction to show that the stationar distribution has finite K moments. Specificall, assume that E q x j is finite for j 1,...,p, where p < K. Then, using the Lapunov function V x x p+1 and the fact that the noise process has finite moments, we can show that E q x p+1 is finite. Thus, the stationar distribution has its first K moments finite 13. So for an given ǫ, there exists a state z, such that >z f K K q < ǫ. Hence the lemma is proved. The next lemma relates the absolute difference in the cost function to the 1 g norm difference between the actual distribution f m,t and its average f. Lemma 6: For the LQ model described above, there exist
7 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 constants M 1,M and M 3 such that c x i,t,a i,t,f m,t c x i,t,a i,t, f M 1 c x i,t,a i,t, f f m,t f + M f m,t f + M 3 f m,t f Proof: For simplicit of notation, we drop the subscripts i,t. Also denote Γf γ 1 f,...,γ K f, where γ i f is the ith moment of the distribution f. Then, we have c cx,a,γf c x,a,γ f c 1 x,γf c 1 x,γ f 1 x,γ f + αf f dα α0 α 1 f f α0 x,γ f + αf f df dα 15 where we have used the fundamental theorem of calculus. Consider the second term in the above equation. We have x,γ f + αf f df γ j γ j1 j f γ j1 j j K K 1 K γ j j1 where the last inequalit holds for some K 1 since Z. Substituting the above equation in equation 15, we get c K 1 f f j1 1 α0 x,γ f + αf f dα 16 γ j where we use the fact that f f is independent of α and we also changed the order of integral and the summation. Define a vector z γ 1,γ,...γ K,x,1 and note that c 1 is a strictl convex quadratic function of z. Thus, we can write c 1 x,γ 1,...,γ K z T Qz for some positive definite smmetric matrix Q. Note that Γ f + αf f j γ j + αγ j γ j. Since c 1 is a quadratic function of γ j, the partial derivative of c 1 with respect to γ j is a linear function of γ j and consequentl a linear function of α. Thus, we have x,γ f + αf f γ j Q sj z s + Q js z s + Q K+1,j x + Q j,k+ s1 Q K+1,j x + Q K+1,j x + s1 Q sj + Q js z s + Q j,k+ s1 Q sj + Q js γ s + α γ s γ s s1 + Q j,k+ 17 where we have used the fact that z s γ s + αγ s γ s for s 1,...,K. Now, γ s γ s f s f s f f s f f K f f Substituting above equation in equation 17, and integrating with respect to α we get that 1 x,γ f + αf f α0 γ j dα Dj T f z + K f, for some K. Here we used the notation z γ 1,... γ K, x,1 and D j is a vector of length K+ with coefficients from equation 17. Substituting above equation in equation 16, we get c K 1 f f Dj T z + K K f f j1 K 1 f f D T z + K 1 K K f f 18 Now consider D T z. We show that there exists a constant δ > 0 such that δd T z z T Q z + ǫ 0. Define w δ/q 1/ D. Then, we have z T Q z + ǫ 0 δd T z z T Q z w T Q 1/ z + ǫ 0 + w T w w T w T Q 1/ z w Q 1/ z w + ǫ 0 w T w T z Q 1/ w Q z Q 1/ w + ǫ 0 w T w ǫ 0 w T w 5537
8 47th IEEE CDC, Cancun, Mexico, Dec. 9-11, 008 where the last inequalit follows since the first term is a quadratic form. So we need to ensure that ǫ 0 w T w 0, which implies that δ 4 DT Q 1 D ǫ 0 δ 4ǫ 0 D T Q 1 D If we choose K 3 > 1/δ, then we have D T z K 3 z T Q z + K 3 ǫ 0. Next we show that there exists a K 4 such that z T Q z K 4 zq z, where z γ 1,... γ K,x,1. Note that z T z z T z. We have z T Q z λ Q max z T z λ Q max z T z K 4 λ Q min zt z where the last inequalit is true for a suitable choice of K 4. Note that Q is positive definite matrix so λ Q min > 0. Since K 4 λ Q min zt z K 4 z T Q z, we have that z T Q z K 4 zq z for some value of K 4. Now, D T z K 3 z T Q z + K 3 ǫ 0 K 3 K 4 zq z + K 3 ǫ 0 K 3 K 4 c 1 x, γ 1,... γ K + K 3 ǫ 0 19 Substituting the above equation in equation 18, we prove the lemma. We restrict attention to sequences µ m, µ m that satisf assumption 5. We conjecture that this assumption would be satisfied for the LQ model if the noise process has finite fourth moment. Under this assumption, the AME propert holds for linear sstems with quadratic costs. VI. CONCLUSIONS AND DISCUSSION In this paper, we studied stochastic dnamic games with man plaers, where the plaers are coupled via their cost functions. Similar to 1, we showed that for a certain class of stochastic games, we can approximate MPE b a computationall simpler concept called OE. Previous work done in this area 1, 6 where the AME propert was established for profit maximization assumed a uniform bound on the cost function. In this paper, we extended the notion of OE to a class of games where the cost function is unbounded. The lack of a uniform bound necessitated a new proof technique. We showed that games with linear dnamics and quadratic costs are a special case of the model considered. This generalizes a similar result derived in. In the development throughout the paper, we have assumed that all plaers are uniform, i.e., the have same form of cost function c. The results of this paper can be easil extended to the case where the plaers are non-uniform and their cost functions are drawn from a finite set of tpes. The reader is referred to 6, where a similar development was done albeit for a different set of assumptions. However, the same technique can be used here to extend our model; the details of which are omitted due to space constraints. As mentioned before, the concept of oblivious equilibrium was first introduced in 1, where it was used to establish AME propert for industr dnamic models. Common single period profit functions used in those models are bounded over states. A tpical example is a profit function arising from price competition among firms that face a logit demand sstem generated b consumers with bounded income. For such functions, the AME propert can be established using a slightl different approach. Specificall, assumptions 4 and 5 can be replaced b a uniform bound on the cost function as well as assuming that the cost functions are Gateaux differentiable with respect to f m. For such models, it is possible to show that log cx,a,f 1,m log cx,a,f,m f 1 f where g sup x,a,f m,m log c x,a,f m,m f m The statement of lemma then follows b a similar argument as given in lemma 4 of 6. As a part of our future work, we hope to develop a model that unifies these two approaches. REFERENCES 1 G. Y. Weintraub, C. L. Benkard, and B. Van Ro, Markov Perfect Industr Dnamics with Man Firms, accepted for publication, Econometrica. M. Huang, P. E. Caines and R. P. Malhamé, Large-Population Cost- Coupled LQG Problems With Nonuniform Agents : Individual-Mass Behavior and Decentralized ǫ-nash Equilibria, IEEE Transactions on Automatic Control, Vol. 5, No. 9, pp , M. Huang, P. E. Caines and R. P. Malhamé, Uplink Power Adjustment in Wireless Communication Sstems: A Stochastic Control Analsis, IEEE Transactions on Automatic Control, Vol. 49, No. 10, pp , October S. Ulukus and R. D. Yates, Stochastic Power Control for Cellular Radio Sstems, IEEE Transactions on Communications, Vol. 46, No. 6, pp , R. Ericson and A. Pakes, Markov-perfect industr dnamics: A framework for empirical work, Review of Economic Studies 61, pp. 53-8, V. Abhishek, S. Adlakha, R. Johari and G. Weintraub, Oblivious Equilibrium for General Stochastic Games with Man Plaers, Allerton Conference on Communication, Control and Computing, A. Pakes and P. McGuire, Computing Markov-Perfect Nash Equilibria: Numerical Implications of a Dnamic Differentiated Product Model, RAND Journal of Economics, 54, pp , A. Pakes and P. McGuire, Stochastic Algorithms, Smmetric Markov Perfect Equilibrium, and the Curse of Dimensionalit, Econometrica, 695, pp , A. C. Antoulas and D. C. Sorensen, Approximation of Large-Scale Dnamical Sstems: An Overview, Int. J. Appl. Math. Comput. Sci, Vol. 11, No. 5, pp , D. D. Siljak, Decentralized Control of Complex Sstems, Academic Pr., D. P. Bertsekas, Dnamic Programming and Optimal Control, Athena Scientific, S. P. Men and R. L. Tweedie, Markov Chains and Stochastic Stabilit, Springer-Verlag, J. G. Dai and S. P. Men, Stabilit and Convergence of moments for Multiclass Queuing Networks Via Fluid Limit Models, IEEE Transactions on Automatic Control, Vol. 40, No. 11, pp , M. Huang, R. P. Malhamé, and P. E. Caines, Nash Equilibria for Large-Population Linear Stochastic Sstems of Weakl Coupled Agents, Analsis, Control and Optimization of Complex Dnamical Sstems, pp. 15-5, J. M. Lasr and P. L Lions, Mean Field Games, Japanese Journal of Mathematics,, 9-60,
Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games
Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games Gabriel Y. Weintraub, Lanier Benkard, and Benjamin Van Roy Stanford University {gweintra,lanierb,bvr}@stanford.edu Abstract
More informationMean-field equilibrium: An approximation approach for large dynamic games
Mean-field equilibrium: An approximation approach for large dynamic games Ramesh Johari Stanford University Sachin Adlakha, Gabriel Y. Weintraub, Andrea Goldsmith Single agent dynamic control Two agents:
More informationLecture 3: Computing Markov Perfect Equilibria
Lecture 3: Computing Markov Perfect Equilibria April 22, 2015 1 / 19 Numerical solution: Introduction The Ericson-Pakes framework can generate rich patterns of industry dynamics and firm heterogeneity.
More informationMean field equilibria of multiarmed bandit games
Mean field equilibria of multiarmed bandit games Ramesh Johari Stanford University Joint work with Ramki Gummadi (Stanford University) and Jia Yuan Yu (IBM Research) A look back: SN 2000 2 Overview What
More informationSF2972 Game Theory Problem set on extensive form games
SF2972 Game Theor Problem set on etensive form games Mark Voorneveld There are five eercises, to be handed in at the final lecture (March 0). For a bonus point, please answer all questions; at least half
More information3.0 PROBABILITY, RANDOM VARIABLES AND RANDOM PROCESSES
3.0 PROBABILITY, RANDOM VARIABLES AND RANDOM PROCESSES 3.1 Introduction In this chapter we will review the concepts of probabilit, rom variables rom processes. We begin b reviewing some of the definitions
More informationTrade Liberalization and Pollution in Melitz-Ottaviano Model
Trade Liberalization and Pollution in Melitz-Ottaviano Model Zhou Mohan November 30, 20 Abstract I incorporate Copeland-Talor tpe pollution technolog into Melitz and Ottaviano 2008) s work horse model.
More informationStability Analysis for Linear Systems under State Constraints
Stabilit Analsis for Linear Sstems under State Constraints Haijun Fang Abstract This paper revisits the problem of stabilit analsis for linear sstems under state constraints New and less conservative sufficient
More informationMarkov Decision Processes and Dynamic Programming
Master MVA: Reinforcement Learning Lecture: 2 Markov Decision Processes and Dnamic Programming Lecturer: Alessandro Lazaric http://researchers.lille.inria.fr/ lazaric/webpage/teaching.html Objectives of
More informationFebruary 11, JEL Classification: C72, D43, D44 Keywords: Discontinuous games, Bertrand game, Toehold, First- and Second- Price Auctions
Eistence of Mied Strateg Equilibria in a Class of Discontinuous Games with Unbounded Strateg Sets. Gian Luigi Albano and Aleander Matros Department of Economics and ELSE Universit College London Februar
More informationRobust Nash Equilibria and Second-Order Cone Complementarity Problems
Robust Nash Equilibria and Second-Order Cone Complementarit Problems Shunsuke Haashi, Nobuo Yamashita, Masao Fukushima Department of Applied Mathematics and Phsics, Graduate School of Informatics, Koto
More informationAdvanced Microeconomic Analysis Solutions to Midterm Exam
Advanced Microeconomic Analsis Solutions to Midterm Exam Q1. (0 pts) An individual consumes two goods x 1 x and his utilit function is: u(x 1 x ) = [min(x 1 + x x 1 + x )] (a) Draw some indifference curves
More informationNBER WORKING PAPER SERIES INDUSTRY DYNAMICS: FOUNDATIONS FOR MODELS WITH AN INFINITE NUMBER OF FIRMS
NBER WORKING PAPER SERIES INDUSTRY DYNAMICS: FOUNDATIONS FOR MODELS WITH AN INFINITE NUMBER OF FIRMS Gabriel Y. Weintraub C. Lanier Benkard Benjamin Van Roy Working Paper 16286 http://www.nber.org/papers/w16286
More informationDynamic stochastic game and macroeconomic equilibrium
Dynamic stochastic game and macroeconomic equilibrium Tianxiao Zheng SAIF 1. Introduction We have studied single agent problems. However, macro-economy consists of a large number of agents including individuals/households,
More informationA. Derivation of regularized ERM duality
A. Derivation of regularized ERM dualit For completeness, in this section we derive the dual 5 to the problem of computing proximal operator for the ERM objective 3. We can rewrite the primal problem as
More informationOrdinal One-Switch Utility Functions
Ordinal One-Switch Utilit Functions Ali E. Abbas Universit of Southern California, Los Angeles, California 90089, aliabbas@usc.edu David E. Bell Harvard Business School, Boston, Massachusetts 0163, dbell@hbs.edu
More informationAn Empirical Algorithm for Relative Value Iteration for Average-cost MDPs
2015 IEEE 54th Annual Conference on Decision and Control CDC December 15-18, 2015. Osaka, Japan An Empirical Algorithm for Relative Value Iteration for Average-cost MDPs Abhishek Gupta Rahul Jain Peter
More informationSUPPLEMENT TO MARKOV PERFECT INDUSTRY DYNAMICS WITH MANY FIRMS TECHNICAL APPENDIX (Econometrica, Vol. 76, No. 6, November 2008, )
Econometrica Supplementary Material SUPPLEMENT TO MARKOV PERFECT INDUSTRY DYNAMICS WITH MANY FIRMS TECHNICAL APPENDIX (Econometrica, Vol. 76, No. 6, November 2008, 1375 1411) BY GABRIEL Y. WEINTRAUB,C.LANIER
More informationNear-Potential Games: Geometry and Dynamics
Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo January 29, 2012 Abstract Potential games are a special class of games for which many adaptive user dynamics
More information1 Lattices and Tarski s Theorem
MS&E 336 Lecture 8: Supermodular games Ramesh Johari April 30, 2007 In this lecture, we develop the theory of supermodular games; key references are the papers of Topkis [7], Vives [8], and Milgrom and
More informationON THE OPTIMALITY OF LEPT AND µc RULES FOR PARALLEL PROCESSORS AND DEPENDENT ARRIVAL PROCESSES
ON THE OPTIMALITY OF LEPT AND µc RULES FOR PARALLEL PROCESSORS AND DEPENDENT ARRIVAL PROCESSES Arie Hordijk & Ger Koole* Dept. of Mathematics and Computer Science, Universit of Leiden P.O. Box 9512, 2300
More informationTechnical Note Preservation of Supermodularity in Parametric Optimization Problems with Nonlattice Structures
Published online ahead of print October 1, 2013 OPERATIONS RESEARCH Articles in Advance, pp. 1 8 ISSN 0030-364X (print) ISSN 1526-5463 (online) http://dx.doi.org/10.1287/opre.2013.1203 2013 INFORMS Technical
More informationOne-Dimensional Dynamics
One-Dimensional Dnamics We aim to characterize the dnamics completel describe the phase portrait) of the one-dimensional autonomous differential equation Before proceeding we worr about 1.1) having a solution.
More informationMarkov Perfect Equilibria in the Ramsey Model
Markov Perfect Equilibria in the Ramsey Model Paul Pichler and Gerhard Sorger This Version: February 2006 Abstract We study the Ramsey (1928) model under the assumption that households act strategically.
More informationBlow-up collocation solutions of nonlinear homogeneous Volterra integral equations
Blow-up collocation solutions of nonlinear homogeneous Volterra integral equations R. Benítez 1,, V. J. Bolós 2, 1 Dpto. Matemáticas, Centro Universitario de Plasencia, Universidad de Extremadura. Avda.
More informationRandomized Sampling for Large Zero-Sum Games,
Randomized Sampling for Large Zero-Sum Games, Shaunak D. Bopardikar a,, Alessandro Borri b, João P. Hespanha c, Maria Prandini d, Maria D. Di Benedetto e a United Technologies Research Center Inc., Berkele,
More informationUnified Convergence Proofs of Continuous-time Fictitious Play
Unified Convergence Proofs of Continuous-time Fictitious Play Jeff S. Shamma and Gurdal Arslan Department of Mechanical and Aerospace Engineering University of California Los Angeles {shamma,garslan}@seas.ucla.edu
More informationMATH College Algebra Review for Test 2
MATH 4 - College Algebra Review for Test Sections. and.. For f (x) = x + 4x + 5, give (a) the x-intercept(s), (b) the -intercept, (c) both coordinates of the vertex, and (d) the equation of the axis of
More informationLecture 7: Advanced Coupling & Mixing Time via Eigenvalues
Counting and Sampling Fall 07 Lecture 7: Advanced Coupling & Miing Time via Eigenvalues Lecturer: Shaan Oveis Gharan October 8 Disclaimer: These notes have not been subjected to the usual scrutin reserved
More informationMean Field Competitive Binary MDPs and Structured Solutions
Mean Field Competitive Binary MDPs and Structured Solutions Minyi Huang School of Mathematics and Statistics Carleton University Ottawa, Canada MFG217, UCLA, Aug 28 Sept 1, 217 1 / 32 Outline of talk The
More information4 Inverse function theorem
Tel Aviv Universit, 2013/14 Analsis-III,IV 53 4 Inverse function theorem 4a What is the problem................ 53 4b Simple observations before the theorem..... 54 4c The theorem.....................
More informationPrice and Capacity Competition
Price and Capacity Competition Daron Acemoglu, Kostas Bimpikis, and Asuman Ozdaglar October 9, 2007 Abstract We study the efficiency of oligopoly equilibria in a model where firms compete over capacities
More informationMA Game Theory 2005, LSE
MA. Game Theor, LSE Problem Set The problems in our third and final homework will serve two purposes. First, the will give ou practice in studing games with incomplete information. Second (unless indicated
More informationNear-Potential Games: Geometry and Dynamics
Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo September 6, 2011 Abstract Potential games are a special class of games for which many adaptive user dynamics
More informationNBER WORKING PAPER SERIES PRICE AND CAPACITY COMPETITION. Daron Acemoglu Kostas Bimpikis Asuman Ozdaglar
NBER WORKING PAPER SERIES PRICE AND CAPACITY COMPETITION Daron Acemoglu Kostas Bimpikis Asuman Ozdaglar Working Paper 12804 http://www.nber.org/papers/w12804 NATIONAL BUREAU OF ECONOMIC RESEARCH 1050 Massachusetts
More informationA Framework for Dynamic Oligopoly in Concentrated Industries
A Framework for Dynamic Oligopoly in Concentrated Industries Vivek Farias Bar Ifrach Gabriel Y. Weintraub October, 2011 NEW VERSION COMING SOON Abstract We consider dynamic oligopoly models in the spirit
More informationEconometric Analysis of Games 1
Econometric Analysis of Games 1 HT 2017 Recap Aim: provide an introduction to incomplete models and partial identification in the context of discrete games 1. Coherence & Completeness 2. Basic Framework
More informationMathematical models in economy. Short descriptions
Chapter 1 Mathematical models in economy. Short descriptions 1.1 Arrow-Debreu model of an economy via Walras equilibrium problem. Let us consider first the so-called Arrow-Debreu model. The presentation
More informationECO 2901 EMPIRICAL INDUSTRIAL ORGANIZATION
ECO 2901 EMPIRICAL INDUSTRIAL ORGANIZATION Lecture 8: Dynamic Games of Oligopoly Competition Victor Aguirregabiria (University of Toronto) Toronto. Winter 2017 Victor Aguirregabiria () Empirical IO Toronto.
More informationOperations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads
Operations Research Letters 37 (2009) 312 316 Contents lists available at ScienceDirect Operations Research Letters journal homepage: www.elsevier.com/locate/orl Instability of FIFO in a simple queueing
More informationTwo-sided Mullins-Sekerka flow does not preserve convexity
Differential Equations and Computational Simulations III J. Graef, R. Shivaji, B. Soni, & J. Zhu (Editors) Electronic Journal of Differential Equations, Conference 01, 1997, pp. 171 179. ISSN: 1072-6691.
More informationNotes on Measure Theory and Markov Processes
Notes on Measure Theory and Markov Processes Diego Daruich March 28, 2014 1 Preliminaries 1.1 Motivation The objective of these notes will be to develop tools from measure theory and probability to allow
More informationDiscontinuous Galerkin method for a class of elliptic multi-scale problems
INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN FLUIDS Int. J. Numer. Meth. Fluids 000; 00: 6 [Version: 00/09/8 v.0] Discontinuous Galerkin method for a class of elliptic multi-scale problems Ling Yuan
More informationLarge-population, dynamical, multi-agent, competitive, and cooperative phenomena occur in a wide range of designed
Definition Mean Field Game (MFG) theory studies the existence of Nash equilibria, together with the individual strategies which generate them, in games involving a large number of agents modeled by controlled
More informationYou don't have to be a mathematician to have a feel for numbers. John Forbes Nash, Jr.
Course Title: Real Analsis Course Code: MTH3 Course instructor: Dr. Atiq ur Rehman Class: MSc-II Course URL: www.mathcit.org/atiq/fa5-mth3 You don't have to be a mathematician to have a feel for numbers.
More informationMarkov Perfect Industry Dynamics with Many Firms
Markov Perfect Industry Dynamics with Many Firms Gabriel Y. Weintraub Stanford University C. Lanier Benkard Stanford University and NBER Benjamin Van Roy Stanford University PRELIMINARY October 16, 2005
More informationUncertainty and Disagreement in Equilibrium Models
Uncertainty and Disagreement in Equilibrium Models Nabil I. Al-Najjar & Northwestern University Eran Shmaya Tel Aviv University RUD, Warwick, June 2014 Forthcoming: Journal of Political Economy Motivation
More informationA Geometric Approach to Inference in Set Identi ed Entry Games"
TSE 943 Jul 28 A Geometric Approach to Inference in Set Identi ed Entr Games" Christian Bontemps and Rohit Kumar A Geometric Approach to Inference in Set-Identified Entr Games Christian Bontemps Rohit
More informationCS364A: Algorithmic Game Theory Lecture #13: Potential Games; A Hierarchy of Equilibria
CS364A: Algorithmic Game Theory Lecture #13: Potential Games; A Hierarchy of Equilibria Tim Roughgarden November 4, 2013 Last lecture we proved that every pure Nash equilibrium of an atomic selfish routing
More informationEC5555 Economics Masters Refresher Course in Mathematics September 2013
EC5555 Economics Masters Refresher Course in Mathematics September 013 Lecture 9 Differential equations Francesco Feri Integration and differential equations. Definition: a differential equation includes
More informationNotes on Convexity. Roy Radner Stern School, NYU. September 11, 2006
Notes on Convexit Ro Radner Stern School, NYU September, 2006 Abstract These notes are intended to complement the material in an intermediate microeconomic theor course. In particular, the provide a rigorous
More informationUC San Francisco UC San Francisco Previously Published Works
UC San Francisco UC San Francisco Previousl Published Works Title Radiative Corrections and Quantum Chaos. Permalink https://escholarship.org/uc/item/4jk9mg Journal PHYSICAL REVIEW LETTERS, 77(3) ISSN
More informationAn approximate dynamic programming approach to solving dynamic oligopoly models
RAND Journal of Economics Vol. 43, No. 2, Summer 2012 pp. 253 282 An approximate dynamic programming approach to solving dynamic oligopoly models Vivek Farias Denis Saure and Gabriel Y. Weintraub In this
More informationON MEAN FIELD GAMES. Pierre-Louis LIONS. Collège de France, Paris. (joint project with Jean-Michel LASRY)
Collège de France, Paris (joint project with Jean-Michel LASRY) I INTRODUCTION II A REALLY SIMPLE EXAMPLE III GENERAL STRUCTURE IV TWO PARTICULAR CASES V OVERVIEW AND PERSPECTIVES I. INTRODUCTION New class
More informationAppendix 2 Complex Analysis
Appendix Complex Analsis This section is intended to review several important theorems and definitions from complex analsis. Analtic function Let f be a complex variable function of a complex value, i.e.
More informationMarcinkiewicz Interpolation Theorem by Daniel Baczkowski
1 Introduction Marcinkiewicz Interpolation Theorem b Daniel Baczkowski Let (, µ be a measure space. Let M denote the collection of all extended real valued µ-measurable functions on. Also, let M denote
More information1.3 LIMITS AT INFINITY; END BEHAVIOR OF A FUNCTION
. Limits at Infinit; End Behavior of a Function 89. LIMITS AT INFINITY; END BEHAVIOR OF A FUNCTION Up to now we have been concerned with its that describe the behavior of a function f) as approaches some
More informationOn the Convergence of Optimistic Policy Iteration
Journal of Machine Learning Research 3 (2002) 59 72 Submitted 10/01; Published 7/02 On the Convergence of Optimistic Policy Iteration John N. Tsitsiklis LIDS, Room 35-209 Massachusetts Institute of Technology
More informationDecentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach
Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach 1 Ashutosh Nayyar, Aditya Mahajan and Demosthenis Teneketzis Abstract A general model of decentralized
More informationMultimarket Oligopolies with Restricted Market Access
Multimarket Oligopolies with Restricted Market Access Tobias Harks 1 and Max Klimm 2 1 Department of Quantitative Economics, Maastricht University, the Netherlands. t.harks@maastrichtuniversity.nl 2 Department
More informationOutline. Linear Matrix Inequalities and Robust Control. Outline. Time-Invariant Parametric Uncertainty. Robust stability analysis
Outline Linear Matrix Inequalities and Robust Control Carsten Scherer and Siep Weiland 7th Elgersburg School on Mathematical Sstems Theor Class 4 version March 6, 2015 1 Robust Stabilit Against Uncertainties
More informationPayoff Continuity in Incomplete Information Games
journal of economic theory 82, 267276 (1998) article no. ET982418 Payoff Continuity in Incomplete Information Games Atsushi Kajii* Institute of Policy and Planning Sciences, University of Tsukuba, 1-1-1
More information(1) I. INTRODUCTION (2)
920 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 57, NO. 4, APRIL 2012 Synchronization of Coupled Oscillators is a Game Huibing Yin, Member, IEEE, Prashant G. Mehta, Member, IEEE, Sean P. Meyn, Fellow,
More information1 Stochastic Dynamic Programming
1 Stochastic Dynamic Programming Formally, a stochastic dynamic program has the same components as a deterministic one; the only modification is to the state transition equation. When events in the future
More informationOptimal Decentralized Control of Coupled Subsystems With Control Sharing
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 9, SEPTEMBER 2013 2377 Optimal Decentralized Control of Coupled Subsystems With Control Sharing Aditya Mahajan, Member, IEEE Abstract Subsystems that
More informationDynamic Stochastic Games with Sequential State-to-State Transitions
Dynamic Stochastic Games with Sequential State-to-State Transitions Ulrich Doraszelski Harvard University and CEPR Kenneth L. Judd Hoover Institution and NBER May 2007 Preliminary and incomplete. Introduction
More informationComputation of Csiszár s Mutual Information of Order α
Computation of Csiszár s Mutual Information of Order Damianos Karakos, Sanjeev Khudanpur and Care E. Priebe Department of Electrical and Computer Engineering and Center for Language and Speech Processing
More informationCompetition, Market Coverage, and Quality Choice in Interconnected Platforms
Competition, Market Coverage, and Qualit Choice in Interconnected Platforms The MIT Facult has made this article openl available Please share how this access benefits ou Your stor matters Citation As Published
More informationThe Entropy Power Inequality and Mrs. Gerber s Lemma for Groups of order 2 n
The Entrop Power Inequalit and Mrs. Gerber s Lemma for Groups of order 2 n Varun Jog EECS, UC Berkele Berkele, CA-94720 Email: varunjog@eecs.berkele.edu Venkat Anantharam EECS, UC Berkele Berkele, CA-94720
More informationMinimax Learning for Remote Prediction
1 Minimax Learning for Remote Prediction Cheuk Ting Li *, Xiugang Wu, Afer Ozgur, Abbas El Gamal * Department of Electrical Engineering and Computer Sciences Universit of California, Berkele Email: ctli@berkele.edu
More informationMonte-Carlo SURE: A Black-Box Optimization of Regularization Parameters for General Denoising Algorithms - Supplementary Material
Monte-Carlo SURE: A Blac-Box Optimization of Regularization Parameters for General Denoising Algorithms - Supplementar Material Sathish Ramani, Student Member, Thierr Blu, Senior Member, and Michael Unser,
More informationCONTINUOUS SPATIAL DATA ANALYSIS
CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s
More informationAsymptotics of Efficiency Loss in Competitive Market Mechanisms
Page 1 of 40 Asymptotics of Efficiency Loss in Competitive Market Mechanisms Jia Yuan Yu and Shie Mannor Department of Electrical and Computer Engineering McGill University Montréal, Québec, Canada jia.yu@mail.mcgill.ca
More informationIdentifying Dynamic Games with Serially-Correlated Unobservables
Identifying Dynamic Games with Serially-Correlated Unobservables Yingyao Hu Dept. of Economics, Johns Hopkins University Matthew Shum Division of Humanities and Social Sciences, Caltech First draft: August
More informationEcon 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers
ECO 504 Spring 2009 Chris Sims Econ 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers Christopher A. Sims Princeton University sims@princeton.edu February 4, 2009 0 Example: LQPY The ordinary
More informationFixed Point Theorems
Fixed Point Theorems Definition: Let X be a set and let f : X X be a function that maps X into itself. (Such a function is often called an operator, a transformation, or a transform on X, and the notation
More informationA Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices. Abstract
A Stackelberg Game Model of Dynamic Duopolistic Competition with Sticky Prices Kenji Fujiwara School of Economics, Kwansei Gakuin University Abstract We develop the following Stackelberg game model of
More informationSection 1.5 Formal definitions of limits
Section.5 Formal definitions of limits (3/908) Overview: The definitions of the various tpes of limits in previous sections involve phrases such as arbitraril close, sufficientl close, arbitraril large,
More informationSelecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden
1 Selecting Efficient Correlated Equilibria Through Distributed Learning Jason R. Marden Abstract A learning rule is completely uncoupled if each player s behavior is conditioned only on his own realized
More informationSample-based Optimal Transport and Barycenter Problems
Sample-based Optimal Transport and Barcenter Problems MAX UANG New York Universit, Courant Institute of Mathematical Sciences AND ESTEBAN G. TABA New York Universit, Courant Institute of Mathematical Sciences
More informationHigh-dimensional Problems in Finance and Economics. Thomas M. Mertens
High-dimensional Problems in Finance and Economics Thomas M. Mertens NYU Stern Risk Economics Lab April 17, 2012 1 / 78 Motivation Many problems in finance and economics are high dimensional. Dynamic Optimization:
More informationQUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS
Journal of the Operations Research Societ of Japan 008, Vol. 51, No., 191-01 QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS Tomonari Kitahara Shinji Mizuno Kazuhide Nakata Toko Institute of Technolog
More informationLinear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013
Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,
More informationResearch Article Chaos and Control of Game Model Based on Heterogeneous Expectations in Electric Power Triopoly
Discrete Dnamics in Nature and Societ Volume 29, Article ID 469564, 8 pages doi:.55/29/469564 Research Article Chaos and Control of Game Model Based on Heterogeneous Epectations in Electric Power Triopol
More informationBoundary Behavior of Excess Demand Functions without the Strong Monotonicity Assumption
Boundary Behavior of Excess Demand Functions without the Strong Monotonicity Assumption Chiaki Hara April 5, 2004 Abstract We give a theorem on the existence of an equilibrium price vector for an excess
More information6.207/14.15: Networks Lecture 11: Introduction to Game Theory 3
6.207/14.15: Networks Lecture 11: Introduction to Game Theory 3 Daron Acemoglu and Asu Ozdaglar MIT October 19, 2009 1 Introduction Outline Existence of Nash Equilibrium in Infinite Games Extensive Form
More informationPrice and Capacity Competition
Price and Capacity Competition Daron Acemoglu a, Kostas Bimpikis b Asuman Ozdaglar c a Department of Economics, MIT, Cambridge, MA b Operations Research Center, MIT, Cambridge, MA c Department of Electrical
More informationOPTIMAL CONTROL FOR A PARABOLIC SYSTEM MODELLING CHEMOTAXIS
Trends in Mathematics Information Center for Mathematical Sciences Volume 6, Number 1, June, 23, Pages 45 49 OPTIMAL CONTROL FOR A PARABOLIC SYSTEM MODELLING CHEMOTAXIS SANG UK RYU Abstract. We stud the
More informationANALYTIC SOLUTIONS FOR HAMILTON-JACOBI-BELLMAN EQUATIONS
Electronic Journal of Differential Equations, Vol. 2016 2016), No. 08, pp. 1 11. ISSN: 1072-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu ANALYTIC SOLUTIONS FOR HAMILTON-JACOBI-BELLMAN
More informationLogit Dynamic for Continuous Strategy Games: Existence of Solutions.
Logit Dynamic for Continuous Strategy Games: Existence of Solutions. Ratul Lahkar August 1, 2007 Abstract We define the logit dynamic in the space of probability measures for a game with a compact and
More informationCompeting Teams. Hector Chade 1 Jan Eeckhout 2. SED June, Arizona State University 2 University College London and Barcelona GSE-UPF
Competing Teams Hector Chade 1 Jan Eeckhout 2 1 Arizona State University 2 University College London and Barcelona GSE-UPF SED June, 2014 The Problem We analyze assortative matching with externalities
More informationMULTI-OBJECTIVE MODEL CHECKING OF MARKOV DECISION PROCESSES
Logical Methods in Computer Science Vol. 4 (4:8) 2008, pp. 1 21 www.lmcs-online.org Submitted Oct. 4, 2007 Published Nov. 12, 2008 MULTI-OBJECTIVE MODEL CHECKING OF MARKOV DECISION PROCESSES KOUSHA ETESSAMI
More informationCELLULAR automaton (CA) is a discrete model consists
1622 IEEE TRANSACTIONS ON CYBERNETICS, VOL. 45, NO. 8, AUGUST 2015 Irregular Cellular Learning Automata Mehdi Esnaashari and Mohammad Reza Mebodi Abstract Cellular learning automaton CLA) is a recentl
More informationBasics of Game Theory
Basics of Game Theory Giacomo Bacci and Luca Sanguinetti Department of Information Engineering University of Pisa, Pisa, Italy {giacomo.bacci,luca.sanguinetti}@iet.unipi.it April - May, 2010 G. Bacci and
More informationMean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio
Mean-Field optimization problems and non-anticipative optimal transport Beatrice Acciaio London School of Economics based on ongoing projects with J. Backhoff, R. Carmona and P. Wang Robust Methods in
More informationOn the Unique D1 Equilibrium in the Stackelberg Model with Asymmetric Information Janssen, M.C.W.; Maasland, E.
Tilburg University On the Unique D1 Equilibrium in the Stackelberg Model with Asymmetric Information Janssen, M.C.W.; Maasland, E. Publication date: 1997 Link to publication General rights Copyright and
More informationDynamic Discrete Choice Structural Models in Empirical IO
Dynamic Discrete Choice Structural Models in Empirical IO Lecture 4: Euler Equations and Finite Dependence in Dynamic Discrete Choice Models Victor Aguirregabiria (University of Toronto) Carlos III, Madrid
More informationCOMPUTATIONAL METHODS AND ALGORITHMS Vol. I - Numerical Methods for Ordinary Differential Equations and Dynamic Systems - E.A.
COMPUTATIOAL METHODS AD ALGORITHMS Vol. I umerical Methods for Ordinar Differential Equations and Dnamic Sstems E.A. oviov UMERICAL METHODS FOR ORDIARY DIFFERETIAL EQUATIOS AD DYAMIC SYSTEMS E.A. oviov
More informationLecture notes for Analysis of Algorithms : Markov decision processes
Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with
More informationSupermodular Games. Ichiro Obara. February 6, 2012 UCLA. Obara (UCLA) Supermodular Games February 6, / 21
Supermodular Games Ichiro Obara UCLA February 6, 2012 Obara (UCLA) Supermodular Games February 6, 2012 1 / 21 We study a class of strategic games called supermodular game, which is useful in many applications
More information