Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty

Size: px
Start display at page:

Download "Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty"

Transcription

1 Stochastics and uncertainty underlie all the processes of the Universe. N.N.Moiseev Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty by Iouldouz S. Raguimov Department of Mathematics and Statistics, York University 4700 Keele Street, Toronto, Ontario, Canada, M3J 1P3 November 13, 2014 Abstract The Best Guaranteed Result Principle is introduced. Decision Making in Operations with Stochastic Factors and Uncertainty is studied. Information hypothesis and decision making turn are introduced. Various scenarios of game situations with different turn sequences of decision making and voluntary information exchange between the players are studied. Decision making in operations with stochastic factors under the assumption of the frequency definition of probability is analyzed. In our setting the possibility of utilization of stochastic information about an uncontrollable factor is connected with the possibility (and intelligibility) of transition from the original operation to an extended one with infinitely many repetitions. It is shown that if in the extended operation we tend to maximize the lower limit of the average value of an original criterion, then the best guaranteed result of the Operating Side is determined as a solution to a certain optimization problem and its optimal strategy will be choosing on each step of the extended operation some optimal solution to that optimization problem. Similar results for the problem of maximization of the probability of exceeding of a given level and the problem of maximization of a level with given guarantee are studied. Applications of the obtained results to a two-stage stochastic programming problem are considered.

2 1 Introduction Problems of control in economic, financial as well as engineering systems are often formulated as a decision-making problem under stochastics and uncertainty. The mathematical formalization of such problems usually requires a non-trivial non-classical approach. Let us define an operation as a series of actions to achieve a certain objective (criterion), i.e., as a series of objective-oriented acts. In operation, a rational entity consisting of one or several persons aiming to achieve the given objective is called an operating side (OS). Within an operating side we will distinguish an operation research analyst (ORA) as a group of one or several persons that does not make a decision by itself, but helps the operating side to do so. The operation usually involves other participants with different objectives and the result of an operation depends on their activities also. Consequently, as a rule, there are some uncontrollable (by the operating side) factors in the operation, that having influenced the result of the operation create a so-called operational situation. According to the information that is available to an ORA these uncontrollable factors can be divided into three groups: 1. Constant factors whose values are known to the ORA; 2. Stochastic factors, i.e. stochastic processes with given distributions; 3. Uncertain factors that have known ranges of values, but uncertain exact value at any particular time. (If the uncertain factor is known to be stochastic, then it is assumed that OR analyst does not know its exact distribution.) The uncertain factors in their turn can also be divided into three groups: (a) uncertain factors that are the result of incomplete knowledge about some processes or quantities in an operation (we call them natural uncertainties); (b) uncertain factors related to an ambiguity in the operating side s objective; 2

3 (c) uncertain factors representing the activities of other participants involved in an operation that act independently and have objectives different from those of the OS. A participant with the opposite objective is called an opponent. 2 Decision making under uncertainty 2.1 Mathematical model of operation In operations containing only uncontrollable factors of type (1) a decisionmaking problem is usually reduced to the solution of a deterministic optimization problem. But, if operation contains also uncontrollable factors of the second and third types, i.e. stochastic and uncertain factors, then to analyze it we need to involve the methods of Stochastic Optimization, Game Theory and Decision Theory. There exist a number of decision-making principles that are used widely in Decision Theory: The minimax criterion; Savage s minimax regret criterion; The expected value criterion; Hurwicz s criterion; Bayes s criterion; and others. However, our main here goal is to introduce the approach based on the Best Guaranteed Result Principle. Suppose that in an operation the criterion of OS is formulated as a maximization of function F (x, y, z), where the vector x M 0 denotes control variables, the vector y N denotes uncontrollable uncertain factors and z denotes a vector of uncontrollable stochastic parameters with distribution function θ(z) on the set Z. The function θ(z) is either known exactly or it is just given that θ Θ, where Θ is a set of distribution functions. Usually, the following two ways of determining set Θ are considered: 1. Θ = {θ α α A}. Here the type of distribution rule, θ α is given, while the vector of parameters α, determining the exact distribution function is uncertain factor with the range of values in the set A at any particular time; 3

4 2. Θ = {θ a i a i (z) dθ(z) a i Z i = 1, 2,, l}, where a i (z), i = 1, 2,, l are given integrable functions. Please note that here the type of distribution rule is unknown but we are given some constraints on its integral characteristics. As a rule, the model of operation is analyzed before the operation takes place. Assume, that there another participant is involved in an operation whose criterion is formulated as the maximization of G(x, y, z). In particular, if G(x, y, z) = F (x, y, z), we obtain a Zero-Sum Game and this participant, as mentioned above, is called an opponent. Usually, while an operation evolves, some information about uncontrollable factors is received by the OS. An exact description of the process of receiving this information is called an information hypothesis. An action allowable by an information hypothesis is called a strategy of the operating side. Mathematically a strategy is defined as a mapping x: N Z M 0. Denote the set of all such functions by M. Let M be the set of strategies, i.e., the set of functions in M, corresponding to the information hypothesis. Notice that strategies are usually determined before operations, once the model of operation is analyzed by the ORA. Consider the following examples of information hypotheses and corresponding sets of strategies: (a) no information is expected about uncontrollable factors. Then M = M 0, where M 0 is the set of all constant strategies, i.e. x x; (b) at the beginning of the operation the exact values of the uncontrollable factors y and z are known. Then M = M, where M is the set of all functions x: N Z M 0 ; (c) at the beginning of the operation exact values of uncertain factors y are known. No information is expected about stochastic factors. Then M = M y, where M y is the set of all functions x: N M 0 ; (d) the information hypothesis may be given using the so-called information function R: N Z R m, defined on N Z. In this case, at the 4

5 beginning of the operation the exact values of R(y, z) for all values of y and z are known. Then the set of strategies of the operating side M = { x M R(y 1, z 1 ) = R(y 2, z 2 ) = x(y 1, z 1 ) = x(y 2, z 2 ) y 1, z 1, y 2, z 2 }. Therefore, the the mathemathical model of operation is determined as a set of the following objects: F (x, y, z), x M 0, y N, z Z, θ Θ, x M. 3 The efficiency of strategies How do we compare strategies? There exist two different approaches to the evaluation of the efficiency of strategy x: N Z M 0 based on either of the following principles: (A) averaging of the value of objective function; (B) the value at risk or conditional value at risk measures of the risk of loss. There exist other principles to evaluate strategies. However, our approach is primarily based on the Best Guaranteed Result Principle. Let us start with mixed strategies. 3.1 Expected value criterion In our setting, for every strategy x M, its efficiency, denoted by W ( x) is estimated using the Guaranteed Result Principle, where W ( x) is a certain guaranteed value of the objective function. In comtaring two strategies, we define the better one to be one with a larger efficiency estimate. The expected value criterion suggests estimating the efficiency of strategy x(y, z) using the function F ( x(y, z), y, θ) = F ( x(y, z), y, z) dθ(z). Z instead of the original criterion F ( x(y, z), y, z). As a result, the OS maximizes the average efficiency. Further suppose: 5

6 1. Every uncertain factor in y is either natural uncertainty or a strategy of an opponent who does not have information about the realizations of stochastic factor z; 2. Opponent s objective is either unknown or opposite to the objective of operating side; 3. The OS agrees with abovementioned averaging of the criterion with respect to stochastic factors. Then the estimate of efficiency for strategy x is given as: W ( x) = inf inf F ( x, y, θ) = inf inf F ( x(y, z), y, z) dθ(z). y N θ Θ y N θ Θ Definition: Strategy x M is said to be optimal, if Z W ( x ) = max x M W ( x). This value is called the Best Guaranteed Result of the OS. Let us ε > 0 be given. The strategy x ε is called ε-optimal, if W ( x ε ) sup W ( x) ε. x M Please note that determining the value of max W ( x) x M is a problem of maximization on the set of functions M. Let us now assume that the opponent knows the realization of stochastic factor z and consequently, has chosen vector y. Then the estimate of efficiency for strategy x is given as: W ( x) = inf θ Θ Z inf F ( x(y, z), y, z) dθ(z). y N Moreover, in the worst case, when the opponent knows strategy x, s/he will be able to achieve inf F ( x(y, z), y, z) y N for the given realization of z. It is clear that W ( x) W ( x). Indeed, the more information an opponent has, the less is the efficiency estimate of operating side s strategy. 6

7 3.2 Mixed strategies in operation Suppose no information is expected about uncontrollable factors, that is x = x M = M 0. If the OS is not satisfied with the best guaranteed result sup x M0 W (x), then mixed strategies can be used. Definition: A probability distribution ϕ on M 0 is called a mixed strategy. For the OS to use the mixed strategy ϕ is equivalent to taking strategy x M 0 as the value of a random variable with the probability distribution ϕ on M 0. As a result, the OS introduces one more stochastic factor into the operation. Let us determine the conditions related to the using of mixed strategies in an operation. In addition, suppose that the opponent does not have information about the realizations of x M 0. Then the efficiency of the mixed strategy ϕ will be given as: W (ϕ) = inf inf F (x, y, z) dϕ(x) dθ(z). y N θ Θ Z M 0 In the case when y is in control of the opponent and s/he knows the realizations of z, the efficiency of the strategy ϕ will be: W (ϕ) = inf inf F (x, y, z) dϕ(x) dθ(z). θ Θ Z y N M 0 It is clear that once the operation involves uncontrollable stochastic factors, it is easier for ORA to convince the OS to use mixed strategies. Problem 1: Let i M 0 = {1, 2, 3, 4}, j N = {1, 2, 3} and let the criterion of the OS be given as: [F (i, j)] 4 3 = , where the OS chooses the rows of the matrix, i.e., i M 0 = {1, 2, 3, 4}; and the states of Nature are given by the collumns of the matris, i.e., j 7

8 N = {1, 2, 3}. Denote the strategy ĩ : N M 0 of the OS as a vector ĩ = (ĩ(j), j N). Evaluate the efficiency of ( the pure strategies ĩ 1 = (1, 1, 1), ĩ 2 = (3, 2, 4) and 1 the mixed strategy ϕ = 2, 1 3, 0, 1 ) og the OS, if: 6 a) j is an uncertain factor; b) j is a stochastic factor with the distribution θ = ( 2 3, 1 6, 1 ) Optimal strategies in operation What strategies in an operation are optimal? Definition: Strategy x M is said to be optimal, whenever F Γ (M) def = max x M W ( x) = W ( x ). The value F Γ (M) is called the best guaranteed result of the OS. When the OS wants to minimize the value of the criterion, the maximum is replaced by minimum. In particular, the best guaranteed result of the OS in mixed strategies is defined as: def F m = sup W (ϕ), ϕ {ϕ} where {ϕ} is the set of all mixed strategies. In the case when the maximum: max x M W ( x) is not attained, instead of optimal strategies ϵ-optimal strategies are used. Definition: For every sufficiently small ϵ > 0, the strategy x ϵ M is said to be ϵ-optimal, whenever F Γ (M) def = W ( x ϵ ) sup W ( x) ϵ. x M Since we need to maximize W ( x) on the set of functions M, determining the value of F Γ (M) is a quite challenging problem. However, this problem is simplified if the set M contains so-called absolute optimal strategies. 8

9 Assume that the OS knows the probability distribution θ for the vector of stochastic factors z. Definition: Strategy x a M is called absolute optimal, if for every y N F ( x a, y, θ) = max F ( x, y, θ). x M It is clear that for every value of the vector of uncontrollable uncertain factor an absolute optimal strategy is the best strategy among all strategies in M. Theorem 1: If there exists an absolute optimal strategy x a M, then x a is an optimal strategy of the OS and F Γ (M) = inf max F ( x, y, θ). y N x a M In particular, if y N and z Z the maxima max F (x, y, θ) and max F ( x, y, θ) x M 0 x M 0 exist, then and F y def = F Γ (M y ) = inf y N max x M 0 F (x, y, θ), F def = F Γ ( M) = inf max F (x, y, z) d θ(z). y N Z x M 0 Let us now suppose that there are some y N, such that the maximum max x M F ( x, y, θ) is not attained but the corresponding supremum is finite. Then the notion of ϵ-absolute optimal strategy can be introduced. Definition: For every sufficiently small ϵ > 0, the strategy x ϵ a M is said to be an ϵ-absolute optimal, if y N F ( x ϵ a, y, θ) sup F ( x, y, θ) ϵ. x M Unfortunately, as the next example shows, ϵ-absolute optimal strategy may not exist at all. 9

10 Example 1: Let i M = M 0 = {1, 2} be a controllable factor, j N = {1, 2} be an uncontrollable factor and suppose that there is no stochastic factor in the operation. The criterion of the operation is given by the following matrix: [ ] 1 1 [F (i, j)] 2 2 = 1 1 Then ϵ [0, 2) there is no ϵ-absolute optimal strategy in the set M 0. Indeed, when i = 1, ϵ [0, 2) F (1, 1) = 1 < 1 ϵ = F (2, 1) ϵ = max i M 0 F (i, 1) ϵ, i.e., for j = 1 the inequality from the definition of ϵ-absolute optimal strategy is not satisfied. Thus, the strategy i = 1 is not an ϵ-absolute optimal strategy. Show that i = 2 also is not an ϵ-absolute optimal strategy ϵ [0, 2). Theorem 2: If x ϵ a M is an ϵ-absolute optimal strategy, then it is an ϵ-optimal strategy. Let assume that ϵ > 0 there exists an ϵ-absolute optimal strategy x ϵ a M. Then F Γ (M) = inf sup F ( x, y, θ). y N x a M In particular, if y N and z Z the values are finite. Then sup F (x, y, θ) and sup F ( x, y, θ) x M 0 x M 0 F y def = F Γ (M y ) = inf y N sup x M 0 F (x, y, θ), and F def = F Γ ( M) = inf sup F (x, y, z) d θ(z). y N Z x M 0 Finally, let us compare the best guaranteed results for the following sets of strategies: M 0, {ϕ}, M y and M. Denote F 0 Γ def = F Γ (M 0 ) = max min x M 0 y N F (x, y, θ). 10

11 As before, F m = sup W (ϕ); F y = F Γ (M y ); F = FΓ ( M). ϕ {ϕ} Theorem 3: The following inequalities are satisfied: F 0 Γ F m F y F. Proof: Since M 0 {ϕ}, FΓ 0 = sup inf F (x, y, z) d θ(z) x M 0 y N Z sup inf F (x, y, z) d θ(z) d ϕ(x) = F m, and ϕ {ϕ} ϕ {ϕ} y N Hence, F m F y. Finally, M 0 Z W (ϕ) = inf F (x, y, z) d θ(z) d ϕ(x) y N M 0 Z inf y N sup x M 0 F (x, y, θ) d ϕ(x) = F y. M y M = M y F. Remark: (Interpretation of Theorem 3) Let y be the strategy of the opponent, whose goal is the opposite to the goal of OS. Denote def = F m FΓ. 0 P E The value of P E can be considered as the price that the opponent would be willing to pay for the information about the strategy x of the OS. Indeed, if the opponent knows the realization of x, then the OS may guarantee for itself just the result FΓ 0. The efficiency of any mixed strategy ϕ, W (ϕ) = M 0 min y N 11 F (x, y, θ) d ϕ(x)

12 is not grater than F 0 Γ. Similarly, the value P OS def = F y F m can be considered as the price that the OS would be willing to pay for the information about the strategy y of the opponent. Indeed, if the information about the strategy y is expected, then the OS gets the result F y, otherwise just the result F m. Example 2: Let i M 0 = {1, 2, 3} be a controllable factor, j N = {1, 2, 3, 4} be an uncontrollable factor and suppose that there is no stochastic factor in the operation. The criterion of the operation is given by the following matrix: [F (i, j)] 3 4 = a) M = M y is the set of all functions ĩ : N M 0. There are 3 4 = 81 different strategies. According to Theorem 1, The strategy ĩ 0 is optimal, if. F y = F Γ (M y ) = min max F (i, j) = 2. 1 j 4 1 i 3 min F 1 i 3 (ĩ0 (j), j) = 2 and j N, F (ĩ 0 (j), j) 2. So, the set of all optimal strategies is: ĩ 0 (1) = 2, 3; ĩ 0 (2) = 1, 2, 3; ĩ 0 (3) = 1, 3; ĩ 0 (4) = 1, 2. Here, underlined values correspond to an existing optimal strategy. There is a total of = 24 optimal strategies. Let us now find all absolute optimal strategies ĩ a : F (ĩ a (j), j) = max i M 0 F (i, j) j N. Hence, the possible absolute optimal strategies ĩ a (j) are: ĩ a (1) = 3; ĩ a (2) = 1, 2, 3; ĩ a (3) = 1; ĩ a (4) = 1, 2. In total, there are = 6 of them. 12

13 b) Suppose that at the beginning of the operation, the OS will know to which of the following sets N 1 = {1, 2} or N 2 = {3, 4} the value of the uncontrollable factor j belongs. Then the strategy of the OS will be described as: { i1, if j N ĩ(j) = 1 i 2, if j N 2. The set M consists of 9 strategies. It contains absolute optimal strategies ĩ a : ĩ a 1 = 3 and ĩ a 2 = 1. By Theorem 1, F Γ (M) = 2. There are two optimal strategies in the set M : i 0 1 = 2, 3; i 0 2 = 1. Problem 2: In Example 3, for the cases (a) and (b) determine the optimal strategies of the OS on the set M y. 4 Hierarchic games of two persons (Principal-agent games) Assume that there is no stochastic factor and in an operation we want to maximize the function F (x, y), where x X denotes our control and y Y denotes an uncontrollable uncertain factor. If we do not know the other participant s objective, then we consider this operation as a game with Nature. According to the Maximin Principle, first we calculate the function and then determine by solving the optimization problem W (x) = inf F (x, y) y Y sup inf F (x, y) x X y Y W (x) sup. x X However, in conflict situations with two participants the interests of participants are not always opposite. Consider a two-person game in the normal form, that is the game when every participant selects her/his strategy without knowing the selection of the opponent. A pair (x, y) of strategies is called 13

14 a game situation. Let the payoff of Player 1 (i.e. operating side) be described by function F (x, y), and the payoff of Player 2 (i.e., a partner) - by function G(x, y), where G(x, y) F (x, y). Assume, each player is trying to maximize her/his payoff by choosing strategies x X and y Y, respectively. Therefore, the two-person game in normal form is given by Γ = X, Y, F (x, y), G(x, y). We shall be considering two-person games with a turn sequence and voluntary information exchange between the players. These types of games are called hierarchic games (also, games with non-antagonistic interests or principal-agent games). It is supposed that Player 1 has the right to the first turn and we will be trying to determine the Best Guaranteed Result that Player 1 (Operating Side) is able to obtain. For simplicity, let F (x, y), G(x, y) C X Y, where X and Y are compact sets. Depending on Player 1 s information about the strategies of Player 2, it is possible to formulate different games. 14

15 Game Γ 1 : Player 1 does not expect to have any information about the strategies of Player 2 and she has to select her strategy x X, and inform Player 2 about it. This will be denoted schematically as x p1 p2 y. Then, Player 2 will be able to choose her strategies in the form g : X Y. Denote the set of such strategies by {g}. Hence, Game Γ 1 in normal form will be given as Γ 1 = X, {g}, F (x, g), G(x, g), where F (x, g) = F (x, g(x)) and G(x, g) = G(x, g(x)). Player 2 knowing x, will choose the strategy y Y (x) = Arg max G(x, y), y Y that is a strategy maximizing her payoff G(x, y). Since Player 1 knows G(x, y), s/he will be able to determine set Y (x). (The set Y (x) is called the set of indifference for Player 2 under the given constraints.) Then the value W (x) = min F (x, y) y Y (x) is said to be the efficiency estimate for strategy x. Consequently, the best guaranteed result of Player 1 is determined as F 1 = sup x X W (x) = sup x X min y Y (x) F (x, y). Note that since Y (x) Y, the reward of Player 1 may be greater than a simple maximin. 15

16 Game Γ 2 : Player 1 expects to have (and will really have) full information about the strategy y Y of Player 2. Since she has the right to the first turn, she chooses her strategy as a function f : Y X, and inform Player 2 about it. Denote the set of such strategies by {f}. The scheme of information exchange in Game Γ 2 may be given as f p1 p2 y p2 p1 x = f(y). Player 2 knowing f, will choose her strategy y from the set Y (f) = Arg max G(f(y), y). y Y Note that set Y (f) may be an empty set unless f(y) is continuous on Y. If f(y) is not continuous on Y, Player 2 can choose any strategy y Y. Define { Y (f), if Y (f), Y (f) = Y, if Y (f) =. Then the optimal strategy of Player 2 will be a choice of any y Y and the efficiency estimate for strategy f of Player 1 will be given as W (f) = inf F (f(y), y). y Y (f) Consequently, the best guaranteed result of Player 1 is determined as F 2 = sup W (f) = sup f {f} f {f} inf y Y (f) F (f(y), y). 16

17 Game Γ 3 : Suppose that Player 2 plays Game Γ 2 against Player 1, i.e. Player 2, expecting to have full information about the strategy x X of Player 1, selects her strategy as a function g : X Y, and informs Player 1 about it. However, as before, the right to the first turn belongs to Player 1. So, the latter knowing g, selects some operator f 1 : {g} X and informs Player 2 about it. The scheme of information exchange in Game Γ 3 may be p1 p2 described as f 1 g p2 p1 x = f(g) p1 p2 y = g(x). Correspondingly, the best guaranteed result of Player 1 is determined as F 3 = sup f 1 {f 1 } inf F (f 1(g), g), g Y (f 1 ) and where Y (f 1 ) = { Y (f1 ), if Y (f 1 ), {g}, if Y (f) =, Y (f 1 ) = Arg max G(f 1 (g), g). g {g} 17

18 Therefore, in Games Γ 1, Γ 2 and Γ 3 to find the best guaranteed result of Player 1 it is necessary to calculate so-called maximin with connected variables. Moreover, finding the best guaranteed result of Player 1 in Games Γ 2 and Γ 3 requires us to solve optimization problem on a set of functions. Let us consider Game Γ 2. We will need the following notations: G 2 = max y Y min G(x, y), x X which is the best guaranteed result of Player 2 when Player 1 for every y Y uses against her a so-called punishment strategy f p (y) : f p (y) Y (x) = Arg min G(x, y), y Y ; x X E = Arg max y Y min G(x, y); x X D = {(x, y) X Y G(x, y) > G 2 }; K = { sup(x,y) D F (x, y), if D,, if D = ; M = min max y E x X F (x, y). Theorem 4: (Yu.B.Germeier) The best guaranteed result of Player 1 in Game Γ 2 is: F 2 = max{k, M}. 18

19 Now consider Game Γ 3. Define: G 3 = min max x X y Y G(x, y) = max y Y G(xp, y) which represents the best guaranteed result of Player 2 when Player 1 uses against her punishment strategy x p ; D = {(x, y) X Y G(x, y) > G 3 }; K = { sup(x,y) D F (x, y), if D,, if D =. Theorem 5: (N.S.Kukushkin) The best guaranteed result of Player 1 in Game Γ 3 is: F 3 = max{k, F 1 }. 19

20 Continuing the above defined recursion, it is possible to formulate Game Γ 4, Game Γ 5,..., Game Γ k,. However, it is shown (Yu.B.Germeier and N.S. Kukushkin) in 1970 s [1, 2, 3] that F 2n = F 2, and F 2n+1 = F 3 n = 2, 3,. Moreover, it is shown that F 1 F 3 F 2. The last inequality is a consequence of the fact that in Game Γ 3 Player 2 also receives (and might be able to use) some information about the strategies of Player 1. Therefore, in a hierarchic game (principal-agent game, two-person game with non-antagonistic interests): 1. For both players it is better to be mutually informed. 2. The information exchange between the players is useful only up to a certain level. Above that level, further collection and/or exchange of information will not potentially increase the payoff of Player 1 (operating side). That is after a certain level of recursion additional complication of the situation will not potentially increase the payoff of Player 1 (operating side). 20

21 Example 3: Consider the bimatrix game: (3, 7) (6, 4) (8, 3) (4, 7 (3, 7) (2, 3) (7, 4) ( 5, 6) ( 1, 6) Here, the criteria of Player I and Player II are given by the matrices A = and B = respectively, and their sets of strategies X = Y = {1, 2, 3}. Let us formulate Games Γ 1, Γ 2 and Γ 3 for this bimatrix game. Game Γ 1 : where F 1 = max i X min j Y (i) a ij = max min a ij = max W (i), 1 i 3 j Y (i) 1 i 3 Y (i) = Arg max 1 j 3 b ij. Since W (1) = W (2) = 3 and W (3) = 5, the best guaranteed result of Player 1, F 1 = 3, and optimal strategies are i = 1, 2. Game Γ 2 : G 2 = max 1 j 3 min 1 i 3 b ij = 4; E = Arg max 1 j 3 min 1 i 3 b ij = {1, 2}, i.e. i p = 3; D = {(i, j) b ij > 4} = {(1, 1), (2, 1), (2, 2), (3, 2), (3, 3)}; M = min j E max i X a ij = min 1 j 2 max 1 i 3 a ij = 6. K = max a ij = 4; (i,j) D Hence, the best guaranteed result of Player 1, F 2 = max{k, M} = max{4, 6} = 6, and the optimal strategy is { 3, if j = 1, f (j) = 1, if j = 2, 3. Game Γ 3 : G 3 = min 1 i 3 max 1 j 3 b ij = max 1 j 3 b i p j = max 1 j 3 b 3j = 6; 21

22 D = {(i, j) b ij > 6} = {(1, 1), (2, 1), (2, 2)}. So, since K = max (i,j) D a ij = 4, the best guaranteed result of Player 1, F 3 = max{k, F 1 } = max{4, 3} = 4, and the optimal strategy is { 2, if g(2) = 1, f1 (g) = 3, if g(2) 1. Note that max min a ij = 3 and min max a ij = 6. 1 i 3 1 j 3 1 j 3 1 i 3 22

23 5 Decision-making in operation with stochastic factors As we have seen, the problem of determining the best guaranteed result sup W ( x) x M is not straightforward. In general, it requires solving a problem of maximization of W ( x) on the set of functions M. However, it is substantially simplified if M contains so called absolute optimal strategies. In this section it is assumed the the OS knows exactly the distribution function θ of the stochastic vector z. Definition: Strategy x a M is called absolute optimal, if y N F ( x a, y, θ) = max F ( x, y, θ). x M x ε a M When the maximum is not attained, for the given ε > 0, strategy is called ε-absolute optimal, if y N F ( x ε a, y, θ) sup F ( x, y, θ) ε. x M It is not difficult to show that every absolute optimal strategy x a is an optimal strategy and W ( x a ) = max W ( x) = inf x M max y N x M F ( x, y, θ). Similarly, every ε-absolute optimal strategy x ε a is an ε-optimal strategy and W ( x ε a) = sup W ( x) = inf sup F ( x, y, θ). x M y N x M Unfortunately, absolute optimal strategies, as well as ε-absolute optimal strategies often do not exist in real-world operations, in reality they exist rarely. 23

24 Let us now consider an operation containing the uncontrollable stochastic vector z, with the given distribution function θ(z) on the set Z. Suppose that there are no uncertain factors in the operation and that the operating side wants to maximize function F (x, z), where vector x X is a strategy, and by z is denoted the above-mentioned stochastic vector. It is clear that the statement z is a stochastic vector represents some additional information for the OS and generally speaking, may allow the OS to get a result better than inf z Z F (x, z). To evaluate the efficiency of strategy x in operations with stochastic factors the following two methods are often used: 1. Set Z is replaced by S Z, such that P [z S] 1 β, where β > 0 is quite a small number, and the function W (x) = inf F (x, z) z S is used to estimate the efficiency of strategy x; 2. The objective function F (x, z) is replaced by the new objective function F (x) = F (x, z) dθ(z). Z It is clear that the both methods entail a certain risk and this risk may be quite large if the operation is not repeated a sufficient number of times. Hence, by using either of these methods we introduce into the operation a new element - the consent to certain risk. 24

25 Now, we will accept the frequency definition of probability. Stated simply, the probability of an event is assumed to be the frequency of this event occurring in a very long series of experiments. More formally, it is supposed that probability is defined as a certain limit when the operation is repeated infinitely often. We assume that: 1. The operation might be repeated infinitely, that is the operating side can choose strategy x X any number of times, and some values of vector z will appear the same number of times. Let s denote a sequence of occurred values of z by z, i.e. z = { z 1, z 2, z 3,..., z N,... }, where z i Z denotes an uncontrollable factor at the i-th stage (repetition) of the operation, i = 1, 2,..., N,...; 2. In this infinitely repeated operation (we will call this operation the extended operation), where the uncontrollable factor, instead of vector z, is an infinite sequence z; this sequence z may be any random sequence with the single constraint that the limit of the realization (occurring) frequency for vector z j exists and is equal to p j. So, if we denote by N the current number of repetitions of the operation and χ j (z) is the characteristic function defined as { 1, if z = zj χ j (z) = 0, if z z j, then, for all j such that z j Z the limit exists and is equal to p j. lim N 1 N N χ j (z i ) i=1 Therefore, in our setting the possibility of utilization of stochastic information about z is connected to the possibility (and intelligibility) of the transition from the original operation to the extended operation with infinitely many repetitions. Suppose that the objective function F (x, z) is bounded, that is, there exists a positive real number C, such that C < F (x, z) < C, for all values of x X R n and stochastic vectors z Z, with P (z = z j ) = p j, j = 1, 2,..., j Z p + j = 1. 25

26 Assume that the operating side agreed to transfer from the original onestage operation into the extended operation. Then, in the extended operation, the set of its strategies will be defined as X = { x = {x i } i=1 x i X, i = 1, 2, 3, }, where x i is the choice of the OS at the stage i; and the range of values for the uncontrollable stochastic factor will be Z = { z = {z i } 1 N i=1 zi Z & lim χ j (z i ) = p j, j = 1, 2, 3, }, N N where z i denotes an uncontrollable stochastic factor z at stage i. Define the average value of the objective function for N steps by i=1 F N (x 1, x 2,..., x N ; z 1, z 2,..., z N ) = 1 N N F (x i, z i ), and assume that in the extended operation the operating side wants to maximize F ( x; z) = lim inf F N(x 1, x 2,..., x N ; z 1, z 2,..., z N ) = lim inf N N Under these conditions it has been proved that sup inf x X z Z i=1 1 N N F (x i, z i ). i=1 (1) F ( x; z) = sup EF (x, z). (2) x X Therefore, we have the following theorem: Theorem 6: If, in the extended operation, the OS wants to maximize the lower limit (1), then its best guaranteed result is determined as the optimal objective value of the problem sup x X EF (x, z), and the optimal strategy is to choose at each stage of the operation some x X, such that EF ( x, z) = max EF (x, z), x X if the maximum is attained on X, or the corresponding ε-optimal strategy, otherwise. 26

27 Next, assume; that in the extended operation, the OS is interested in maximization of the functional Φ( x; z) = lim inf N 1 N N χ(x i, z i ), (3) i=1 where χ(x, z) = { 1, if F (x, z) u0 0, if F (x, z) < u 0, and u 0 is some given value of F (x, z), with u 0 ( C, C). Then applying Theorem 6 to the operation with objective function (3), we obtain sup inf x X z Z On the other hand, from definition (4), Φ( x; z) = sup Eχ(x, z). x X (4) Eχ(x, z) = p j χ(x, z j ) = p j, (5) j=1 j J(u 0,x) where Consequently, J(u 0, x) = { j Z + F (x, z j ) u 0 }. sup inf x X z Z Φ( x; z) = sup x X j J(u 0,x) p j. (6) 27

28 Therefore, we have the following corollary: Corollary 1: If, in the extended operation, the OS wants to maximize a functional, then the best guaranteed result is defined as the supremum of the functional (5), and the optimal strategy is to choose, at each stage of the operation, some x X, such that Eχ( x, z) = max Eχ(x, z), x X if the maximum is attained on X, or the corresponding ε-optimal strategy, otherwise. It is not difficult to see that for every x X the value (5) defines the probability of exceeding of the given level u 0. At the same time, it worth noting that although the result (6) is a particular case of (2), it is significantly important for applications. Let us now assume that the level u is not fixed. Then χ(x, z, u) = { 1, if F (x, z) u 0, if F (x, z) < u, (7) Denote Φ( x; z; u) = lim inf N and consider the functional 1 N N χ(x i, z i, u) i=1 Ω(u) = sup inf Φ( x; z; u). x X z Z 28

29 Suppose that in the extended operation the operating side is interested in maximization of the value u, subject to the constraint Ω(u) p, that is solving the following optimization problem { u sup, Ω(u) p (8), for p (0, 1), where p is a given probability. Let u be the solution to problem (8). Then, if we denote U = {u Ω(u) p }, we obtain that u = sup U. Therefore, under these conditions we have the following theorem: Theorem 7: where v(x) = u = sup v(x), x X sup inf F (x, z). S Z:P (S) p z S It is not difficult to see that, in the extended operation, solving problem (8) is equivalent to the maximization of the value of u, subject to the condition that supremum with respect to x X of the probability of exceeding of the level u by F (x, z) is not less than the given p, for z Z. So, it is clear that in this case the operating side is maximizing the level with the given guarantee. Under these circumstances, to evaluate the efficiency of strategy x it is possible to use the function v(x), and the best guaranteed result is equal to sup x X v(x). At the same time, the optimization problem v(x) sup x X is equivalent to the problem of finding sup sup S Ξ:P (S) p x X ω 29

30 subject to the constraints ω F (x, z j ), j J S, where J S = { j 1, z j S }. Note that the maximin problem has been reduced to a constrained maximization problem. 30

31 6 Applications to two-stage stochastic programming problem Let the profit of an enterprise from realization of strategy (plan) x and its correction r for every value of the stochastic factor z Z is determined by the value of the function F (x, r, z), (9) and plan x and its correction r for all z Z must satisfy the constraints: f k (x, r, z) 0, k = 1, m (10) x X, r R, where x = (x 1, x 2,..., x n ) and r = (r 1, r 2,..., r n ) are strategies of the enterprise and z is an uncontrollable stochastic factor. Denote by r(x, z) the correction r that maximizes (9), subject to constraints (10), for given values of x and z. Then the anticipated (expected) profit from the realization of plan x and correction r(x, z) will be determined as EF (x, r(x, z), z). (11) The problem is reduced to the finding of x X, which maximizes function (11), i.e. to the following optimization problem EF (x, r(x, z), z) sup. (12) x X 31

32 The given stochastic programming problem can be considered as a Zero Sum Game, where Player 1 (the enterprise) is trying to maximize F (x, r(x, z), z) by choosing x X, while Player 2 (Nature) is trying to minimize F (x, r(x, z), z) by choosing z Z. Consider this stochastic programming problem under the assumption of frequency definition of probability about probabilistic characteristics of stochastic factor z. Assume that set Z is at most countable. Let J = { j 1, z j Z }. Suppose that the operation of choosing of x may be repeated infinitely, and on the each step i, when the enterprise makes its choice x i, there appears some z i = z j Z, such that j J p j 1 j J, lim N N N i=1 χ j(z i ) = p j, with = 1, where N is the current number of repetitions of the operation. If the enterprise intends to carry out the operation just once and is interested on the result of the one-step operation, its best guaranteed result is determined as sup inf F (x, r(x, z), z). x X z Z On the other hand, if the enterprise agreed to transfer from the one-step operation to the infinitely-repeated extended operation with the following objective function F ( x; z) = lim inf N 1 N N F (x i, r(x i, z i ), z i ), i=1 where x X, z Z, and sets X and Z are defined as in the previous section, then applying Theorem 6, we find that the best guaranteed result is sup inf x X z Z F ( x; z) = sup EF (x, r(x, z), z). x X Hence, the best guaranteed result of the enterprise in the extended operation is determined as the optimal objective value of problem (12). Further, if in the two-stage problem the enterprise wants to maximize the probability of exceeding the given level, it may construct an extended operation and formulate a problem similar to optimization problem (6). Then the solution to problem (6) will determine the best guaranteed result and optimal strategy in the extended operation. Finally, if in the two-stage problem the enterprise is interested in maximization of the level with the given guarantee, it may transfer to an 32

33 extended operation and formulate an optimization problem similar to (8). Then correspondingly, the solution to problem (8) will determine the best guaranteed result and optimal strategy in the extended operation. Let us now assume that only the constraints of the problem contain the stochastic factor z. Enterprise s strategies, as before, are x and r; and suppose that constraints for x are known, but constrains for r depend on z, for instance are given in the form r R(x, z). For simplicity assume that for every z Z, max r R(x,z) F (x, r) is attained. Then, if the enterprise intends to carry out just one-step operation, the best guaranteed result will be determined as sup inf x X z Z max F (x, r). r R(x,z) However, if a stochastic information about z is known and the enterprise intends to use it by the transformation to extended operation (for example, with objective function of type (1)), it will be need to solve the optimization problem sup x X j J p j max F (x, r). r R(x,z j ) Additionally, when the original problem besides r R(x, z) contains also a constraint g(x, z) 0, denote R (x, z) = {r r R(x, z), g(x, z) 0}, and suppose, for simplicity that R (x, z), x X. Then in the original one-step operation the best guaranteed result will be sup inf max F (x, r). x X z Z r R (x,z) However, once the enterprise transfers to the extended operation, it will have to solve the optimization problem p j F (x, r j ), sup x X sup r R(x) j J where R(x) = { r = (r j ) j J r j R(x, z j ), } p j g(x, r j ) 0, j J 33

34 and the optimal behavior at the second stage of the two-stage stochastic programming problem is defined for all realizations of z at the same time. Although the problem is getting complicated, the operating side might obtain a substantial increment in the value of the operation. 7 What can be done further: There is a lot of room for development and generalization of the suggested approach. A particular interest represents an extension of the suggested method on multi-step operations with stochastic factors. Indeed, it is impossible to enter the same river twice. Additional research also needs to be conducted for the case when the set of values of a stochastic factor is uncountable. 34

35 References [1] Germeier, Yu. B., Non-Antagonistic Games, D. Reidel Publishing Company, Boston, [2] Germeier, Yu. B., On two-person games with a fixed turn sequence, The Annals of the Academy of Sciences of the USSR 198, 5,1971, p [3] Kukushkin, N. S., The importance of a mutual information exchange between the sides in two-person games with non-antagonistic interests, USSR Computational Mathematics and Mathematical Physics 1972, vol 4, p [4] Raguimov, I. S., On decision making in operations with stochastic factors, In the proceedings of the 2001 International Conference on Computational Intelligence for Modeling Control and Automation- CIMCA 2001 /ISBN , p [5] Vasin, A. A., Krasnoshokov P.S., and Morozov V. V. Operations Research, Moscow: Academia, (in Russian) 35

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin Game Theory Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin Bimatrix Games We are given two real m n matrices A = (a ij ), B = (b ij

More information

Satisfaction Equilibrium: Achieving Cooperation in Incomplete Information Games

Satisfaction Equilibrium: Achieving Cooperation in Incomplete Information Games Satisfaction Equilibrium: Achieving Cooperation in Incomplete Information Games Stéphane Ross and Brahim Chaib-draa Department of Computer Science and Software Engineering Laval University, Québec (Qc),

More information

Computational Game Theory Spring Semester, 2005/6. Lecturer: Yishay Mansour Scribe: Ilan Cohen, Natan Rubin, Ophir Bleiberg*

Computational Game Theory Spring Semester, 2005/6. Lecturer: Yishay Mansour Scribe: Ilan Cohen, Natan Rubin, Ophir Bleiberg* Computational Game Theory Spring Semester, 2005/6 Lecture 5: 2-Player Zero Sum Games Lecturer: Yishay Mansour Scribe: Ilan Cohen, Natan Rubin, Ophir Bleiberg* 1 5.1 2-Player Zero Sum Games In this lecture

More information

Strategies under Strategic Uncertainty

Strategies under Strategic Uncertainty Discussion Paper No. 18-055 Strategies under Strategic Uncertainty Helene Mass Discussion Paper No. 18-055 Strategies under Strategic Uncertainty Helene Mass Download this ZEW Discussion Paper from our

More information

A Note on the Existence of Ratifiable Acts

A Note on the Existence of Ratifiable Acts A Note on the Existence of Ratifiable Acts Joseph Y. Halpern Cornell University Computer Science Department Ithaca, NY 14853 halpern@cs.cornell.edu http://www.cs.cornell.edu/home/halpern August 15, 2018

More information

Game Theory and its Applications to Networks - Part I: Strict Competition

Game Theory and its Applications to Networks - Part I: Strict Competition Game Theory and its Applications to Networks - Part I: Strict Competition Corinne Touati Master ENS Lyon, Fall 200 What is Game Theory and what is it for? Definition (Roger Myerson, Game Theory, Analysis

More information

Total Expected Discounted Reward MDPs: Existence of Optimal Policies

Total Expected Discounted Reward MDPs: Existence of Optimal Policies Total Expected Discounted Reward MDPs: Existence of Optimal Policies Eugene A. Feinberg Department of Applied Mathematics and Statistics State University of New York at Stony Brook Stony Brook, NY 11794-3600

More information

DECISIONS UNDER UNCERTAINTY

DECISIONS UNDER UNCERTAINTY August 18, 2003 Aanund Hylland: # DECISIONS UNDER UNCERTAINTY Standard theory and alternatives 1. Introduction Individual decision making under uncertainty can be characterized as follows: The decision

More information

Preliminary Results on Social Learning with Partial Observations

Preliminary Results on Social Learning with Partial Observations Preliminary Results on Social Learning with Partial Observations Ilan Lobel, Daron Acemoglu, Munther Dahleh and Asuman Ozdaglar ABSTRACT We study a model of social learning with partial observations from

More information

Lecture 12 November 3

Lecture 12 November 3 STATS 300A: Theory of Statistics Fall 2015 Lecture 12 November 3 Lecturer: Lester Mackey Scribe: Jae Hyuck Park, Christian Fong Warning: These notes may contain factual and/or typographic errors. 12.1

More information

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem

Algorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem Algorithmic Game Theory and Applications Lecture 4: 2-player zero-sum games, and the Minimax Theorem Kousha Etessami 2-person zero-sum games A finite 2-person zero-sum (2p-zs) strategic game Γ, is a strategic

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo January 29, 2012 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

1: PROBABILITY REVIEW

1: PROBABILITY REVIEW 1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following

More information

Lecture Notes 1: Decisions and Data. In these notes, I describe some basic ideas in decision theory. theory is constructed from

Lecture Notes 1: Decisions and Data. In these notes, I describe some basic ideas in decision theory. theory is constructed from Topics in Data Analysis Steven N. Durlauf University of Wisconsin Lecture Notes : Decisions and Data In these notes, I describe some basic ideas in decision theory. theory is constructed from The Data:

More information

Decision Analysis. An insightful study of the decision-making process under uncertainty and risk

Decision Analysis. An insightful study of the decision-making process under uncertainty and risk Decision Analysis An insightful study of the decision-making process under uncertainty and risk Decision Theory A philosophy, articulated by a set of logical axioms and a methodology, for analyzing the

More information

Quitting games - An Example

Quitting games - An Example Quitting games - An Example E. Solan 1 and N. Vieille 2 January 22, 2001 Abstract Quitting games are n-player sequential games in which, at any stage, each player has the choice between continuing and

More information

Near-Potential Games: Geometry and Dynamics

Near-Potential Games: Geometry and Dynamics Near-Potential Games: Geometry and Dynamics Ozan Candogan, Asuman Ozdaglar and Pablo A. Parrilo September 6, 2011 Abstract Potential games are a special class of games for which many adaptive user dynamics

More information

AN EXTENSION OF THE PROBABILITY LOGIC LP P 2. Tatjana Stojanović 1, Ana Kaplarević-Mališić 1 and Zoran Ognjanović 2

AN EXTENSION OF THE PROBABILITY LOGIC LP P 2. Tatjana Stojanović 1, Ana Kaplarević-Mališić 1 and Zoran Ognjanović 2 45 Kragujevac J. Math. 33 (2010) 45 62. AN EXTENSION OF THE PROBABILITY LOGIC LP P 2 Tatjana Stojanović 1, Ana Kaplarević-Mališić 1 and Zoran Ognjanović 2 1 University of Kragujevac, Faculty of Science,

More information

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1 EC9A0: Pre-sessional Advanced Mathematics Course Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1 1 Infimum and Supremum Definition 1. Fix a set Y R. A number α R is an upper bound of Y if

More information

Economics 201B Economic Theory (Spring 2017) Bargaining. Topics: the axiomatic approach (OR 15) and the strategic approach (OR 7).

Economics 201B Economic Theory (Spring 2017) Bargaining. Topics: the axiomatic approach (OR 15) and the strategic approach (OR 7). Economics 201B Economic Theory (Spring 2017) Bargaining Topics: the axiomatic approach (OR 15) and the strategic approach (OR 7). The axiomatic approach (OR 15) Nash s (1950) work is the starting point

More information

Interval values for strategic games in which players cooperate

Interval values for strategic games in which players cooperate Interval values for strategic games in which players cooperate Luisa Carpente 1 Balbina Casas-Méndez 2 Ignacio García-Jurado 2 Anne van den Nouweland 3 September 22, 2005 Abstract In this paper we propose

More information

Probability theory basics

Probability theory basics Probability theory basics Michael Franke Basics of probability theory: axiomatic definition, interpretation, joint distributions, marginalization, conditional probability & Bayes rule. Random variables:

More information

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad

More information

1 Stochastic Dynamic Programming

1 Stochastic Dynamic Programming 1 Stochastic Dynamic Programming Formally, a stochastic dynamic program has the same components as a deterministic one; the only modification is to the state transition equation. When events in the future

More information

Game Theory- Normal Form Games

Game Theory- Normal Form Games Chapter 6 Game Theory- Normal Form Games Key words: Game theory, strategy, finite game, zero-sum game, pay-off matrix, dominant strategy, value of the game, fair game, stable solution, saddle point, pure

More information

Three contrasts between two senses of coherence

Three contrasts between two senses of coherence Three contrasts between two senses of coherence Teddy Seidenfeld Joint work with M.J.Schervish and J.B.Kadane Statistics, CMU Call an agent s choices coherent when they respect simple dominance relative

More information

THE STUDY OF THE SET OF NASH EQUILIBRIUM POINTS FOR ONE TWO-PLAYERS GAME WITH QUADRATIC PAYOFF FUNCTIONS

THE STUDY OF THE SET OF NASH EQUILIBRIUM POINTS FOR ONE TWO-PLAYERS GAME WITH QUADRATIC PAYOFF FUNCTIONS Yugoslav Journal of Operations Research 5 015 Number 91 97 DOI: 10.98/YJOR13091701N THE STUDY OF THE SET OF NASH EQUILIBRIUM POINTS FOR ONE TWO-PLAYERS GAME WITH QUADRATIC PAYOFF FUNCTIONS Mikhail Sergeevich

More information

HOMEWORK ASSIGNMENT 6

HOMEWORK ASSIGNMENT 6 HOMEWORK ASSIGNMENT 6 DUE 15 MARCH, 2016 1) Suppose f, g : A R are uniformly continuous on A. Show that f + g is uniformly continuous on A. Solution First we note: In order to show that f + g is uniformly

More information

The Game of Normal Numbers

The Game of Normal Numbers The Game of Normal Numbers Ehud Lehrer September 4, 2003 Abstract We introduce a two-player game where at each period one player, say, Player 2, chooses a distribution and the other player, Player 1, a

More information

Economics 204 Fall 2011 Problem Set 1 Suggested Solutions

Economics 204 Fall 2011 Problem Set 1 Suggested Solutions Economics 204 Fall 2011 Problem Set 1 Suggested Solutions 1. Suppose k is a positive integer. Use induction to prove the following two statements. (a) For all n N 0, the inequality (k 2 + n)! k 2n holds.

More information

In N we can do addition, but in order to do subtraction we need to extend N to the integers

In N we can do addition, but in order to do subtraction we need to extend N to the integers Chapter The Real Numbers.. Some Preliminaries Discussion: The Irrationality of 2. We begin with the natural numbers N = {, 2, 3, }. In N we can do addition, but in order to do subtraction we need to extend

More information

Sequential Decisions

Sequential Decisions Sequential Decisions A Basic Theorem of (Bayesian) Expected Utility Theory: If you can postpone a terminal decision in order to observe, cost free, an experiment whose outcome might change your terminal

More information

Lecture notes on statistical decision theory Econ 2110, fall 2013

Lecture notes on statistical decision theory Econ 2110, fall 2013 Lecture notes on statistical decision theory Econ 2110, fall 2013 Maximilian Kasy March 10, 2014 These lecture notes are roughly based on Robert, C. (2007). The Bayesian choice: from decision-theoretic

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

1 Basic Game Modelling

1 Basic Game Modelling Max-Planck-Institut für Informatik, Winter 2017 Advanced Topic Course Algorithmic Game Theory, Mechanism Design & Computational Economics Lecturer: CHEUNG, Yun Kuen (Marco) Lecture 1: Basic Game Modelling,

More information

Introduction to Game Theory

Introduction to Game Theory COMP323 Introduction to Computational Game Theory Introduction to Game Theory Paul G. Spirakis Department of Computer Science University of Liverpool Paul G. Spirakis (U. Liverpool) Introduction to Game

More information

Information, Utility & Bounded Rationality

Information, Utility & Bounded Rationality Information, Utility & Bounded Rationality Pedro A. Ortega and Daniel A. Braun Department of Engineering, University of Cambridge Trumpington Street, Cambridge, CB2 PZ, UK {dab54,pao32}@cam.ac.uk Abstract.

More information

Convex analysis and profit/cost/support functions

Convex analysis and profit/cost/support functions Division of the Humanities and Social Sciences Convex analysis and profit/cost/support functions KC Border October 2004 Revised January 2009 Let A be a subset of R m Convex analysts may give one of two

More information

Lower Semicontinuity of the Solution Set Mapping in Some Optimization Problems

Lower Semicontinuity of the Solution Set Mapping in Some Optimization Problems Lower Semicontinuity of the Solution Set Mapping in Some Optimization Problems Roberto Lucchetti coauthors: A. Daniilidis, M.A. Goberna, M.A. López, Y. Viossat Dipartimento di Matematica, Politecnico di

More information

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable) Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f :

More information

Lecture 6: April 25, 2006

Lecture 6: April 25, 2006 Computational Game Theory Spring Semester, 2005/06 Lecture 6: April 25, 2006 Lecturer: Yishay Mansour Scribe: Lior Gavish, Andrey Stolyarenko, Asaph Arnon Partially based on scribe by Nataly Sharkov and

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

MATH31011/MATH41011/MATH61011: FOURIER ANALYSIS AND LEBESGUE INTEGRATION. Chapter 2: Countability and Cantor Sets

MATH31011/MATH41011/MATH61011: FOURIER ANALYSIS AND LEBESGUE INTEGRATION. Chapter 2: Countability and Cantor Sets MATH31011/MATH41011/MATH61011: FOURIER ANALYSIS AND LEBESGUE INTEGRATION Chapter 2: Countability and Cantor Sets Countable and Uncountable Sets The concept of countability will be important in this course

More information

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes Ellida M. Khazen * 13395 Coppermine Rd. Apartment 410 Herndon VA 20171 USA Abstract

More information

For general queries, contact

For general queries, contact PART I INTRODUCTION LECTURE Noncooperative Games This lecture uses several examples to introduce the key principles of noncooperative game theory Elements of a Game Cooperative vs Noncooperative Games:

More information

Computation of Efficient Nash Equilibria for experimental economic games

Computation of Efficient Nash Equilibria for experimental economic games International Journal of Mathematics and Soft Computing Vol.5, No.2 (2015), 197-212. ISSN Print : 2249-3328 ISSN Online: 2319-5215 Computation of Efficient Nash Equilibria for experimental economic games

More information

On maxitive integration

On maxitive integration On maxitive integration Marco E. G. V. Cattaneo Department of Mathematics, University of Hull m.cattaneo@hull.ac.uk Abstract A functional is said to be maxitive if it commutes with the (pointwise supremum

More information

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden 1 Selecting Efficient Correlated Equilibria Through Distributed Learning Jason R. Marden Abstract A learning rule is completely uncoupled if each player s behavior is conditioned only on his own realized

More information

Informed Principal in Private-Value Environments

Informed Principal in Private-Value Environments Informed Principal in Private-Value Environments Tymofiy Mylovanov Thomas Tröger University of Bonn June 21, 2008 1/28 Motivation 2/28 Motivation In most applications of mechanism design, the proposer

More information

Stochastic dominance with imprecise information

Stochastic dominance with imprecise information Stochastic dominance with imprecise information Ignacio Montes, Enrique Miranda, Susana Montes University of Oviedo, Dep. of Statistics and Operations Research. Abstract Stochastic dominance, which is

More information

Zero sum games Proving the vn theorem. Zero sum games. Roberto Lucchetti. Politecnico di Milano

Zero sum games Proving the vn theorem. Zero sum games. Roberto Lucchetti. Politecnico di Milano Politecnico di Milano General form Definition A two player zero sum game in strategic form is the triplet (X, Y, f : X Y R) f (x, y) is what Pl1 gets from Pl2, when they play x, y respectively. Thus g

More information

1 Definition of the Riemann integral

1 Definition of the Riemann integral MAT337H1, Introduction to Real Analysis: notes on Riemann integration 1 Definition of the Riemann integral Definition 1.1. Let [a, b] R be a closed interval. A partition P of [a, b] is a finite set of

More information

A remark on discontinuous games with asymmetric information and ambiguity

A remark on discontinuous games with asymmetric information and ambiguity Econ Theory Bull DOI 10.1007/s40505-016-0100-5 RESEARCH ARTICLE A remark on discontinuous games with asymmetric information and ambiguity Wei He 1 Nicholas C. Yannelis 1 Received: 7 February 2016 / Accepted:

More information

Payoff Continuity in Incomplete Information Games

Payoff Continuity in Incomplete Information Games journal of economic theory 82, 267276 (1998) article no. ET982418 Payoff Continuity in Incomplete Information Games Atsushi Kajii* Institute of Policy and Planning Sciences, University of Tsukuba, 1-1-1

More information

Bargaining Efficiency and the Repeated Prisoners Dilemma. Bhaskar Chakravorti* and John Conley**

Bargaining Efficiency and the Repeated Prisoners Dilemma. Bhaskar Chakravorti* and John Conley** Bargaining Efficiency and the Repeated Prisoners Dilemma Bhaskar Chakravorti* and John Conley** Published as: Bhaskar Chakravorti and John P. Conley (2004) Bargaining Efficiency and the repeated Prisoners

More information

Lebesgue measure and integration

Lebesgue measure and integration Chapter 4 Lebesgue measure and integration If you look back at what you have learned in your earlier mathematics courses, you will definitely recall a lot about area and volume from the simple formulas

More information

Practicable Robust Markov Decision Processes

Practicable Robust Markov Decision Processes Practicable Robust Markov Decision Processes Huan Xu Department of Mechanical Engineering National University of Singapore Joint work with Shiau-Hong Lim (IBM), Shie Mannor (Techion), Ofir Mebel (Apple)

More information

Imitation and Minimax Regret 1

Imitation and Minimax Regret 1 Imitation and Minimax Regret 1 Karl H. Schlag December 3 4 1 This rst version of this paper was written for an imitation workshop at UL, London 5.1.3. Economics Department, European University Institute,

More information

Could Nash equilibria exist if the payoff functions are not quasi-concave?

Could Nash equilibria exist if the payoff functions are not quasi-concave? Could Nash equilibria exist if the payoff functions are not quasi-concave? (Very preliminary version) Bich philippe Abstract In a recent but well known paper (see [11]), Reny has proved the existence of

More information

Mean-Payoff Games and the Max-Atom Problem

Mean-Payoff Games and the Max-Atom Problem Mean-Payoff Games and the Max-Atom Problem Albert Atserias Universitat Politècnica de Catalunya Barcelona, Spain Elitza Maneva Universitat Politècnica de Catalunya Barcelona, Spain February 3, 200 Abstract

More information

On Decision Making under Interval Uncertainty: A New Justification of Hurwicz Optimism-Pessimism Approach and Its Use in Group Decision Making

On Decision Making under Interval Uncertainty: A New Justification of Hurwicz Optimism-Pessimism Approach and Its Use in Group Decision Making On Decision Making under Interval Uncertainty: A New Justification of Hurwicz Optimism-Pessimism Approach and Its Use in Group Decision Making Van Nam Huynh 1, Chenyi Hu, Yoshiteru Nakamori 1, and Vladik

More information

Area I: Contract Theory Question (Econ 206)

Area I: Contract Theory Question (Econ 206) Theory Field Exam Summer 2011 Instructions You must complete two of the four areas (the areas being (I) contract theory, (II) game theory A, (III) game theory B, and (IV) psychology & economics). Be sure

More information

1 Lattices and Tarski s Theorem

1 Lattices and Tarski s Theorem MS&E 336 Lecture 8: Supermodular games Ramesh Johari April 30, 2007 In this lecture, we develop the theory of supermodular games; key references are the papers of Topkis [7], Vives [8], and Milgrom and

More information

Recursive Methods Recursive Methods Nr. 1

Recursive Methods Recursive Methods Nr. 1 Nr. 1 Outline Today s Lecture Dynamic Programming under Uncertainty notation of sequence problem leave study of dynamics for next week Dynamic Recursive Games: Abreu-Pearce-Stachetti Application: today

More information

Decision theory. 1 We may also consider randomized decision rules, where δ maps observed data D to a probability distribution over

Decision theory. 1 We may also consider randomized decision rules, where δ maps observed data D to a probability distribution over Point estimation Suppose we are interested in the value of a parameter θ, for example the unknown bias of a coin. We have already seen how one may use the Bayesian method to reason about θ; namely, we

More information

The sample complexity of agnostic learning with deterministic labels

The sample complexity of agnostic learning with deterministic labels The sample complexity of agnostic learning with deterministic labels Shai Ben-David Cheriton School of Computer Science University of Waterloo Waterloo, ON, N2L 3G CANADA shai@uwaterloo.ca Ruth Urner College

More information

The Value of Information in Zero-Sum Games

The Value of Information in Zero-Sum Games The Value of Information in Zero-Sum Games Olivier Gossner THEMA, Université Paris Nanterre and CORE, Université Catholique de Louvain joint work with Jean-François Mertens July 2001 Abstract: It is not

More information

Robust Learning Equilibrium

Robust Learning Equilibrium Robust Learning Equilibrium Itai Ashlagi Dov Monderer Moshe Tennenholtz Faculty of Industrial Engineering and Management Technion Israel Institute of Technology Haifa 32000, Israel Abstract We introduce

More information

Chapter 9. Mixed Extensions. 9.1 Mixed strategies

Chapter 9. Mixed Extensions. 9.1 Mixed strategies Chapter 9 Mixed Extensions We now study a special case of infinite strategic games that are obtained in a canonic way from the finite games, by allowing mixed strategies. Below [0, 1] stands for the real

More information

Mechanism Design: Basic Concepts

Mechanism Design: Basic Concepts Advanced Microeconomic Theory: Economics 521b Spring 2011 Juuso Välimäki Mechanism Design: Basic Concepts The setup is similar to that of a Bayesian game. The ingredients are: 1. Set of players, i {1,

More information

Chapter 4. Measure Theory. 1. Measure Spaces

Chapter 4. Measure Theory. 1. Measure Spaces Chapter 4. Measure Theory 1. Measure Spaces Let X be a nonempty set. A collection S of subsets of X is said to be an algebra on X if S has the following properties: 1. X S; 2. if A S, then A c S; 3. if

More information

Entropy Games and Matrix Multiplication Games. EQINOCS seminar

Entropy Games and Matrix Multiplication Games. EQINOCS seminar Entropy Games and Matrix Multiplication Games Eugene Asarin Julien Cervelle Aldric Degorre Cătălin Dima Florian Horn Victor Kozyakin IRIF, LACL, IITP EQINOCS seminar 2016-05-11 A game of freedom The story

More information

Area I: Contract Theory Question (Econ 206)

Area I: Contract Theory Question (Econ 206) Theory Field Exam Winter 2011 Instructions You must complete two of the three areas (the areas being (I) contract theory, (II) game theory, and (III) psychology & economics). Be sure to indicate clearly

More information

Cowles Foundation for Research in Economics at Yale University

Cowles Foundation for Research in Economics at Yale University Cowles Foundation for Research in Economics at Yale University Cowles Foundation Discussion Paper No. 1904 Afriat from MaxMin John D. Geanakoplos August 2013 An author index to the working papers in the

More information

Tijmen Daniëls Universiteit van Amsterdam. Abstract

Tijmen Daniëls Universiteit van Amsterdam. Abstract Pure strategy dominance with quasiconcave utility functions Tijmen Daniëls Universiteit van Amsterdam Abstract By a result of Pearce (1984), in a finite strategic form game, the set of a player's serially

More information

In N we can do addition, but in order to do subtraction we need to extend N to the integers

In N we can do addition, but in order to do subtraction we need to extend N to the integers Chapter 1 The Real Numbers 1.1. Some Preliminaries Discussion: The Irrationality of 2. We begin with the natural numbers N = {1, 2, 3, }. In N we can do addition, but in order to do subtraction we need

More information

Change-point models and performance measures for sequential change detection

Change-point models and performance measures for sequential change detection Change-point models and performance measures for sequential change detection Department of Electrical and Computer Engineering, University of Patras, 26500 Rion, Greece moustaki@upatras.gr George V. Moustakides

More information

Birgit Rudloff Operations Research and Financial Engineering, Princeton University

Birgit Rudloff Operations Research and Financial Engineering, Princeton University TIME CONSISTENT RISK AVERSE DYNAMIC DECISION MODELS: AN ECONOMIC INTERPRETATION Birgit Rudloff Operations Research and Financial Engineering, Princeton University brudloff@princeton.edu Alexandre Street

More information

Robust Partially Observable Markov Decision Processes

Robust Partially Observable Markov Decision Processes Submitted to manuscript Robust Partially Observable Markov Decision Processes Mohammad Rasouli 1, Soroush Saghafian 2 1 Management Science and Engineering, Stanford University, Palo Alto, CA 2 Harvard

More information

Bayesian Persuasion Online Appendix

Bayesian Persuasion Online Appendix Bayesian Persuasion Online Appendix Emir Kamenica and Matthew Gentzkow University of Chicago June 2010 1 Persuasion mechanisms In this paper we study a particular game where Sender chooses a signal π whose

More information

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define Homework, Real Analysis I, Fall, 2010. (1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define ρ(f, g) = 1 0 f(x) g(x) dx. Show that

More information

The Pre-nucleolus for Fuzzy Cooperative Games

The Pre-nucleolus for Fuzzy Cooperative Games Journal of Game Theory 207, 6(2): 43-5 DOI: 05923/jjgt207060203 The Pre-nucleolus for Fuzzy Cooperative Games Yeremia Maroutian 39, E98 Str Broolyn, NY 236 Abstract In this paper invented by D Schmeidler

More information

Chapter 3 Continuous Functions

Chapter 3 Continuous Functions Continuity is a very important concept in analysis. The tool that we shall use to study continuity will be sequences. There are important results concerning the subsets of the real numbers and the continuity

More information

Chapter 5. Measurable Functions

Chapter 5. Measurable Functions Chapter 5. Measurable Functions 1. Measurable Functions Let X be a nonempty set, and let S be a σ-algebra of subsets of X. Then (X, S) is a measurable space. A subset E of X is said to be measurable if

More information

DO FIVE OUT OF SIX ON EACH SET PROBLEM SET

DO FIVE OUT OF SIX ON EACH SET PROBLEM SET DO FIVE OUT OF SIX ON EACH SET PROBLEM SET 1. THE AXIOM OF FOUNDATION Early on in the book (page 6) it is indicated that throughout the formal development set is going to mean pure set, or set whose elements,

More information

Econ 2148, spring 2019 Statistical decision theory

Econ 2148, spring 2019 Statistical decision theory Econ 2148, spring 2019 Statistical decision theory Maximilian Kasy Department of Economics, Harvard University 1 / 53 Takeaways for this part of class 1. A general framework to think about what makes a

More information

Coherent Choice Functions Under Uncertainty* OUTLINE

Coherent Choice Functions Under Uncertainty* OUTLINE Coherent Choice Functions Under Uncertainty* Teddy Seidenfeld joint work with Jay Kadane and Mark Schervish Carnegie Mellon University OUTLINE 1. Preliminaries a. Coherent choice functions b. The framework

More information

The Folk Theorem for Finitely Repeated Games with Mixed Strategies

The Folk Theorem for Finitely Repeated Games with Mixed Strategies The Folk Theorem for Finitely Repeated Games with Mixed Strategies Olivier Gossner February 1994 Revised Version Abstract This paper proves a Folk Theorem for finitely repeated games with mixed strategies.

More information

Recitation 7: Uncertainty. Xincheng Qiu

Recitation 7: Uncertainty. Xincheng Qiu Econ 701A Fall 2018 University of Pennsylvania Recitation 7: Uncertainty Xincheng Qiu (qiux@sas.upenn.edu 1 Expected Utility Remark 1. Primitives: in the basic consumer theory, a preference relation is

More information

Towards Decision Making under Interval Uncertainty

Towards Decision Making under Interval Uncertainty Journal of Uncertain Systems Vol.1, No.3, pp.00-07, 018 Online at: www.us.org.uk Towards Decision Making under Interval Uncertainty Andrze Pownuk, Vladik Kreinovich Computational Science Program, University

More information

REPEATED GAMES. Jörgen Weibull. April 13, 2010

REPEATED GAMES. Jörgen Weibull. April 13, 2010 REPEATED GAMES Jörgen Weibull April 13, 2010 Q1: Can repetition induce cooperation? Peace and war Oligopolistic collusion Cooperation in the tragedy of the commons Q2: Can a game be repeated? Game protocols

More information

REVIEW OF ESSENTIAL MATH 346 TOPICS

REVIEW OF ESSENTIAL MATH 346 TOPICS REVIEW OF ESSENTIAL MATH 346 TOPICS 1. AXIOMATIC STRUCTURE OF R Doğan Çömez The real number system is a complete ordered field, i.e., it is a set R which is endowed with addition and multiplication operations

More information

1 Markov decision processes

1 Markov decision processes 2.997 Decision-Making in Large-Scale Systems February 4 MI, Spring 2004 Handout #1 Lecture Note 1 1 Markov decision processes In this class we will study discrete-time stochastic systems. We can describe

More information

Optimism in the Face of Uncertainty Should be Refutable

Optimism in the Face of Uncertainty Should be Refutable Optimism in the Face of Uncertainty Should be Refutable Ronald ORTNER Montanuniversität Leoben Department Mathematik und Informationstechnolgie Franz-Josef-Strasse 18, 8700 Leoben, Austria, Phone number:

More information

Persuading a Pessimist

Persuading a Pessimist Persuading a Pessimist Afshin Nikzad PRELIMINARY DRAFT Abstract While in practice most persuasion mechanisms have a simple structure, optimal signals in the Bayesian persuasion framework may not be so.

More information

Principal-Agent Games - Equilibria under Asymmetric Information -

Principal-Agent Games - Equilibria under Asymmetric Information - Principal-Agent Games - Equilibria under Asymmetric Information - Ulrich Horst 1 Humboldt-Universität zu Berlin Department of Mathematics and School of Business and Economics Work in progress - Comments

More information

On the Existence of Price Equilibrium in Economies with Excess Demand Functions

On the Existence of Price Equilibrium in Economies with Excess Demand Functions On the Existence of Price Equilibrium in Economies with Excess Demand Functions Guoqiang Tian Abstract This paper provides a price equilibrium existence theorem in economies where commodities may be indivisible

More information

Convex Functions and Optimization

Convex Functions and Optimization Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized

More information

Cooperative assignment games with the inverse Monge property

Cooperative assignment games with the inverse Monge property Cooperative assignment games with the inverse Monge property F. Javier Martínez-de-Albéniz a,1, Carles Rafels a a Dep. de Matemàtica Econòmica, Financera i Actuarial Universitat de Barcelona Av. Diagonal

More information

Ergodicity and Non-Ergodicity in Economics

Ergodicity and Non-Ergodicity in Economics Abstract An stochastic system is called ergodic if it tends in probability to a limiting form that is independent of the initial conditions. Breakdown of ergodicity gives rise to path dependence. We illustrate

More information