Maximum Process Problems in Optimal Control Theory

Size: px
Start display at page:

Download "Maximum Process Problems in Optimal Control Theory"

Transcription

1 J. Appl. Math. Stochastic Anal. Vol. 25, No., 25, (77-88) Research Report No. 423, 2, Dept. Theoret. Statist. Aarhus (2 pp) Maximum Process Problems in Optimal Control Theory GORAN PESKIR 3 Given a standard Brownian motion (B t ) t and the equation of motion: p dx t = v t dt + 2 db t we set S t = max st X s and consider the optimal control problem: sup E S c v where c > and the supremum is taken over all admissible controls v satisfying v t 2 [ ; ] for all t up to = inf f t > j X t =2 (`; `) g with < < and ` < < ` given and fixed. The following control v 3 is proved to be optimal: Pull as hard as possible that is v 3 = t if X t < g3(s t ), and push as hard as possible that is v 3 = t if X t > g3(s t ), where s 7! g3(s) is a switching curve that is determined explicitly (as the unique solution to a nonlinear differential equation). The solution found demonstrates that the problem formulations based on a maximum functional can be successfully included in optimal control theory (calculus of variations) in addition to the classic problem formulations due to Lagrange, Mayer and Bolza.. Introduction Stochastic control theory deals with three basic problem formulations which were inherited from classical calculus of variations (cf. [4] pp ). Given the equation of motion: (.) dx t = (X t ; u t ) dt + (X t ; u t ) db t where (B t ) t is standard Brownian motion, consider the optimal control problem: (.2) inf E u xz D L(X t ; u t ) dt + M (X D ) where the infimum is taken over all admissible controls u = (u t ) t applied before the exit time D = inf f t > j X t =2 C g for some open set C = D c and the process (X t ) t starts at x under P x. If M and L 6=, the problem (.2) is said to be Lagrange formulated. If L and M 6=, the problem (.2) is said to be Mayer formulated. If both L 6= and M 6=, the problem (.2) is said to be Bolza formulated. The Lagrange formulation goes back to the 8th century, the Mayer formulation originated in the 9th century, and the Bolza formulation [2] was introduced in 93. We refer to [] (pp ) with the references for a historical account of the Lagrange, Mayer and Bolza problems. 3 Centre for Mathematical Physics and Stochastics (funded by the Danish National Research Foundation). Mathematics Subject Classification 2. Primary 49J55, 93E2, 6J65. Secondary 6J6, 34A34, 35G5. Key words and phrases: Optimal control theory, maximum process, the Hamilton-Jacobi-Bellman equation, bang-bang control, switching curve, diffusion, Brownian motion, nonlinear differential equation. goran@imf.au.dk

2 Although the three problem formulations are formally known to be equivalent (see e.g. [] pp or [4] pp ), this fact is rarely proved to be essential when solving a concrete problem. Setting Z t = L(X t ; u t ) or Z t = M (X t ), and focusing upon the sample path t 7! Z t for t 2 [; D ], we see that the three problem formulations measure the performance associated with a control u by means of the following two functionals: (.3) Z D Z t dt and Z D where the first one represents the surface area below (or above) the sample path, and the second one represents the sample-path s terminal value. In addition to these two functionals, it is suggested by elementary geometric considerations that the maximal value of the sample path: (.4) max Z t t D provides yet another performance measure which, to a certain extent, is more sensitive than the previous two ones. Clearly, a sample path can have a small integral but still a large maximum, while a large maximum cannot be detected by the terminal value either. The main purpose of the present paper is to show that the problem formulations based on a maximum functional can be successfully added to optimal control theory (calculus of variations). This is done by formulating a specific problem of this type (Section 2) and solving it in a closed form (Section 3). The result suggests a number of new avenues for further research upon extending the Bolza formulation (.2) to optimize the following expression: (.5) E x max t D K(X t ; u t ) +Z D L(X t ; u t ) dt + M (X D ) where some of the maps K ; L and M may also be identically zero. Optimal stopping problems for the maximum process have been studied by a number of authors in the 99 s (see e.g. [6, 3,, 8, 5]) and the subject seems to be well-understood now. 2. Formulation of the problem Consider a process X = (X t ) t (2.) dx t = v t dt + solving the stochastic differential equation: p 2 db t where B = (B t ) t is standard Brownian motion, and associate with X the maximum process: (2.2) S t = max X r rt so that X = x and S = s under P x;s where x s. Introduce the exit time: _ s (2.3) = inf f t > j X t =2 (`; `) g where ` < < ` are given and fixed, and let so be c > in the sequel. 2

3 The optimal control problem to be examined in this paper is formulated as follows: (2.4) J(x; s) := sup v E x;s S c where the supremum is taken over all admissible controls v satisfying v t 2 [ ; ] for all t with some < < given and fixed. By admissible we mean that the s.d.e. (2.) can be solved in Itô s sense (either strongly or weakly). Since v t is required to be uniformly bounded, it is well-known that a weak solution (unique in law) always exists under a measurability condition (see e.g. [9] p. 55) where v t may depend on the entire sample path r 7! X r up to time t. Moreover, if v t = v(x t ) for some (bounded) measurable function v, then a strong solution (pathwise unique) also exists (see e.g. [9] pp. 79-8). The optimal control problem (2.4) has some interesting interpretations. The equation (2.) may be viewed as describing the motion of a Brownian particle (subject to a fluctuating force B _ t ) that is under influence of a (slowly varying) external force v t (see [7] pp ). The objective in (2.4) is therefore to determine an optimum of the external force one needs to exert upon the particle so to make its maximal height at the time of exit as large as possible in the course of time needed for the same exit to happen as short as possible. Clearly, the interpretation and objective just described carry over to many other problems where the equation (2.) plays a role. It appears intuitively clear that the optimal control should be of the following bang-bang type: At each time either push or pull as hard as possible so to reach either ` or ` as soon as possible. The solution of the problem presented in the next section confirms this guess and makes the statement precise in analytic terms. It is also apparent that at each time t we need to keep track of both X t and S t so that the problem is inherently two-dimensional. 3. Solution of the problem. In the setting of the previous section consider the optimal control problem (2.4). Note that X t = (X t ; S t ) is a two-dimensional process with the state space S = f (x; s) 2 IR 2 j x s g that changes (increases) in the second coordinate only after hitting the diagonal x = s in IR 2. Off the diagonal, the process X = ( Xt ) t changes only in the first coordinate and thus may be identified with X. Moreover, if v t = v(x t ) for some (bounded) measurable function v in (2.), then X is a Markov process. The later feedback controls are sufficient to be considered under fairly general hypotheses (see e.g. [4] pp ), and this fact will also be proved below. The infinitesimal generator of X may be therefore formally described as follows: (3.) IL X = IL X in x < = at x = s where IL X is the infinitesimal generator of X. This means that the infinitesimal generator of X is acting on a space of C 2 functions f on S satisfying lim x"s (@f =@s)(x; s) =. The formal description (3.)+(3.2) appears in [8] (pp ) where the latter fact is also verified. The condition of normal reflection (3.2) was used for the first time in [3] in the case of Bessel processes (it was also noted in [6] (p. 8) in the case of a Bessel process of dimension one). 3

4 2. Assuming for a moment that the supremum in (2.4) is attained at some feedback control, and making use of the formal description of the infinitesimal generator (3.)+(3.2), we are naturally led to formulate the following HJB system: (3.3) sup IL vx (J) c = in C for x < s (x; s) = (normal reflection) x=s (3.5) J(x; s) = s (instantaneous stopping) x=`+ (3.6) J(x; s) = s (instantaneous stopping) x=` for ` < s < ` where the infinitesimal generator of X given v is expressed by (3.7) IL v X 2 and we set C = f (x; s) 2 IR 2 j ` < x s < ` g. Our main effort in the sequel will be directed to solving the system (3.3)-(3.6) in a closed form. More explicitly, the HJB equation (3.3) with J = J(x; s) reads as follows: (3.8) sup v v J x + J xx c = so that we may expect a bang-bang solution v 3 depending on the sign of J x : If J x < then v 3 t =, and if J x > then v 3 t =. The equation J x = defines an optimal switching curve s 7! g 3 (s), and the main task in the sequel will be to determine it explicitly. Further heuristic considerations based on the bang-bang principle just stated (when close to ` apply so to exit at `, and when close to ` apply so to exit at ` ) suggest to partition C into the following three subsets (modulo two curves x = g 3 (s) and s = s 3 to be found): (3.9) C = f (x; s) 2 IR 2 j ` < x < g 3 (s) ; s 3 < s < ` g (3.) C 2 = f (x; s) 2 IR 2 j g 3 (s) < x s ; s 3 < s < ` g (3.) C 3 = f (x; s) 2 IR 2 j ` < x s < s 3 g where s 3 is a unique point in (`; `) satisfying: (3.2) g 3 (s 3 ) = `. In addition to (3.) we also set: (3.3) g 3 (`) = x 3 to denote another point in (`; `) playing a role. We refer to Figure below to obtain a better geometric understanding of (3.9)-(3.3). 4

5 3. We construct a solution to the system (3.3)-(3.6) in three steps. In the first two steps we determine C [ C 2 together with a boundary curve s 7! g 3 (s) separating C from C 2. Step. Consider the HJB equation: (3.4) J x + J xx c = in C to be found. The general solution of (3.4) is given by (3.5) J(x; s) = a (s) e x + c x + b (s) where a (s) and b (s) are some undetermined functions of s. Using (3.5) we can eliminate b (s) from (3.5) and this yields: (3.6) J(x; s) = a (s) e x e ` + c (x `) + s. Solving J x (x; s) = for x gives x 3 = g 3 (s) as a candidate for the switching curve, and also that a (s) can be expressed in terms of g 3 (s) as follows: (3.7) a (s) = c Inserting this back into (3.6) gives: (3.8) J () (x; s) = c ( ) 2 e g3(s) e g3(s). e x e ` as a candidate for the value function (2.4) when (x; s) 2 C. Step 2. Consider the HJB equation: (3.9) J x + J xx c = in C 2 to be found. The general solution of (3.9) is given by (3.2) J(x; s) = a (s) e x + c x + b (s) + c (x `) + s where a (s) and b (s) are some undetermined functions of s. Solving J x (x; s) = for x gives x 3 = g 3 (s) as a candidate for the switching curve, and also that a (s) can be expressed in terms of g 3 (s) as follows: (3.2) a (s) = c Inserting this back into (3.2) gives: (3.22) J (2) (x; s) = c e g3(s). ( ) 2 e (g3(s)x) + c x + b (s). The two functions (3.8) and (3.22) must coincide at the switching curve, i.e. 5

6 (3.23) J () (x; s) x=g3(s)+ = J (2) (x; s) x=g3(s) giving a closed expression for b (s), which after inserted back into (3.22) yields: (3.24) J (2) (x; s) = c ( ) 2 e (g3(s)x) c ( ) 2 e (g3(s)`) c g3(s) x + c g3(s) ` + s + c c ( ) 2 ( ) 2 as a candidate for the value function (2.4) when (x; s) 2 C 2. The condition (3.6) with x = s = ` can now be used to determine a unique x3 2 (`; `) satisfying (3.3). It follows from the expression (3.24) above that x3 solves (and is uniquely determined by) the following transcendental equation: (3.25) ( ) 2 e (x3`) ( ) 2 e (x3`) + ` x3 ` =. It may be noted that this equation, and therefore x3 as well, does not depend on c. Finally, applying the condition (3.4) to the expression (3.24) we obtain a differential equation for the switching curve s 7! g3(s) that takes the following form: (3.26) g (s) = c e (g(s)s) c e (g(s)`) for all s 2 (s; `) where s < ` happens to be a singularity point at which dg=ds = +. The equation (3.26) is solved backwards under the initial condition (3.3) where x3 2 (`; `) is found by solving (3.25). The switching curve s 7! g3(s) is a unique solution of (3.26) obtained in this way. It can also be verified that (3.2) holds for some s3 2 (s; `). 4. It turns out that the solution of (3.26) satisfying (3.3) can hit the diagonal in IR 2 at a point s 3 2 (`; `) taken to be closest to x3, if c c3 for some large c3 to be determined below. When this happens, the construction of the solution becomes more complicated, and the solution of (3.26) for s 2 (s; s 3 ) is of no use. We thus first treat the simpler case when the solution of (3.26) stays below the diagonal ( the case of small c ), and this case is then followed by the more complicated case when the solution of (3.26) hits the diagonal ( the case of large c ). Step 3 ( The case of small c ). Having characterized the curve s 7! g 3 (s) for [s 3 ; `] we have obtained a candidate (3.8)+(3.24) for the value function (2.4) when (x; s) 2 C [ C 2. It remains to determine J(x; s) for (x; s) 2 C 3, and this is what we do in the final step. As clearly the control should be applied, consider the HJB equation: (3.27) J x + J xx c = in C 3 given and fixed. The general solution of (3.27) is given by (3.28) J(x; s) = a 2 (s) e x + c x + b 2 (s) 6

7 where a2(s) and b2(s) are some undetermined functions of s. Using the condition (3.5) we can eliminate b2(s) from (3.28) and this yields: (3.29) J (3) (x; s) = a 2(s) e x e ` + c (x `) + s. Applying the condition (3.4) to the expression (3.29) we find that a2(s) (3.3) a 2 (s) = e s e ` for ` < s < s 3. This equation can be solved explicitly and this gives: (3.3) a2(s) = e ` s + log e s e ` + d should solve: where d is an undetermined constant. To determine d we may use the fact that the maps (3.24) and (3.29) must coincide at (s 3 ; s 3 ), i.e. (3.32) J (2) (x; s) x=s=s3 = J (3) (x; s) the latter being known explicitly due to (3.2). With this d (3.33) a2(s) = e ` (s 3 s) log x=s=s3 we can then rewrite (3.3) as follows: c. e s e ` e s3 e ` Inserting this expression back into (3.29) we obtain a candidate for the value function (2.4) when (x; s) 2 C3, thus completing the construction of a solution to the system (3.3)-(3.6). Steps 3-5 ( The case of large c ). In this case the solution s 7! g 3 (s) of (3.26) satisfying (3.3) hits the diagonal in IR 2 at some s 2 (`; `). The set C 3 from (3.9) naturally splits into the following three subsets (modulo two curves): (3.34) C ; = f (x; s) 2 IR 2 j ` < x < g 3 (s) ; s 3 < s < ` g (3.35) C ;2 = f (x; s) 2 IR 2 j ` < x < s ; s 3 < s < s 3 g (3.36) C ;3 = f (x; s) 2 IR 2 j ` < x < h 3 (s) ; s 3 < s < s 3 g where s C 2 3, the map s 7! h 3(s), and s 3 in this context will soon be defined. Similarly, the set from (3.) naturally splits into the following two subsets (modulo one curve): (3.37) C 2; = f (x; s) 2 IR 2 j g 3 (s) < x s ; s 3 < s < ` g (3.38) C 2;2 = f (x; s) 2 IR 2 j h 3 (s) < x s ; s 3 < s < s 3 g. The set C 3 from (3.) remains the same, however, with a new value for s 3 to be introduced. We refer to Figure 2 below to obtain a better geometric understanding of (3.34)-(3.38). It is clear from the construction above that in C ; the value function (2.4) is given by (3.8), and that in C 2; the value function (2.4) is given by (3.24), where s 7! g 3 (s) is the solution of (3.26) satisfying (3.3). For the points s < s 3 (close to s 3 ) we can no longer make use 7

8 of the solution g 3 (s), and the expression (3.8), as clearly the condition (3.4) fails to hold. We need, moreover, to apply the control instead of, and thus Step 3 presented above must be modified. The HJB equation (3.27) is considered with in place of, and this again using (3.5) leads to the expression (3.29) with in place of : (3.39) J (;2) (x; s) = a 3(s) e x e ` + c (x `) + s as well as the equation (3.3) and its solution (3.3). To determine d (3.4) J (;2) (x; s) = J (i;) (x; s) x=s=s 3 x=s=s 3 in (3.3) we may use that for i either or 2, where we set J (;) = J () and J (2;) = J (2) with J () from (3.8) and J (2) from (3.24). The latter expression in (3.4) is known explicitly due to the fact that. With this d we can then rewrite an analogue of (3.33) as follows: g 3 (s 3 ) = s 3 (3.4) a3(s) = e ` e s (s 3 s) log e ` c e s 3 e ` e s Inserting this expression back into (3.39) we obtain a candidate for the value function (2.4) when (x; s) 2 C;2. Moreover, the equation J (;2) x = determines an optimal switching curve: (3.42) h 3 (s) = log c a 3(s) and the points s 3 2 (`; s 3 ) and s 3 2 (`; s 3 ) appearing in (3.35)+(3.36) and (3.38) are determined by the following identities: (3.43) h 3 (s 3 ) = s 3 (3.44) h 3 (s 3 ) = `. Note that (3.44) is reminiscent of (3.2) above, and so is h 3 (s) of g 3 (s) above. However, the two functions are different for s < s 3. It is clear that the value function (2.4) is also given by the formula (3.39) for (x; s) 2 C;3. To determine the value function (2.4) in C2;2, where clearly the control is to be applied, we can use the result of Step 2 above which leads to the following analogue of (3.24) above: (3.45) J (2;2) (x; s) = c () 2 e (h3(s)x) c () 2 e (h3(s)`) c h 3 (s) x + c h 3 (s) ` + s + c () 2 3. c () 2 for (x; s) 2 C2;2. Finally, in C3 we should also apply the control, and thus the result of Step 3 ( the case of small c ) above can be used. Due to (3.44) the expressions (3.29)+(3.33) carry over unchanged to the present case. It must be kept in mind, however, that s 3 satisfies (3.44) and not (3.2). This completes the construction of a solution to the system (3.3)-(3.6). 8

9 5. In this way we have obtained a candidate for the value function (2.4) when (x; s) 2 C in both cases of small and large c. The preceding considerations can now be summarized as follows. Theorem 3. In the setting of (2.)-(2.3) consider the optimal control problem (2.4) where the supremum is taken over all admissible controls as explained above.. In the case of small c in the sense that the solution of (3.26) stays below the diagonal in IR 2, the value function J = J(x; s) solves the system (3.3)-(3.6) and is explicitly given by (3.8) in C from (3.9), by (3.24) in C 2 from (3.), and by (3.29)+(3.33) in C 3 from (3.) with s 3 from (3.2), where s 7! g 3 (s) is the unique solution of the differential equation (3.26) for s 2 (s 3 ; `) satisfying g 3 (`) = x 3, and x 3 2 (`; `) is the unique solution of the transcendental equation (3.25). (The values of J at the two boundary curves in C not mentioned are determined by continuity.) 2. In the case of large c in the sense that the solution of (3.26) hits the diagonal in IR 2, the value function J = J(x; s) solves the system (3.3)-(3.6) and is explicitly given by (3.8) in C ; from (3.34), by (3.24) in C 2; from (3.37), by (3.39)+(3.4) in C ;2 [ C ;3 from (3.35)+(3.36), by (3.45) in C 2;2 from (3.38), and by (3.29)+(3.33) in C 3 from (3.) with s 3 from (3.44), where s 7! g 3 (s) is the unique solution of the differential equation (3.26) for s 2 (s ; `) satisfying 3 g 3 (`) = x 3 with x 3 2 (`; `) as above, and s 7! h 3 (s) is given by (3.42)+(3.4). (The values of J at the four boundary curves in C not mentioned are determined by continuity.) 3. The optimal control v 3 in the problem (2.4) is described as follows: Pull as hard as possible that is vt 3 = if X t < g 3 (S t ), and push as hard as possible that is vt 3 = if X t > g 3 (S t ), both as long as t where is given in (2.3). This description holds in the case of large c by modifying the solution g 3 (s) of (3.26) for s < s as follows: g 3 3 (s) := s + if s 2 (s ; ) 3 s 3 and g 3 (s) := h 3 (s) for s 2 (s 3 ; s ). In both cases of small and large c we formally set 3 g 3 (s) := ` for s 2 (`; s 3 ). (The control vt 3 for X t = g 3 (S t ) is arbitrary and has no effect on optimality of the result.) To verify the statements denote the candidate function by J(x; s) for (x; s) 2 C. Then, by construction, it is clear that (x; s) 7! J(x; s) is C 2 outside the set f (g 3 (s); s) j s 3 < s < ` g in C, and x 7! J(x; s) is C at g(s) when s is fixed. (Actually, the latter mapping is also C 2 as is easily seen from (3.4) and (3.9) above.) Thus Itô s formula can be applied to the process J(X t ; S t ) with any admissible v, and under P x;s in this way we get: (3.46) J(X t ; S t ) = J(x; s)+ = J(x; s) + Z t Z t J x (X r ; S r )dx r + Z t p 2 J x (X r ; S r ) db r + Z t J s (X r ; S r )ds r + 2 Z t ((IL v X )(J))(X r; S r ) dr J xx (X r ; S r )d X; X r where the integral with respect to ds r is zero, since the increment S r equals zero off the diagonal in IR 2, and at the diagonal we have (3.4). Since X equals either ` or `, we see by (3.5)+(3.6) that J(X ; S ) = S, and thus (3.46) implies the following identity: (3.47) S = J(x; s) + M + Z ((IL v X)(J))(X r ; S r ) dr 9

10 R t p where M t = 2 Jx (X r ; S r ) db r is a continuous (local) martingale. Since (x; s) 7! J x (x; s) is uniformly bounded in C, it follows by the optional sampling theorem that E x;s (M ) =. (Note that there is no restriction to consider only those controls v for which E x;s () < so that E x;s ( p ) < as well.) On the other hand, by (3.3) we know that (IL v X(J))(x; s) c for all (x; s) 2 C, with equality if v = v 3, and thus (3.47) yields: (3.48) J(x; s) = E x;s S Z ((IL v X)(J))(X r ; S r ) dr Ex;s S c with equality if v = v 3. This establishes that the candidate function equals the value function, and that the optimal control equals v 3, thus completing the proof. REFERENCES [] BLISS, G. A. (946). Lectures on the Calculus of Variations. Univ. Chicago Press. [2] BOLZA, O. (93). Über den Anormalen Fall beim Lagrangeschen und Mayerschen Problem mit gemischten Bedingungen und variablen Endpunkten. Math. Ann. 74 (43-446). [3] DUBINS, L. E. SHEPP, L. A. and SHIRYAEV, A. N. (993). Optimal stopping rules and maximal inequalities for Bessel processes. Theory Probab. Appl. 38 (226-26). [4] FLEMING, W. H. and RISHEL, R. W. (975). Deterministic and Stochastic Optimal Control. Springer-Verlag. [5] GRAVERSEN, S. E. PESKIR, G. and SHIRYAEV, A. N. (999). Stopping Brownian motion without anticipation as close as possible to its ultimate maximum. Research Report No. 45, 999, Dept. Theoret. Statist. Aarhus ( pp). Theory Probab. Appl. 45, 2 (25-36). [6] JACKA, S. D. (99). Optimal stopping and best constants for Doob-like inequalities I: The case p =. Ann. Probab. 9 (798-82). [7] NELSON, E. (967). Dynamical Theories of Brownian Motion. Princeton Univ. Press. [8] PESKIR, G. (997). Optimal stopping of the maximum process: The maximality principle. Research Report No. 377, 997, Dept. Theoret. Statist. Aarhus (3 pp). Ann. Probab. 26, 998 (64-64). [9] ROGERS, L. C. G. and WILLIAMS, D. (987). Diffusions, Markov Processes, and Martingales. Volume 2: Itô Calculus. John Wiley and Sons. [] SHEPP, L. A. and SHIRYAEV, A. N. (993). The Russian option: Reduced regret. Ann. Appl. Probab. 3 (63-64). Goran Peskir Department of Mathematical Sciences University of Aarhus, Denmark Ny Munkegade, DK-8 Aarhus home.imf.au.dk/goran goran@imf.au.dk

11 x * C C 2 s g * (s) s x s * C 3 Figure. The case of small c in the problem (2.4). The state space (triangle) of the process (X t; S t) from (2.)+(2.2) splits into three regions. In the region C the optimal control 3 vt equals, and in the region C 2 [ C 3 the optimal control 3 vt equals. The switching curve s 7! g3(s) is determined as the unique solution of the differential equation (3.26) satisfying g3(`) = x3 where x3 2 (`; `) is the unique solution of the transcedental equation (3.25). In this specific case we took ` = =, ` = = and c = 2. It turns out that x3 =, s3 = :574 8 and s = : (The point s is a singularity point at which dg3=ds = + and g3 takes a finite value.)

12 x * C 2, C, s g * (s) s x s h * (s) s * C,3 s * s * C,2 C 2,2 C 3 Figure 2. The case of large c in the problem (2.4). The state space (triangle) of the process (X t; S t) from (2.)+(2.2) splits into six regions. In the region C ; [ C ;2 [ C ;3 the optimal control v 3 t equals, and in the region C 2; [ C 2;2 [ C 3 the optimal control v 3 t equals. The switching curve s 7! g3(s) is determined as the unique solution of the differential equation (3.26) satisfying g3(`) = x3 where x3 2 (`; `) is the unique solution of the transcedental equation (3.25). The switching curve s 7! h3(s) differs from the solution s 7! g3(s) of (3.26) and is determined to meet the condition (3.4) as explained in the text. In this specific case we took ` = =, ` = = and c = 3:8. It turns out that x3 =, s 3 = :492 2, s 3 = :82273 and s3 = : For comparison we state that the smaller (of two) ~s3 satisfying g3(~s3 ) = ~s3 equals :8247 4, that ~s3 satisfying g3(~s3) = ` equals :9658 5, and that s = : (The point s is a singularity point at which dg3=ds = + and g3 takes a finite value.) 2

MAXIMUM PROCESS PROBLEMS IN OPTIMAL CONTROL THEORY

MAXIMUM PROCESS PROBLEMS IN OPTIMAL CONTROL THEORY MAXIMUM PROCESS PROBLEMS IN OPTIMAL CONTROL THEORY GORAN PESKIR Receied 16 January 24 and in reised form 23 June 24 Gien a standard Brownian motion B t ) t and the equation of motion dx t = t dt+ 2dBt,wesetS

More information

Solving the Poisson Disorder Problem

Solving the Poisson Disorder Problem Advances in Finance and Stochastics: Essays in Honour of Dieter Sondermann, Springer-Verlag, 22, (295-32) Research Report No. 49, 2, Dept. Theoret. Statist. Aarhus Solving the Poisson Disorder Problem

More information

The Azéma-Yor Embedding in Non-Singular Diffusions

The Azéma-Yor Embedding in Non-Singular Diffusions Stochastic Process. Appl. Vol. 96, No. 2, 2001, 305-312 Research Report No. 406, 1999, Dept. Theoret. Statist. Aarhus The Azéma-Yor Embedding in Non-Singular Diffusions J. L. Pedersen and G. Peskir Let

More information

On Doob s Maximal Inequality for Brownian Motion

On Doob s Maximal Inequality for Brownian Motion Stochastic Process. Al. Vol. 69, No., 997, (-5) Research Reort No. 337, 995, Det. Theoret. Statist. Aarhus On Doob s Maximal Inequality for Brownian Motion S. E. GRAVERSEN and G. PESKIR If B = (B t ) t

More information

Predicting the Time of the Ultimate Maximum for Brownian Motion with Drift

Predicting the Time of the Ultimate Maximum for Brownian Motion with Drift Proc. Math. Control Theory Finance Lisbon 27, Springer, 28, 95-112 Research Report No. 4, 27, Probab. Statist. Group Manchester 16 pp Predicting the Time of the Ultimate Maximum for Brownian Motion with

More information

Albert N. Shiryaev Steklov Mathematical Institute. On sharp maximal inequalities for stochastic processes

Albert N. Shiryaev Steklov Mathematical Institute. On sharp maximal inequalities for stochastic processes Albert N. Shiryaev Steklov Mathematical Institute On sharp maximal inequalities for stochastic processes joint work with Yaroslav Lyulko, Higher School of Economics email: albertsh@mi.ras.ru 1 TOPIC I:

More information

On Reflecting Brownian Motion with Drift

On Reflecting Brownian Motion with Drift Proc. Symp. Stoch. Syst. Osaka, 25), ISCIE Kyoto, 26, 1-5) On Reflecting Brownian Motion with Drift Goran Peskir This version: 12 June 26 First version: 1 September 25 Research Report No. 3, 25, Probability

More information

On the American Option Problem

On the American Option Problem Math. Finance, Vol. 5, No., 25, (69 8) Research Report No. 43, 22, Dept. Theoret. Statist. Aarhus On the American Option Problem GORAN PESKIR 3 We show how the change-of-variable formula with local time

More information

Solution of Stochastic Optimal Control Problems and Financial Applications

Solution of Stochastic Optimal Control Problems and Financial Applications Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty

More information

Optimal Prediction of the Ultimate Maximum of Brownian Motion

Optimal Prediction of the Ultimate Maximum of Brownian Motion Optimal Prediction of the Ultimate Maximum of Brownian Motion Jesper Lund Pedersen University of Copenhagen At time start to observe a Brownian path. Based upon the information, which is continuously updated

More information

A Barrier Version of the Russian Option

A Barrier Version of the Russian Option A Barrier Version of the Russian Option L. A. Shepp, A. N. Shiryaev, A. Sulem Rutgers University; shepp@stat.rutgers.edu Steklov Mathematical Institute; shiryaev@mi.ras.ru INRIA- Rocquencourt; agnes.sulem@inria.fr

More information

The Wiener Sequential Testing Problem with Finite Horizon

The Wiener Sequential Testing Problem with Finite Horizon Research Report No. 434, 3, Dept. Theoret. Statist. Aarhus (18 pp) The Wiener Sequential Testing Problem with Finite Horizon P. V. Gapeev and G. Peskir We present a solution of the Bayesian problem of

More information

OPTIMAL STOPPING OF A BROWNIAN BRIDGE

OPTIMAL STOPPING OF A BROWNIAN BRIDGE OPTIMAL STOPPING OF A BROWNIAN BRIDGE ERIK EKSTRÖM AND HENRIK WANNTORP Abstract. We study several optimal stopping problems in which the gains process is a Brownian bridge or a functional of a Brownian

More information

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012 1 Stochastic Calculus Notes March 9 th, 1 In 19, Bachelier proposed for the Paris stock exchange a model for the fluctuations affecting the price X(t) of an asset that was given by the Brownian motion.

More information

Introduction to Random Diffusions

Introduction to Random Diffusions Introduction to Random Diffusions The main reason to study random diffusions is that this class of processes combines two key features of modern probability theory. On the one hand they are semi-martingales

More information

Optimal Stopping Problems for Time-Homogeneous Diffusions: a Review

Optimal Stopping Problems for Time-Homogeneous Diffusions: a Review Optimal Stopping Problems for Time-Homogeneous Diffusions: a Review Jesper Lund Pedersen ETH, Zürich The first part of this paper summarises the essential facts on general optimal stopping theory for time-homogeneous

More information

ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER

ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER GERARDO HERNANDEZ-DEL-VALLE arxiv:1209.2411v1 [math.pr] 10 Sep 2012 Abstract. This work deals with first hitting time densities of Ito processes whose

More information

Pavel V. Gapeev, Neofytos Rodosthenous Perpetual American options in diffusion-type models with running maxima and drawdowns

Pavel V. Gapeev, Neofytos Rodosthenous Perpetual American options in diffusion-type models with running maxima and drawdowns Pavel V. Gapeev, Neofytos Rodosthenous Perpetual American options in diffusion-type models with running maxima and drawdowns Article (Accepted version) (Refereed) Original citation: Gapeev, Pavel V. and

More information

On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem

On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem Koichiro TAKAOKA Dept of Applied Physics, Tokyo Institute of Technology Abstract M Yor constructed a family

More information

On the principle of smooth fit in optimal stopping problems

On the principle of smooth fit in optimal stopping problems 1 On the principle of smooth fit in optimal stopping problems Amir Aliev Moscow State University, Faculty of Mechanics and Mathematics, Department of Probability Theory, 119992, Moscow, Russia. Keywords:

More information

STOPPING AT THE MAXIMUM OF GEOMETRIC BROWNIAN MOTION WHEN SIGNALS ARE RECEIVED

STOPPING AT THE MAXIMUM OF GEOMETRIC BROWNIAN MOTION WHEN SIGNALS ARE RECEIVED J. Appl. Prob. 42, 826 838 (25) Printed in Israel Applied Probability Trust 25 STOPPING AT THE MAXIMUM OF GEOMETRIC BROWNIAN MOTION WHEN SIGNALS ARE RECEIVED X. GUO, Cornell University J. LIU, Yale University

More information

On Wald-Type Optimal Stopping for Brownian Motion

On Wald-Type Optimal Stopping for Brownian Motion J Al Probab Vol 34, No 1, 1997, (66-73) Prerint Ser No 1, 1994, Math Inst Aarhus On Wald-Tye Otimal Stoing for Brownian Motion S RAVRSN and PSKIR The solution is resented to all otimal stoing roblems of

More information

n E(X t T n = lim X s Tn = X s

n E(X t T n = lim X s Tn = X s Stochastic Calculus Example sheet - Lent 15 Michael Tehranchi Problem 1. Let X be a local martingale. Prove that X is a uniformly integrable martingale if and only X is of class D. Solution 1. If If direction:

More information

Stochastic Hamiltonian systems and reduction

Stochastic Hamiltonian systems and reduction Stochastic Hamiltonian systems and reduction Joan Andreu Lázaro Universidad de Zaragoza Juan Pablo Ortega CNRS, Besançon Geometric Mechanics: Continuous and discrete, nite and in nite dimensional Ban,

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010 1 Stochastic Calculus Notes Abril 13 th, 1 As we have seen in previous lessons, the stochastic integral with respect to the Brownian motion shows a behavior different from the classical Riemann-Stieltjes

More information

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME SAUL D. JACKA AND ALEKSANDAR MIJATOVIĆ Abstract. We develop a general approach to the Policy Improvement Algorithm (PIA) for stochastic control problems

More information

I forgot to mention last time: in the Ito formula for two standard processes, putting

I forgot to mention last time: in the Ito formula for two standard processes, putting I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy

More information

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011 Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance

More information

ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS

ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS PORTUGALIAE MATHEMATICA Vol. 55 Fasc. 4 1998 ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS C. Sonoc Abstract: A sufficient condition for uniqueness of solutions of ordinary

More information

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS APPLICATIONES MATHEMATICAE 29,4 (22), pp. 387 398 Mariusz Michta (Zielona Góra) OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS Abstract. A martingale problem approach is used first to analyze

More information

The Russian option: Finite horizon. Peskir, Goran. MIMS EPrint: Manchester Institute for Mathematical Sciences School of Mathematics

The Russian option: Finite horizon. Peskir, Goran. MIMS EPrint: Manchester Institute for Mathematical Sciences School of Mathematics The Russian option: Finite horizon Peskir, Goran 25 MIMS EPrint: 27.37 Manchester Institute for Mathematical Sciences School of Mathematics The University of Manchester Reports available from: And by contacting:

More information

Applications of Optimal Stopping and Stochastic Control

Applications of Optimal Stopping and Stochastic Control Applications of and Stochastic Control YRM Warwick 15 April, 2011 Applications of and Some problems Some technology Some problems The secretary problem Bayesian sequential hypothesis testing the multi-armed

More information

Optimal Stopping and Maximal Inequalities for Poisson Processes

Optimal Stopping and Maximal Inequalities for Poisson Processes Optimal Stopping and Maximal Inequalities for Poisson Processes D.O. Kramkov 1 E. Mordecki 2 September 10, 2002 1 Steklov Mathematical Institute, Moscow, Russia 2 Universidad de la República, Montevideo,

More information

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES ERHAN BAYRAKTAR AND MIHAI SÎRBU Abstract. We adapt the Stochastic Perron s

More information

Lecture 12. F o s, (1.1) F t := s>t

Lecture 12. F o s, (1.1) F t := s>t Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let

More information

Optimal Stopping Games for Markov Processes

Optimal Stopping Games for Markov Processes SIAM J. Control Optim. Vol. 47, No. 2, 2008, (684-702) Research Report No. 15, 2006, Probab. Statist. Group Manchester (21 pp) Optimal Stopping Games for Markov Processes E. Ekström & G. Peskir Let X =

More information

Worst Case Portfolio Optimization and HJB-Systems

Worst Case Portfolio Optimization and HJB-Systems Worst Case Portfolio Optimization and HJB-Systems Ralf Korn and Mogens Steffensen Abstract We formulate a portfolio optimization problem as a game where the investor chooses a portfolio and his opponent,

More information

Stochastic Calculus (Lecture #3)

Stochastic Calculus (Lecture #3) Stochastic Calculus (Lecture #3) Siegfried Hörmann Université libre de Bruxelles (ULB) Spring 2014 Outline of the course 1. Stochastic processes in continuous time. 2. Brownian motion. 3. Itô integral:

More information

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Yipeng Yang * Under the supervision of Dr. Michael Taksar Department of Mathematics University of Missouri-Columbia Oct

More information

On Distributions Associated with the Generalized Lévy s Stochastic Area Formula

On Distributions Associated with the Generalized Lévy s Stochastic Area Formula On Distributions Associated with the Generalized Lévy s Stochastic Area Formula Raouf Ghomrasni Abstract A closed-form ression is obtained for the conditional probability distribution of t R s ds given

More information

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009 A new approach for investment performance measurement 3rd WCMF, Santa Barbara November 2009 Thaleia Zariphopoulou University of Oxford, Oxford-Man Institute and The University of Texas at Austin 1 Performance

More information

Stochastic Differential Equations.

Stochastic Differential Equations. Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)

More information

Lectures in Mathematics ETH Zürich Department of Mathematics Research Institute of Mathematics. Managing Editor: Michael Struwe

Lectures in Mathematics ETH Zürich Department of Mathematics Research Institute of Mathematics. Managing Editor: Michael Struwe Lectures in Mathematics ETH Zürich Department of Mathematics Research Institute of Mathematics Managing Editor: Michael Struwe Goran Peskir Albert Shiryaev Optimal Stopping and Free-Boundary Problems Birkhäuser

More information

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS (2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS Svetlana Janković and Miljana Jovanović Faculty of Science, Department of Mathematics, University

More information

STAT 331. Martingale Central Limit Theorem and Related Results

STAT 331. Martingale Central Limit Theorem and Related Results STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal

More information

Reflected Brownian Motion

Reflected Brownian Motion Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide

More information

Malliavin Calculus in Finance

Malliavin Calculus in Finance Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x

More information

Stochastic optimal control with rough paths

Stochastic optimal control with rough paths Stochastic optimal control with rough paths Paul Gassiat TU Berlin Stochastic processes and their statistics in Finance, Okinawa, October 28, 2013 Joint work with Joscha Diehl and Peter Friz Introduction

More information

On pathwise stochastic integration

On pathwise stochastic integration On pathwise stochastic integration Rafa l Marcin Lochowski Afican Institute for Mathematical Sciences, Warsaw School of Economics UWC seminar Rafa l Marcin Lochowski (AIMS, WSE) On pathwise stochastic

More information

Lecture 21 Representations of Martingales

Lecture 21 Representations of Martingales Lecture 21: Representations of Martingales 1 of 11 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 21 Representations of Martingales Right-continuous inverses Let

More information

Bessel Processes. Maximal Inequalities for. 0_<,<_:rmaX Z2t) S.E. GRAVERSEN and G. PEKIR * < G(oOV/-(T) 8000 Aarhus, Denmark

Bessel Processes. Maximal Inequalities for. 0_<,<_:rmaX Z2t) S.E. GRAVERSEN and G. PEKIR * < G(oOV/-(T) 8000 Aarhus, Denmark J. of Inequal. & Appl., 1998, Vol. 2, pp. 99-119 Reprints available directly from the publisher Photocopying permitted by license only (C) 1998 OPA (Overseas Publishers Association) Amsterdam B.V. Published

More information

Applied Mathematics Letters. Stationary distribution, ergodicity and extinction of a stochastic generalized logistic system

Applied Mathematics Letters. Stationary distribution, ergodicity and extinction of a stochastic generalized logistic system Applied Mathematics Letters 5 (1) 198 1985 Contents lists available at SciVerse ScienceDirect Applied Mathematics Letters journal homepage: www.elsevier.com/locate/aml Stationary distribution, ergodicity

More information

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner STOCHASTIC STABILITY H.J. Kushner Applied Mathematics, Brown University, Providence, RI, USA. Keywords: stability, stochastic stability, random perturbations, Markov systems, robustness, perturbed systems,

More information

MAXIMAL COUPLING OF EUCLIDEAN BROWNIAN MOTIONS

MAXIMAL COUPLING OF EUCLIDEAN BROWNIAN MOTIONS MAXIMAL COUPLING OF EUCLIDEAN BOWNIAN MOTIONS ELTON P. HSU AND KAL-THEODO STUM ABSTACT. We prove that the mirror coupling is the unique maximal Markovian coupling of two Euclidean Brownian motions starting

More information

Chapter 2 Event-Triggered Sampling

Chapter 2 Event-Triggered Sampling Chapter Event-Triggered Sampling In this chapter, some general ideas and basic results on event-triggered sampling are introduced. The process considered is described by a first-order stochastic differential

More information

Change detection problems in branching processes

Change detection problems in branching processes Change detection problems in branching processes Outline of Ph.D. thesis by Tamás T. Szabó Thesis advisor: Professor Gyula Pap Doctoral School of Mathematics and Computer Science Bolyai Institute, University

More information

A Change of Variable Formula with Local Time-Space for Bounded Variation Lévy Processes with Application to Solving the American Put Option Problem 1

A Change of Variable Formula with Local Time-Space for Bounded Variation Lévy Processes with Application to Solving the American Put Option Problem 1 Chapter 3 A Change of Variable Formula with Local Time-Space for Bounded Variation Lévy Processes with Application to Solving the American Put Option Problem 1 Abstract We establish a change of variable

More information

Pathwise Construction of Stochastic Integrals

Pathwise Construction of Stochastic Integrals Pathwise Construction of Stochastic Integrals Marcel Nutz First version: August 14, 211. This version: June 12, 212. Abstract We propose a method to construct the stochastic integral simultaneously under

More information

Estimating transition times for a model of language change in an age-structured population

Estimating transition times for a model of language change in an age-structured population Estimating transition times for a model of language change in an age-structured population W. Garrett Mitchener MitchenerG@cofc.edu http://mitchenerg.people.cofc.edu Abstract: Human languages are stable

More information

Lecture 22 Girsanov s Theorem

Lecture 22 Girsanov s Theorem Lecture 22: Girsanov s Theorem of 8 Course: Theory of Probability II Term: Spring 25 Instructor: Gordan Zitkovic Lecture 22 Girsanov s Theorem An example Consider a finite Gaussian random walk X n = n

More information

Reaction-Diffusion Equations In Narrow Tubes and Wave Front P

Reaction-Diffusion Equations In Narrow Tubes and Wave Front P Outlines Reaction-Diffusion Equations In Narrow Tubes and Wave Front Propagation University of Maryland, College Park USA Outline of Part I Outlines Real Life Examples Description of the Problem and Main

More information

First passage time for Brownian motion and piecewise linear boundaries

First passage time for Brownian motion and piecewise linear boundaries To appear in Methodology and Computing in Applied Probability, (2017) 19: 237-253. doi 10.1007/s11009-015-9475-2 First passage time for Brownian motion and piecewise linear boundaries Zhiyong Jin 1 and

More information

When is a Moving Average a Semimartingale?

When is a Moving Average a Semimartingale? 29 Barrett Lectures Ph.D.-student under supervision of Jan Pedersen, Thiele Centre, University of Aarhus, Denmark. 29 Barrett Lectures at The University of Tennessee: Stochastic Analysis and its Applications

More information

Noncooperative continuous-time Markov games

Noncooperative continuous-time Markov games Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and

More information

On the Multi-Dimensional Controller and Stopper Games

On the Multi-Dimensional Controller and Stopper Games On the Multi-Dimensional Controller and Stopper Games Joint work with Yu-Jui Huang University of Michigan, Ann Arbor June 7, 2012 Outline Introduction 1 Introduction 2 3 4 5 Consider a zero-sum controller-and-stopper

More information

Continuous Time Finance

Continuous Time Finance Continuous Time Finance Lisbon 2013 Tomas Björk Stockholm School of Economics Tomas Björk, 2013 Contents Stochastic Calculus (Ch 4-5). Black-Scholes (Ch 6-7. Completeness and hedging (Ch 8-9. The martingale

More information

On Stopping Times and Impulse Control with Constraint

On Stopping Times and Impulse Control with Constraint On Stopping Times and Impulse Control with Constraint Jose Luis Menaldi Based on joint papers with M. Robin (216, 217) Department of Mathematics Wayne State University Detroit, Michigan 4822, USA (e-mail:

More information

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p

More information

Simulation of diffusion. processes with discontinuous coefficients. Antoine Lejay Projet TOSCA, INRIA Nancy Grand-Est, Institut Élie Cartan

Simulation of diffusion. processes with discontinuous coefficients. Antoine Lejay Projet TOSCA, INRIA Nancy Grand-Est, Institut Élie Cartan Simulation of diffusion. processes with discontinuous coefficients Antoine Lejay Projet TOSCA, INRIA Nancy Grand-Est, Institut Élie Cartan From collaborations with Pierre Étoré and Miguel Martinez . Divergence

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Lecture 1: Brief Review on Stochastic Processes

Lecture 1: Brief Review on Stochastic Processes Lecture 1: Brief Review on Stochastic Processes A stochastic process is a collection of random variables {X t (s) : t T, s S}, where T is some index set and S is the common sample space of the random variables.

More information

On Mean-Square and Asymptotic Stability for Numerical Approximations of Stochastic Ordinary Differential Equations

On Mean-Square and Asymptotic Stability for Numerical Approximations of Stochastic Ordinary Differential Equations On Mean-Square and Asymptotic Stability for Numerical Approximations of Stochastic Ordinary Differential Equations Rózsa Horváth Bokor and Taketomo Mitsui Abstract This note tries to connect the stochastic

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Some Tools From Stochastic Analysis

Some Tools From Stochastic Analysis W H I T E Some Tools From Stochastic Analysis J. Potthoff Lehrstuhl für Mathematik V Universität Mannheim email: potthoff@math.uni-mannheim.de url: http://ls5.math.uni-mannheim.de To close the file, click

More information

On the submartingale / supermartingale property of diffusions in natural scale

On the submartingale / supermartingale property of diffusions in natural scale On the submartingale / supermartingale property of diffusions in natural scale Alexander Gushchin Mikhail Urusov Mihail Zervos November 13, 214 Abstract Kotani 5 has characterised the martingale property

More information

1 Brownian Local Time

1 Brownian Local Time 1 Brownian Local Time We first begin by defining the space and variables for Brownian local time. Let W t be a standard 1-D Wiener process. We know that for the set, {t : W t = } P (µ{t : W t = } = ) =

More information

Risk-Minimality and Orthogonality of Martingales

Risk-Minimality and Orthogonality of Martingales Risk-Minimality and Orthogonality of Martingales Martin Schweizer Universität Bonn Institut für Angewandte Mathematik Wegelerstraße 6 D 53 Bonn 1 (Stochastics and Stochastics Reports 3 (199, 123 131 2

More information

EULER MARUYAMA APPROXIMATION FOR SDES WITH JUMPS AND NON-LIPSCHITZ COEFFICIENTS

EULER MARUYAMA APPROXIMATION FOR SDES WITH JUMPS AND NON-LIPSCHITZ COEFFICIENTS Qiao, H. Osaka J. Math. 51 (14), 47 66 EULER MARUYAMA APPROXIMATION FOR SDES WITH JUMPS AND NON-LIPSCHITZ COEFFICIENTS HUIJIE QIAO (Received May 6, 11, revised May 1, 1) Abstract In this paper we show

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

Stochastic Integration and Stochastic Differential Equations: a gentle introduction

Stochastic Integration and Stochastic Differential Equations: a gentle introduction Stochastic Integration and Stochastic Differential Equations: a gentle introduction Oleg Makhnin New Mexico Tech Dept. of Mathematics October 26, 27 Intro: why Stochastic? Brownian Motion/ Wiener process

More information

Chapter 10. Definition of the Derivative Velocity and Tangents

Chapter 10. Definition of the Derivative Velocity and Tangents Chapter 10 Definition of the Derivative 10.1 Velocity and Tangents 10.1 Notation. If E 1 (x, y) and E 2 (x, y) denote equations or inequalities in x and y, we will use the notation {E 1 (x, y)} = {(x,

More information

Stationary distribution and pathwise estimation of n-species mutualism system with stochastic perturbation

Stationary distribution and pathwise estimation of n-species mutualism system with stochastic perturbation Available online at www.tjnsa.com J. Nonlinear Sci. Appl. 9 6), 936 93 Research Article Stationary distribution and pathwise estimation of n-species mutualism system with stochastic perturbation Weiwei

More information

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations 2005 American Control Conference June 8-10, 2005. Portland, OR, USA WeB10.3 A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations Arik Melikyan, Andrei Akhmetzhanov and Naira

More information

Random G -Expectations

Random G -Expectations Random G -Expectations Marcel Nutz ETH Zurich New advances in Backward SDEs for nancial engineering applications Tamerza, Tunisia, 28.10.2010 Marcel Nutz (ETH) Random G-Expectations 1 / 17 Outline 1 Random

More information

AN EXTREME-VALUE ANALYSIS OF THE LIL FOR BROWNIAN MOTION 1. INTRODUCTION

AN EXTREME-VALUE ANALYSIS OF THE LIL FOR BROWNIAN MOTION 1. INTRODUCTION AN EXTREME-VALUE ANALYSIS OF THE LIL FOR BROWNIAN MOTION DAVAR KHOSHNEVISAN, DAVID A. LEVIN, AND ZHAN SHI ABSTRACT. We present an extreme-value analysis of the classical law of the iterated logarithm LIL

More information

Likelihood Functions for Stochastic Signals in White Noise* TYRONE E. DUNCAN

Likelihood Functions for Stochastic Signals in White Noise* TYRONE E. DUNCAN INFORMATION AND CONTROL 16, 303-310 (1970) Likelihood Functions for Stochastic Signals in White Noise* TYRONE E. DUNCAN Computer, Information and Control Engineering, The University of Nlichigan, Ann Arbor,

More information

Stochastic Mechanics of Particles and Fields

Stochastic Mechanics of Particles and Fields Stochastic Mechanics of Particles and Fields Edward Nelson Department of Mathematics, Princeton University These slides are posted at http://math.princeton.edu/ nelson/papers/xsmpf.pdf A preliminary draft

More information

WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES. 1. Introduction

WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES. 1. Introduction WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES BRIAN D. EWALD 1 Abstract. We consider the weak analogues of certain strong stochastic numerical schemes considered

More information

Scaling Limits of Waves in Convex Scalar Conservation Laws under Random Initial Perturbations

Scaling Limits of Waves in Convex Scalar Conservation Laws under Random Initial Perturbations Scaling Limits of Waves in Convex Scalar Conservation Laws under Random Initial Perturbations Jan Wehr and Jack Xin Abstract We study waves in convex scalar conservation laws under noisy initial perturbations.

More information

Prof. Erhan Bayraktar (University of Michigan)

Prof. Erhan Bayraktar (University of Michigan) September 17, 2012 KAP 414 2:15 PM- 3:15 PM Prof. (University of Michigan) Abstract: We consider a zero-sum stochastic differential controller-and-stopper game in which the state process is a controlled

More information

Introduction to Algorithmic Trading Strategies Lecture 3

Introduction to Algorithmic Trading Strategies Lecture 3 Introduction to Algorithmic Trading Strategies Lecture 3 Trend Following Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com References Introduction to Stochastic Calculus with Applications.

More information

Convergence at first and second order of some approximations of stochastic integrals

Convergence at first and second order of some approximations of stochastic integrals Convergence at first and second order of some approximations of stochastic integrals Bérard Bergery Blandine, Vallois Pierre IECN, Nancy-Université, CNRS, INRIA, Boulevard des Aiguillettes B.P. 239 F-5456

More information

Asymptotic Perron Method for Stochastic Games and Control

Asymptotic Perron Method for Stochastic Games and Control Asymptotic Perron Method for Stochastic Games and Control Mihai Sîrbu, The University of Texas at Austin Methods of Mathematical Finance a conference in honor of Steve Shreve s 65th birthday Carnegie Mellon

More information

for all f satisfying E[ f(x) ] <.

for all f satisfying E[ f(x) ] <. . Let (Ω, F, P ) be a probability space and D be a sub-σ-algebra of F. An (H, H)-valued random variable X is independent of D if and only if P ({X Γ} D) = P {X Γ}P (D) for all Γ H and D D. Prove that if

More information

CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67

CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67 CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Chapter2 p. 1/67 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Main Purpose: Introduce the maximum principle as a necessary condition to be satisfied by any optimal

More information

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications The multidimensional Ito Integral and the multidimensional Ito Formula Eric Mu ller June 1, 215 Seminar on Stochastic Geometry and its applications page 2 Seminar on Stochastic Geometry and its applications

More information

March 16, Abstract. We study the problem of portfolio optimization under the \drawdown constraint" that the

March 16, Abstract. We study the problem of portfolio optimization under the \drawdown constraint that the ON PORTFOLIO OPTIMIZATION UNDER \DRAWDOWN" CONSTRAINTS JAKSA CVITANIC IOANNIS KARATZAS y March 6, 994 Abstract We study the problem of portfolio optimization under the \drawdown constraint" that the wealth

More information