Solvability of G-SDE with Integral-Lipschitz Coefficients

Similar documents
arxiv:math/ v1 [math.pr] 13 Feb 2007

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem

Stochastic Dynamical System for SDE driven by Nonlinear Brownian Motion

Law of Large Number and Central Limit Theorem under Uncertainty, the Related/ New 53. and Applications to Risk Measures

Mean-field SDE driven by a fractional BM. A related stochastic control problem

Robust Mean-Variance Hedging via G-Expectation

On continuous time contract theory

Random G -Expectations

Quasi-sure Stochastic Analysis through Aggregation

Backward Stochastic Differential Equations with Infinite Time Horizon

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

1. Stochastic Processes and filtrations

Optimal stopping for non-linear expectations Part I

Contents. 1. Introduction

A Concise Course on Stochastic Partial Differential Equations

Uniformly Uniformly-ergodic Markov chains and BSDEs

Stochastic Calculus (Lecture #3)

EULER MARUYAMA APPROXIMATION FOR SDES WITH JUMPS AND NON-LIPSCHITZ COEFFICIENTS

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Weak solutions of mean-field stochastic differential equations

Solving Stochastic Partial Differential Equations as Stochastic Differential Equations in Infinite Dimensions - a Review

BSDEs and PDEs with discontinuous coecients Applications to homogenization K. Bahlali, A. Elouain, E. Pardoux. Jena, March 2009

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Minimal Supersolutions of Backward Stochastic Differential Equations and Robust Hedging

Chapter 2 Random Ordinary Differential Equations

Properties of an infinite dimensional EDS system : the Muller s ratchet

Approximating solutions of nonlinear second order ordinary differential equations via Dhage iteration principle

Reflected Brownian Motion

SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE

MA8109 Stochastic Processes in Systems Theory Autumn 2013

Exercises. T 2T. e ita φ(t)dt.

On a general definition of transition waves and their properties

Wiener Measure and Brownian Motion

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Convergence at first and second order of some approximations of stochastic integrals

LAN property for sde s with additive fractional noise and continuous time observation

I forgot to mention last time: in the Ito formula for two standard processes, putting

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009

Theory of Ordinary Differential Equations

Martingale Representation Theorem for the G-expectation

Stochastic Differential Equations.

Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.

µ X (A) = P ( X 1 (A) )

A Narrow-Stencil Finite Difference Method for Hamilton-Jacobi-Bellman Equations

Markov Chain BSDEs and risk averse networks

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Stochastic integration. P.J.C. Spreij

Stabilization of persistently excited linear systems

Probability and Measure

Infinite-dimensional calculus under weak spatial regularity of the processes

Approximation of BSDEs using least-squares regression and Malliavin weights

On the Multi-Dimensional Controller and Stopper Games

Optimal Stopping under Adverse Nonlinear Expectation and Related Games

Half of Final Exam Name: Practice Problems October 28, 2014

A class of globally solvable systems of BSDE and applications

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

Stochastic optimal control with rough paths

Main topics and some repetition exercises for the course MMG511/MVE161 ODE and mathematical modeling in year Main topics in the course:

Lecture 4: Ito s Stochastic Calculus and SDE. Seung Yeal Ha Dept of Mathematical Sciences Seoul National University

In essence, Dynamical Systems is a science which studies differential equations. A differential equation here is the equation

Rough Burgers-like equations with multiplicative noise

On linear and non-linear equations. (Sect. 1.6).

1 Stochastic Dynamic Programming

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

I. The space C(K) Let K be a compact metric space, with metric d K. Let B(K) be the space of real valued bounded functions on K with the sup-norm

Numerical schemes of resolution of stochastic optimal control HJB equation

Discretization of SDEs: Euler Methods and Beyond

The Wiener Itô Chaos Expansion

The De Giorgi-Nash-Moser Estimates

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

Stability of Stochastic Differential Equations

Information and Credit Risk

n E(X t T n = lim X s Tn = X s

Lecture 22 Girsanov s Theorem

THE FUNDAMENTAL THEOREM OF SPACE CURVES

GENERALIZED COVARIATION FOR BANACH SPACE VALUED PROCESSES, ITÔ FORMULA AND APPLICATIONS

Solutions to the Exercises in Stochastic Analysis

Calculus of Variations. Final Examination

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE

Nonlinear Lévy Processes and their Characteristics

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Minimal Surface equations non-solvability strongly convex functional further regularity Consider minimal surface equation.

Some Properties of NSFDEs

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

13 The martingale problem

Existence and Comparisons for BSDEs in general spaces

NORMS ON SPACE OF MATRICES

Lecture 19 L 2 -Stochastic integration

Some SDEs with distributional drift Part I : General calculus. Flandoli, Franco; Russo, Francesco; Wolf, Jochen

Convergence & Continuity

Supermodular ordering of Poisson arrays

Lecture 4 Lebesgue spaces and inequalities

Functional Differential Equations with Causal Operators

Stochastic Differential Equations

Invariant measures for iterated function systems

Transcription:

Solvability of G-SDE with Integral-Lipschitz Coefficients Yiqing LIN joint work with Xuepeng Bai IRMAR, Université de Rennes 1, FRANCE http://perso.univ-rennes1.fr/yiqing.lin/ ITN Marie Curie Workshop Stochastic Control and Finance 18th, March 21, Roscoff 1 / 42

Content Statement of the Problem G-Brownian motion and G-capacity Solvability of G-SDE with Integral-Lipschitz Coefficients Solvability of G-BSDE with Integral-Lipschitz Coefficients 2 / 42

1. Statement of the Problem Motivated by uncertainty problems, risk measures and the superhedging in finance, Peng introduced G-Brownian motion. The expectation E[ ] associated with G-Brownian motion is a sublinear expectation which is called G-expectation. In the papers, Peng [27] & Gao [29], the stochastic calculus with respect to the G-Brownian motion has been established. In this work, we study the solvability of the following stochastic differential equation driven by G-Brownian motion: X(t) = x + b(s, X(s))ds + h(s,x(s))d B,B s + σ(s,x(s))db s, (1) where t [,T], the initial condition x R n is given and ( B,B t ) t is the quadratic variation process of G-Brownian motion (B t ) t. The solvability of (1) has been studied in Peng [27] & Gao [29], where the coefficients b, h and σ are subject to a Lipschitz condition, and the existence and uniqueness of the solution to (1) has been obtained by the contraction mapping principle and Picard iteration respectively. 3 / 42

In this work, we establish the existence and uniqueness of the solution to (1) under the following so-called integral-lipschitz condition: b(t,x 1 ) b(t,x 2 ) 2 + h(t,x 1 ) h(t,x 2 ) 2 + σ(t,x 1 ) σ(t,x 2 ) 2 ρ( x 1 x 2 2 ), (2) where ρ : (,+ ) (,+ ) is a continuous, increasing, concave function satisfying 1 dr ρ(+) =, ρ(r) = +. A typical example of (2) is: b(t,x 1 ) b(t,x 2 ) + h(t,x 1 ) h(t,x 2 ) + σ(t,x 1 ) σ(t,x 2 ) x 1 x 2 (ln 1 x 1 x 2 ) 1 2. Furthermore, we also establish the existence and uniqueness of the solution to equation (1) under a weaker condition on b and h, i.e., b(t,x 1 ) b(t,x 2 ) ρ( x 1 x 2 ); h(t,x 1 ) h(t,x 2 ) ρ( x 1 x 2 ). (3) A typical example of (3) is: 1 b(t,x 1 ) b(t,x 2 ) x 1 x 2 ln x 1 x 2 ; h(t,x 1 1) h(t,x 2 ) x 1 x 2 ln x 1 x 2. 4 / 42

Reference: Watanabe-Yamada [1971] & Yamada [1981] studied the solvability of (1) under the condition (2) for classical finite dimensional case and the uniqueness of solutions to (1) under the condition (3) for classical case. Hu-Lerner [22] studied the solvability of (1) under the condition (2) for classical infinite dimensional case and the existence of solutions to (1) under the condition (3) for classical case. In this work, we prove the existence and uniqueness of the solution to (1) still hold under either condition (2) or condition (3) 5 / 42

At the end of the work, we also consider the following type of G-BSDE: T T Y t = E[ξ + f (s,y s )ds + g(s,y s )d B,B s F t ], (4) t t where t [,T] and ξ L 1 G (F T;R n ). While the solvability for (4) under a Lipschitz condition on the coefficients f and g has been given by Peng [27], we prove that under the integral-lipschitz condition: g(s,y 1 ) g(s,y 2 ) + f (s,y 1 ) f (s,y 2 ) ρ( y 1 y 2 ), the existence and uniqueness of the solution to (4) are obtained as well. 6 / 42

2. G-Brownian Motion and G-Capacity Adapting approach in Peng [27], we have the following basic notions: Ω: a given nonempty fundamental space. H : a space of random variables which is a linear space of real functions defined on Ω such that : 1 H ; H is stable with respect to local Lipschitz functions, i.e., for all n 1, and for all X 1,...,X n H, ϕ C l,lip (R n ), ϕ(x 1,...,X n ) H. E: a sublinear expectation which is a functional E : H R with the following properties : for all X,Y H, we have Monotonicity: if X Y, then E[X] E[Y]; Preservation of constants: E[c] = c, for all c R; Sub-additivity: E[X] E[Y] E[X Y]; Positive homogeneity: E[λX] = λe[x], for all λ. 7 / 42

Definition (Independent) For arbitrary n,m 1, a random vector Y H n is said to be independent of X H m under E[ ] if for each test function ϕ C l,lip (R n+m ) we have E[ϕ(X,Y)] = E[E[ϕ(x,Y)] x=x ]. Definition (Identically Distributed) Given two sublinear expectation spaces (Ω,H,E) and ( Ω, H, Ẽ), two random vectors X H n and Y H n are said to be identically distributed if for each test function ϕ C l,lip (R n ) F X [ϕ] = F Y [ϕ], where F X [ϕ] := E[ϕ(X)] and F Y [ϕ] := E[ϕ(Y)] respectively. 8 / 42

Definition (G-Normal Distributed) A d-dimensional random vector X in a sublinear expectation space (Ω,H,E) is called G-normal distributed if for each ϕ C l,lip (R d ), u(t,x) := E[ϕ(x + tx)], t,x R d is the viscosity solution of the following PDE defined on [, ) R d : u t G(D2 u) =, u t= = ϕ, where G = G X (A) : S d R is defined by and D 2 u = ( 2 x i x j u) d i,j=1. G X (A) := 1 2 E[ AX,X ],A Sd, 9 / 42

Remark In particular, E[ϕ(X)] = u(1,), and by Peng [27] it is easy to check that, for a G-normal distributed random vector X, there exists a bounded, convex and closed subset Γ of R d, which is the space of all d d matrices, such that for each A S d, G(A) = G X (A) can be represented as G(A) = 1 2 sup tr[γγ T A]. γ Γ Consequencely, we can denote the G-normal distribution by N(, Σ), where Σ := {γγ T,γ Γ}. 1 / 42

In order to establish the stochastic calculus in the framework of G-expectation, we firstly give a more explicit definition of the triple (Ω,H,E): Sample space: Ω := {ω : [, ) R d (ω t ) t is continuous and ω = }. Distance on Ω: Ramdom variable space: and ρ(ω 1,ω 2 ) := 2 i [( max i=1 t [,i] ω1 t ωt 2 ) 1], L ip (F T) := {ϕ(b t1,...,b tn ) : n 1,t 1,...,t n [,T],ϕ C l,lip (R d n )}, Lip (F ) := Lip (F n), n=1 where B t (ω) = ω t,t [,+ ) is the canonical process. 11 / 42

Definition (G-Expectation) Sublinear expectation E : Lip (F ) R on (Ω,L ip (F ) is a G-expectation if and only if (B t ) t is a G-Brownian motion under E, that is, B (ω) = ; B t+s B t is N(,sΣ)-distributed and independent of (B t1,...,b tn ), for each n N, t 1... t n t, and t,s. Remark We denote by L p G (F T) (resp. L p G (F )) the topological completion of Lip (F T) (resp. Lip (F )) under the Banach norm E[ p ] 1 p, 1 p. 12 / 42

Definition The related conditional expectation of X L ip (F T) under L ip (F t j ): E[X F tj ] = E[ϕ(B t1,b t2 B t1,...,b tn B tn 1 ) F tj ] = E[ψ(B t1,b t2 B t1,...,b tj B tj 1 )], where ψ(x 1,...,x j ) = Ē[ϕ(x 1,...,x j, t j+1 t j ξ j+1,..., t n t n 1 ξ n )], and t 1... t n T, (ξ 1,...,ξ n ) are a sequence of G-normal distributed random vectors. Remark Since, for X, Y L ip (F t j ), E[ E[X F tj ] E[Y F tj ] ] E[ X Y ], the mapping E[ F tj ] : L ip (F T) L ip (F t j ) can be continuously extended to L 1 G (F T) L 1 G (F t j ). 13 / 42

From the above definition, we know each G-expectation is determined by the parameter G, which coincides with Γ, so we have the following representation theorem according to Denis-Hu-Peng [28]. Let A, Γ be the collection of all Γ valued natural filtration {F t, t } adapted processes on the interval [, ). For each fixed θ A, Γ, set P θ be the law of the process ( θ sdb s ) t under the Wiener measure P, and P := {P θ : θ A, Γ }. Definition (Quasi-Surely) For each A B(Ω), we define C(A) := sup P(A), P P then a set A is called polar if C(A) =. Moreover, a property is said to hold quasi-surely (q.s.) if it holds outside a polar set. 14 / 42

Theorem (Denis-Hu-Peng [28]) For each X L (Ω), we define then we have Ē[X] := sup E P (X), P P L 1 G(F ) = {X L (Ω) : X is q.s.continuous and Furthermore for all X LG 1 (F ), E[X] = Ē[X]. lim Ē[ X I { X >n} ] = }. n + From Denis-Hu-Peng [28] and Gao [29], we also have the following monotone convergence theorem: Theorem X n L 1 G (F ), X n X, q.s. E[X n ] Ē[X]. X n B(Ω), X n X, q.s., E P (X 1 ) > for all P P Ē[X n ] Ē[X]. (5) 15 / 42

In Gao [29], a generalized Itô integral and a generalized Itô formula with respect to G-Brownian motion are established: A partion of [,T]: π N T = {t,t 1,...,t N }, = t < t 1 <... < t N = T. A collection of simple processes M p, G (,T;R), p 1: M p, N 1 G (,T;R) := {η t = ξ j I [tj,t j+1 )(t);ξ j L p G (F t j ),j =,1,,...,N 1}. j= Norm on M p, G (,T;R): ( T ) 1 E[ η t p ]dt ( p N 1 = j= ) 1 E[ ξ j p p ](t j+1 t j ). We denote by M p G (,T;R) the topological completion of Mp, G (,T;R) under the Banach norm ( T E[ η t p ]dt) 1 p. 16 / 42

Let a = (a 1,...,a d ) T be a given vector in R d, we set (B a t ) t = (a,b t ) t, where (a,b t ) denotes the scalar product of a and B t. Definition For each η M 2, G ([,T]) with the form we define N 1 η t (ω) = ξ j I [tj,t j+1 )(t), j= T I (η) = η s db a N 1 s := ξ j (B a t j+1 B a t j ), j= the mapping can be continuously extended to I : MG 2 ([,T]) L2 G (F T). Then, for each η MG 2 ([,T]), the stochastic integral is defined by T η s db a s := I (η). 17 / 42

Denote by ( B a t ) t the quadratic variation process of process (B a t ) t, we know from Peng [27] that ( B a t ) t is an increasing process with B a =, and for each fixed s, B a t+s B a s = (B s ) a t, where B s t = B t+s B s,t,(b s ) a t = (a,b s t ). The mutual variation process of B a and Bā is defined by B a,bā t := 1 4 ( Ba + Bā t B a Bā t ). Definition Define the mapping M 1, G ([,T]) L1 G (F T) as follows: T Q(η) = η(s)d B a N 1 s := ξ k ( B a tk+1 B a tk ). k= Then Q can be uniquely extended to MG 1 ([,T]). We still use Q(η) to denote the mapping T η(s)d Ba s, η MG 1 ([,T]). 18 / 42

We introduce two important inequalities for G-stochastic integrals which we will need in the sequel. Theorem (Gao [29], BDG Inequality) Let p 2 and η = {η s,s [,T]} M p G ([,T]). For a Rd, set X t = η sdb a s. Then there exists a continuous modification X of X, i.e., on some Ω Ω, with C( Ω c ) =, X (ω) C [,T] and C( X t X t = ) = for all t [,T], such that ( p/2 Ē[ sup X u X s p ] C p σ p/2 E[ η aa s u t T u du) 2 ], s where < C p < is a positive constant independent of a, η and Γ. 19 / 42

Theorem (Gao [29]) Let p 1 and a,ā R d. Let η M p G ([,T]). Then there exists a continuous modification X t a,ā of Xt a,ā := η sd B a,bā s such that for any s t T, E[ sup u [s,t] X u a,ā X s a,ā p ] ( 1 4 σ (a+ā)(a+ā) T + 1 ) p 4 σ (a ā)(a ā) T (t s) p 1 E[ η u p du]. s Remark By the above two Theorems, we can assume that the stochastic integrals η sdb a s, η sd B a,bā s and η sds are continuous in t for all ω Ω. 2 / 42

Theorem (G-Itô s formula) Let α ν, η νij and β νj MG 2 ([,T]), ν = 1,...,n, i,j = 1,...,d be bounded processes and consider Xt ν = X ν + α ν d s ds + ηs νij d B i,b j d s + βs νj db j s, i,j=1 j=1 where X ν R, ν = 1,...,n. Let Φ C2 (R n ) be a real function with bounded derivatives such that { x 2 µ x ν Φ} n µ,ν=1 are uniformly Lipschitz. Then for each s,t [,T], in LG 2 (F t) Φ(X t ) Φ(X s ) = s + 1 2 s + s x ν Φ(X u )αu ν du + s 2 x µ x ν Φ(X u)β µi u β νj u d B i,b j u x ν Φ(X u )β νj u db j u, x ν Φ(X u )ηu νij d B i,b j u where the repeated indices ν, µ, i and j imply the summation. 21 / 42

3. Solvability of G-SDE with Integral-Lipschitz Coefficients Consider the following stochastic differential equation (1) driven by a d-dimensional G-Brownian motion, and we rewrite it in an equivalent form: d X t = x + b(s,x s )ds + h ij (s,x s )d B i,b j d s + σ j (s,x s )db j s, (6) i,j=1 j=1 b(,x), h ij (,x), σ j (,x) M 2 G ([,T];Rn ); b(t,x) 2 + d i,j=1 h ij(t,x) 2 + d j=1 σ j(t,x) 2 β 2 1 (t) + β 2 2 (t) x 2 ; b(t,x 1 ) b(t,x 2 ) 2 + d i,j=1 h ij(t,x 1 ) h ij (t,x 2 ) 2 + d j=1 σ j(t,x 1 ) σ j (t,x 2 ) 2 β 2 (t)ρ( x 1 x 2 2 ), (H1) (H2) where β 1 MG 2 ([,T]), β : [,T] R+, β 2 : [,T] R + are square integrable, and ρ : (,+ ) (,+ ) is continuous, increasing, concave function satisfying 1 dr ρ(+) =, = +. (7) ρ(r) 22 / 42

Theorem 1 We suppose (H1) and (H2), then there exists a unique continuous process X( ;x) L 2 G ([,T];Rn ) (for all t, X(t;x) L 2 G (F t;r n )) which satisfies (6). Lemma 1 (Hu-Lerner [22]) Let ρ : (,+ ) (,+ ) be a continuous, increasing function satisfying (7) and let u be a measurable, non-negative function defined on (,+ ) satisfying u(t) a + β(s)ρ(u(s))ds, t (,+ ), where a [,+ ), and β : [,T] R + is Lebesgue integrable. We have: If a =, then u(t) =, for t [,+ ); If a >, we define v(t) = t (ds/ρ(s)), t [,+ ), where t (,+ ), then u(t) v 1 (v(a) + β(s)ds). (8) 23 / 42

Lemma 2 (Jensen s Inequality) Let ρ : R R be a continuous increasing, concave function defined on R, then for each X LG 1 (F ), by a classical argument, the following inequality holds, ρ(ē[x]) Ē[ρ(X)]. Proof to Theorem 1: Uniqueness X(t;x 1 ) X(t;x 2 ) 2 4 x 1 x 2 2 + 4 b(s,x(s;x 1 )) b(s,x(s;x 2 ))ds 2 d + 4 h ij (s,x(s;x 1 )) h ij (s,x(s;x 2 ))d B i,b j s 2 i,j=1 d + 4 σ j (s,x(s;x 1 )) σ j (s,x(s;x 2 ))db j s 2. j=1 24 / 42

From the theorems in Gao [29], we notice that r Ē[ sup (b(s,x(s;x 1 )) b(s,x(s;x 2 )))ds 2 ] r t K 1 t β 2 (s)ē[ρ( X(s,x 1 ) X(s,x 2 ) 2 )]ds; r Ē[ sup (h ij (s,x(s;x 1 )) h ij (s,x(s;x 2 )))d B i,b j s 2 ] r t K 2 t β 2 (s)ē[ρ( X(s,x 1 ) X(s,x 2 ) 2 )]ds; r Ē[ sup (σ j (s,x(s;x 1 )) σ j (s,x(s;x 2 )))db j s 2 ] r t K 3 β 2 (s)ē[ρ( X(s;x 1 ) X(s;x 2 ) 2 )]ds. 25 / 42

Let u(t) = Ē[ sup X(r;x 1 ) X(r;x 2 ) 2 ], r t due to the sub-additivity property of Ē[ ] and Lemma 2, we have u(t) C 1 x 1 x 2 2 + C 2 β 2 (s)ē[ρ( X(s;x 1 ) X(s;x 2 ) 2 )]ds C 1 x 1 x 2 2 + C 2 β 2 (s)ρ(ē[ X(s;x 1 ) X(s;x 2 ) 2 ])ds C 1 x 1 x 2 2 + C 2 β 2 (s)ρ(u(s))ds. In particular, if x 1 = x 2, we obtain the uniqueness of the solution to (6). 26 / 42

Proof to Theorem 1: Existence Picard Iteration: X m+1 (t) = x + + d j=1 b(s,x m (s))ds + Priori estimate for {Ē[ X m (t) 2 ],m } : d i,j=1 h ij (s,x m (s))d B i,b j s σ j (s,x m (s))db j s, t [,T]. (9) Ē[ X m+1 (t) 2 ] C 1 x 2 + C 2 Ē[β1 2 (s) + β 2 2 (s) Xm (s) 2 ]ds. Dominated by the solution p of: p(t) = C 1 x 2 + C 2 Ē[β1 2 (s)]ds + C 2 β2 2 (s)p(s)ds, which is bounded from the explicit form. 27 / 42

Set u k+1,m (t) = sup Ē[ X k+1+m (r) X k+1 (r) 2 ]. r t From the definition of the sequence {X m ( ),m }, we have Set then, Finally, we have: u k+1,m (t) C β 2 (s)ρ(u k,m (s))ds. v k (t) = supu k,m (t), t T, m v k+1 (t) C β 2 (s)ρ(v k (s))ds. limsupv k (t) =, t T. k + Hence, {X m ( ),m } is a Cauchy sequence in L 2 G ([,T];Rn ), set X(t) = (X m (t) X m 1 (t)), m=1 after checking the right side of the iteration, the proof of the existence is complete. 28 / 42

Theorem 2 We assume the following one-sided integral-lipschitz conditions for b, h and σ, i.e., for all x,x 1,x 2 R n and i,j = 1,...d, (H1 ) b(,x), h ij (,x), σ j (,x) M 2 G ([,T];Rn ) are uniformly bounded; (H2 ) 2 x 1 x 2,b(t,x 1 ) b(t,x 2 ) β 2 (t)ρ( x 1 x 2 2 ); 2 x 1 x 2,h ij (t,x 1 ) h ij (t,x 2 ) β 2 (t)ρ( x 1 x 2 2 ); σ j (t,x 1 ) σ j (t,x 2 ) 2 β 2 (t)ρ( x 1 x 2 2 ) where β : [,T] R + is square integrable. Then there exists at most one solution X( ) in L 2 G ([,T],Rn ) to (6). 29 / 42

Proof to Theorem 2: Applying G-Itô s formula to X 1 (t) X 2 (t) 2, we obtain: d( X 1 (t) X 2 (t) 2 ) =2 X 1 (t) X 2 (t),b(t,x 1 (t)) b(t,x 2 (t)) dt +2 X 1 (t) X 2 (t),h ij (t,x 1 (t)) h ij (t,x 2 (t)) d B i,b j t +(σ i (t,x 1 (t)) σ i (t,x 2 (t))) k (σ j (t,x 1 (t)) σ j (t,x 2 (t))) k d B i,b j t +2 X 1 (t) X 2 (t),σ j (t,x 1 (t)) σ j (t,x 2 (t)) db j t, where the repeated indices k, i and j imply the summation and σ j = ((σ j ) 1,,(σ j ) n ) T. From the theorems in Gao [29], we can deduce that Ē[ X 1 (t) X 2 (t) 2 ] C β 2 (s)ρ(ē[ X 1 (s) X 2 (s) 2 ])ds. Finally, Lemma 1 gives the uniqueness result. 3 / 42

Theorem 3 We suppose (H1 ) and the following condition: for any x 1, x 2 R n b(t,x 1 ) b(t,x 2 ) β(t)ρ 1 ( x 1 x 2 ); (H2 ) h ij (t,x 1 ) h ij (t,x 2 ) β(t)ρ 1 ( x 1 x 2 ); σ j (t,x 1 ) σ j (t,x 2 ) 2 β(t)ρ 2 ( x 1 x 2 2 ), where β : [,T] R + is square integrable, ρ 1, ρ 2 : (,+ ) (,+ ) are continuous, concave and increasing, and both of them satisfy (7). Furthermore, we assume that ρ 3 (r) = ρ 2(r 2 ), r (,+ ) r is also continuous, concave and increasing, and ρ 3 (+) =, 1 dr ρ 1 (r) + ρ 3 (r) = +. Then there exists a unique solution to the equation (6). 31 / 42

Example ρ 1 (r) = r ln 1 r ; ρ 2 (r) = r ln 1 r. As for existence, we need some stronger conditions, nevertheless, the example above satisfies the conditions of Theorem 3 but not Theorem 1. 32 / 42

Proof of Theorem 3: Picard Iteration: X m+1 (t) =x + + d i,j=1 b(s,x m (s))ds h ij (s,x m (s))d B i,b j s + d j=1 σ j (s,x m+1 (s))db j s. Because of the assumptions of this theorem and thanks to Theorem 1, the sequence {X m ( ),m } is well defined. Note that as x is not C 2, we approximate x by F ε C 2, where We notice that F ε (x) = ( x 2 + ε) 1 2, x R n. F ε(x) 1, F ε (x) 2, ( x 2 + ε) 1 2 and F ε(x), F ε (x) are bounded and uniformly Lipschitz. 33 / 42

Applying G-Itô formula to F ε (X k+1+m (t) X k+1 (t)), and taking the G-expectation, we get from Theorem 16 that, for some positive constant K, Ē[F ε (X k+1+m (t) X k+1 (t))] Ē[ b(s,x k+m (s)) b(s,x k (s)) ds] d +K i,j=1 Ē[ h ij (s,x k+m (s)) h ij (s,x k (s)) ds] d σ j (s,x +K k+m+1 (s)) σ j (s,x k+1 (s)) 2 Ē[ ds]. j=1 ( X k+m+1 (s) X k+1 (s) 2 + ε) 1 2 Letting ε, we deduce from Lemma 1 and (5) that, for some positive constant C, u k+1,m (t) = sup Ē[ X k+1+m (r) X k+1 (r) ] r t C β(s)(ρ 1 (u k,m (s)) + ρ 3 (u k+1,m (s)))ds. 34 / 42

Set v k (t) = supu k,m (t), t T, m then, v k+1 (t) C β(s)(ρ 1 (v k (s)) + ρ 3 (v k+1 (s)))ds. By Lemma 1, we deduce that limsupv k (t) =, t [,T]. k + Hence, {X m ( ),m } is a Cauchy sequence in L 1 G ([,T],Rn ). Then there exists X( ) L 1 G ([,T],Rn ) and a subsequence {X m l( ),l 1} {X m ( ),m 1} such that X m l X, as l +, q.s.. 35 / 42

Since b, h ij and σ j are bounded, it is easy to check that, for some positive constant M >, and for each P P, sup sup Ē[ X m (t) p ] M, where p > 2, m t T E P ( X(t) p ) =E P (liminf l + Xm l (t) p ) liminf l + E P( X m l (t) p ) liminf sup E P ( X m l (t) p ) = liminfē[ X m l (t) p ] l + P P l + M. Hence, and sup Ē[ X(t) p ] = sup ( sup E P ( X(t) p )) M t T t T P P sup Ē[ X m (t) X(t) p ] 2 p sup Ē[ X m (t) p ] + 2 p sup Ē[ X(t) p ] 2 p+1 M. t T t T t T 36 / 42

Consequencely, for a fixed ε >, limsupē[ X m (t) X(t) 2 ] m + limsup(ε 2 Ē[I { X m (t) X(t) ε}] + Ē[ X m (t) X(t) 2 I { X m (t) X(t) ε}]) m + ε 2 + limsup(ē[ X m (t) X(t) p ]) 2 p (Ē[ I { X m (t) X(t) ε} p p 2 ]) p 2 p m + ε 2 + 8M 2 p limsup(ē[i { X m (t) X(t) ε}]) p 2 p m + =ε 2. The last step above can be easily deduced from lim m + (sup t T Ē X m (t) X(t) ) =. Since ε can be arbitrary small, we have lim m + Ē[ X m (t) X(t) 2 ] =. 37 / 42

On the other hand, since ρ 1 : (,+ ) (,+ ) are continuous, concave and increasing, then for arbitrary fixed ε >, there exists a constant K ε, such that ρ 1 (x) K ε x, for x > ε. Hence, for some positive constant C, we have lim Ē[ sup (b(s,x m (s)) b(s,x(s)))ds 2 ] m t T T C lim β 2 (s)ē[ρ m 1 2 ( Xm (s) X(s) )]ds, ( T C lim β 2 (s)ē[ρ m 1 2 ( Xm (s) X(s) )I { X m (s) X(s) >ε}]ds T ) +ρ1 2 (ε2 ) β 2 (s)ē[i { X m (s) X(s) ε}]ds C lim m ( T T =Cρ1 2 (ε2 ) β 2 (s)ds ) T β 2 (s)ē[kε 2 X m (s) X(s) 2 ]]ds + Cρ1 2 (ε2 ) β 2 (s)ds 38 / 42

Notice that ρ 1 is continuous, ρ 1 (+) = and ε can arbitrary small, then we have lim Ē[ sup (b(s,x m (s)) b(s,x(s)))ds 2 ] =. m t T Similarly we get lim Ē[ sup (h ij (s,x m (s)) h ij (s,x(s)))d B i,b j s 2 ] = m t T and lim Ē[ sup (σ j (s,x m (s)) σ j (s,x(s)))db j s 2 ] m + t T T lim C β 2 (s)ρ 2 (Ē[ X m (s) X(s) 2 ])ds m + =. Then the proof of the existence of the solution to (6) is complete. 39 / 42

4. Solvability of G-BSDE with Integral-Lipschitz Coefficients Consider the following type of G-backward stochastic differential equation (G-BSDE): T Y t = E[ξ + t f (s,y s )ds + d T i,j=1 t g ij (s,y s )d B i,b j s F t ], t [,T]. (1) f (,x),g ij (,x) M 1 G ([,T];Rn ); g(s,y) + f (s,y) β(t) + c y ; g(s,y 1 ) g(s,y 2 ) + f (s,y 1 ) f (s,y 2 ) ρ( y 1 y 2 ). where c >, β M 1 G ([,T];R +) and ρ : (,+ ) (,+ ) is a continuous, concave, increasing function satisfying (7). Theorem 4 Under the assumptions above, (1) admits a unique solution Y L 1 G ([,T],Rn ). 4 / 42

Proof to Theorem 4: Due to the sub-additivity property of E[ F t ], we obtain: T Yt 1 Yt 2 E[ (f (s,ys 1 ) f (s,ys 2 ))ds t d T + (g ij (s,ys 1 ) g ij (s,ys 2 ))d B i,b j s F t ] i,j=1 t Taking the G-expectation on both sides, we have from theorem in Gao [29] and Lemma 2 that T E[ Yt 1 Yt 2 ] (E[ (f (s,ys 1 ) f (s,ys 2 ))ds t d T + E[ (g ij (s,ys 1 ) g ij (s,ys 2 ))d B i,b j s ]) i,j=1 t T KE ρ( Ys 1 Ys 2 )ds t T K ρ(e[ Ys 1 Ys 2 ])ds. t 41 / 42

Set u(t) = E[ Y 1 t Y 2 t ], then T u(t) K ρ(u(s))ds, t and we deduce from Lemma 1 that u(t) =. Then the uniqueness of the solution can be now easily proved. As for the existence of solution, we proceed as Theorem 1: define a sequence of (Y m,m ), as follows: T Yt m+1 = E[ξ + t f (s,y m s )ds + d T i,j=1 t g ij (s,y m s )d B i,b j s F t ], Y =. Then the rest of the proof goes in a similar way as that in Theorem 1. Acknowledgement The authors express special thanks to Prof. Ying Hu, who provided both the initial inspiration for the work and useful suggestions. 42 / 42