An Overview of Risk-Sensitive Stochastic Optimal Control

Similar documents
Feedback Control of Quantum Systems. Matt James ANU

Risk-Sensitive Control with HARA Utility

Robustness and Risk-Sensitive Control of Quantum Systems. Matt James Australian National University

Continuous dependence estimates for the ergodic problem with an application to homogenization

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Thuong Nguyen. SADCO Internal Review Metting

Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations

Linear-Quadratic Stochastic Differential Games with General Noise Processes

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

Risk-Sensitive Filtering and Smoothing via Reference Probability Methods

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

Lipschitz continuity for solutions of Hamilton-Jacobi equation with Ornstein-Uhlenbeck operator

The principle of least action and two-point boundary value problems in orbital mechanics

Risk-Sensitive and Robust Mean Field Games

Linear Quadratic Zero-Sum Two-Person Differential Games

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations

Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control

H 1 optimisation. Is hoped that the practical advantages of receding horizon control might be combined with the robustness advantages of H 1 control.

H CONTROL FOR NONLINEAR STOCHASTIC SYSTEMS: THE OUTPUT-FEEDBACK CASE. Nadav Berman,1 Uri Shaked,2

On differential games with long-time-average cost

Bridging the Gap between Center and Tail for Multiscale Processes

HJ equations. Reachability analysis. Optimal control problems

Nonlinear L 2 -gain analysis via a cascade

SUB-OPTIMAL RISK-SENSITIVE FILTERING FOR THIRD DEGREE POLYNOMIAL STOCHASTIC SYSTEMS

Stochastic Nonlinear Stabilization Part II: Inverse Optimality Hua Deng and Miroslav Krstic Department of Mechanical Engineering h

On continuous time contract theory

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem

Thomas Knispel Leibniz Universität Hannover

Robust control and applications in economic theory

1. Numerical Methods for Stochastic Control Problems in Continuous Time (with H. J. Kushner), Second Revised Edition, Springer-Verlag, New York, 2001.

Sébastien Chaumont a a Institut Élie Cartan, Université Henri Poincaré Nancy I, B. P. 239, Vandoeuvre-lès-Nancy Cedex, France. 1.

The main motivation of this paper comes from the following, rather surprising, result of Ecker and Huisken [13]: for any initial data u 0 W 1,

Mean-Field Games with non-convex Hamiltonian

Closed-Loop Impulse Control of Oscillating Systems

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables

Stochastic optimal control with rough paths

On the Stability of the Best Reply Map for Noncooperative Differential Games

Chain differentials with an application to the mathematical fear operator

Master Thesis. Nguyen Tien Thinh. Homogenization and Viscosity solution

Stochastic contraction BACS Workshop Chamonix, January 14, 2008

An Introduction to Noncooperative Games

Viscosity Solutions for Dummies (including Economists)

Controlled Markov Processes and Viscosity Solutions

APPROXIMATIONS FOR NONLINEAR FILTERING

Large Deviations Theory, Induced Log-Plus and Max-Plus Measures and their Applications

Calculating the domain of attraction: Zubov s method and extensions

On the stability of receding horizon control with a general terminal cost

Asymptotic Perron Method for Stochastic Games and Control

Weak convergence and large deviation theory

Solution of Stochastic Optimal Control Problems and Financial Applications

Dynamical systems with Gaussian and Levy noise: analytical and stochastic approaches

Characterizing controllability probabilities of stochastic control systems via Zubov s method

Shape from Shading with discontinuous image brightness

A Model of Optimal Portfolio Selection under. Liquidity Risk and Price Impact

An Uncertain Control Model with Application to. Production-Inventory System

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations

A generalization of Zubov s method to perturbed systems

Mean Field Games on networks

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

I R TECHNICAL RESEARCH REPORT. Risk-Sensitive Probability for Markov Chains. by Vahid R. Ramezani and Steven I. Marcus TR

INFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION

AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS. 1. Introduction

Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs

Strong uniqueness for stochastic evolution equations with possibly unbounded measurable drift term

Controlled Markov Processes and Viscosity Solutions

A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS

The Convergence of the Minimum Energy Estimator

An iterative procedure for constructing subsolutions of discrete-time optimal control problems

Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini

A Narrow-Stencil Finite Difference Method for Hamilton-Jacobi-Bellman Equations

A probabilistic approach to second order variational inequalities with bilateral constraints

Markov Chain BSDEs and risk averse networks

Zubov s method for perturbed differential equations

Exam February h

Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations.

mü/wm 9rwsm TITLE AND SUBTITLE COMPUTATIONAL NONLINEAR CONTROL

Numerical methods for Mean Field Games: additional material

Robust Control of Linear Quantum Systems

Prof. Erhan Bayraktar (University of Michigan)

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt +

LMI Methods in Optimal and Robust Control

Balancing of Lossless and Passive Systems

HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS. CIME Courses-Cetraro August 29-September COURSES

The Important State Coordinates of a Nonlinear System

Optimal Control Using an Algebraic Method for Control-Affine Nonlinear Systems. Abstract

Nonlinear Control Systems

L -Bounded Robust Control of Nonlinear Cascade Systems

Deterministic Minimax Impulse Control

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points

On the infinity Laplace operator

AN ASYMPTOTIC MEAN VALUE CHARACTERIZATION FOR p-harmonic FUNCTIONS. To the memory of our friend and colleague Fuensanta Andreu

Introduction to Nonlinear Control Lecture # 3 Time-Varying and Perturbed Systems

Dynamical properties of Hamilton Jacobi equations via the nonlinear adjoint method: Large time behavior and Discounted approximation

High-Gain Observers in Nonlinear Feedback Control

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES

State-Feedback Optimal Controllers for Deterministic Nonlinear Systems

Uniformly Uniformly-ergodic Markov chains and BSDEs

Transcription:

An Overview of Risk-Sensitive Stochastic Optimal Control Matt James Australian National University Workshop: Stochastic control, communications and numerics perspective UniSA, Adelaide, Sept. 2004. 1

Outline A brief history Approaches to controller design Relationships among approaches Some limit results (state feedback) Robust interpretation Partially observed case Quantum systems References 2

A brief history The risk-sensitive criterion introduced by Jacobson 1973. Linear system with Gaussian noise: dx = (Ax + Bu)dt + dw [Linear-Exponential-Quadratic-Gaussian (LEQG)] Minimize average of exponential of quadratic cost: ( )] T J µ 1 = E x,t [expµ 2 [x Qx + 1 2 u 2 ] ds + 1 2 x Rx µ > 0 - risk-sensitive µ < 0 - risk-seeking t 3

solution Optimal state feedback control is u (x, t) = B P µ t x where P µ t = P µ t A A P µ t + P µ t (BB µi) P µ t Q, P µ T = R 0 t T [control type Riccati equation] 4

Limit µ 0: recover standard LQG control (risk-neutral). Linear system with Gaussian noise: dx = (Ax + Bu)dt + dw Minimize average of quadratic cost: E x,t [ T t 1 2 [x Qx + 1 2 u 2 ] ds + 1 2 x Rx Solution: Optimal state feedback control is u (x, t) = B P t x [Linear-Quadratic-Gaussian (LQG)] ] where P t = P t A A P t + P t BB P t Q, P T = R 5

partially observed case Linear system with Gaussian noise: dx dy = (Ax + Bu)dt + dw = Cxdt + dv Minimize average of exponential of quadratic cost ( )] T J µ 1 = E x,t [expµ 2 [x Qx + 1 2 u 2 ] ds + 1 2 x Rx over causal maps y( ) u( ). t 6

Solution: Whittle 1981, Bensoussan-van Schuppen 1985. Optimal partially observed control is u t = B P µ t ˆx µ t where dˆx µ = ([A + µy µ Q]ˆx µ + Bu)dt + Y µ C (dy Cˆx µ dt) and Ẏ µ = AY µ + Y µ A + µy µ QY µ + I Y µ C CY µ Significantly, this is not the Kalman filter. When µ 0, LEQG LQG, whose solution is expressed in terms of the Kalman filter. 7

Some subsequent developments: LEQG. Connection with dynamic games. (Jacobson, 1973) LEQG. Connection with robust control (H ). (Glover-Doyle, 1988) Risk-sensitive control for nonlinear systems, state feedback. Connection with dynamic games. (Fleming-McEneaney 1992, James, 1992, Nagai 1996) Solution of risk-sensitive control for nonlinear systems, output feedback. Solution and connection with dynamic games and nonlinear robust control. (James-Baras-Elliott 1994, Charalambous-Hibey 1996, Dai Pra-Meneghini-Runggaldier 1996. ) Robustness properties of risk-sensitive criterion. Also stochastic IQC. (James, Fleming, Boel, Dupuis, Petersen, 1995, 1998) Risk-sensitive control for finance. (Nagai 2000) Risk-sensitive control for quantum systems. (James 2004) 8

Approaches to controller design The basic aim of robust control is: Design a controller to achieve satisfactory performance in the presence of disturbances and uncertainty. Stochastic control has similar aims. 9

Dynamics ẋ s = f(x s ) + g(x s )d s, 0 < s < T, z s = h(x s ) H robust control Disturbance Output Map f(0) = 0, h(0) = 0, x 0 = 0 H Norm Note: T d, z γ iff T d, z : d( ) z( ) T d, z = T 0 z(t) 2 dt sup d L 2 [0,T] T 0 z 2 d 2 γ 2 d(t) 2 dt d L 2 [0, T] 10

D-H 2 deterministic optimal control disturbance: none controller: Optimal cost gives optimal performance without special consideration of disturbances. W(x, t) = inf u [ T t L(x s, u s ) ds + Φ(x T ) ] Dynamics ẋ s = b(x s, u s ), t < s < T, x t = x 11

S-H 2 (RN) stochastic optimal control disturbance: controller: Optimal cost noise gives optimal average performance W ε (x, t) = inf u E x,t [ T t L(x ε s, u s ) ds + Φ(x T ) ] Dynamics dx ε s = b(x ε s, u s ) ds + ε db s, t < s < T, x ε t = x (ε > 0 - noise intensity) 12

RS disturbance: noise stochastic risk-sensitive optimal control controller: gives optimal average performance using exponential cost (heavily penalizes large values) Optimal cost S µ,ε (x, t) = inf u E x,t [ exp µ ε ( T t )] L(x ε s, u s ) ds + Φ(x ε T) Dynamics dx ε s = b(x ε s, u s ) ds + ε db s, t < s < T, x ε t = x (µ > 0 - risk sensitivity) 13

H H robust control disturbance: L 2 signal controller: Find u( ) such that Dynamics achieves specified H norm constraint (worst-case design) T d, z γ ẋ s = b(x s, u s ) + d s t < s < T x t = x z s = h(x s ) (γ = 1/ µ - H constraint) 14

D-DG deterministic differential game disturbance: L 2 signal controller: Optimal cost Dynamics W µ (x, t) = sup d gives optimal performance subject to opposing efforts of disturbance (worst-case design) ẋ s x t inf u [ T t ] L(x s, u s ) 1 d 2µ s 2 ds + Φ(x T ) = b(x s, u s ) + d s, t < s < T = x 15

S-DG stochastic differential game disturbances: controller: Optimal cost noise W µ,ε (x, t) = sup d L 2 signal gives optimal average performance subject to opposing efforts of disturbance (worst-case design) inf u E x,t [ T t L(x ε s, u s ) 1 2µ d s 2 ds + Φ(x ε T) ] Dynamics dx ε s x ε t = (b(x ε s, u s ) + d s ) ds + ε db s, t < s < T = x 16

Relationships among approaches RS µ 0 S-H 2 ε 0 ε 0 D-DG µ 0 D-H 2 RS, S-DG, D-DG, and S-H 2 are perturbations of D-H 2. 17

(James, Fleming-McEneaney, Whittle) RS S-DG D-H 2 + (noise variance term) + (disturbance energy term) Equivalence via logarithmic transformation, dynamic programming PDEs, and representation theorem Expansion valid for small noise intensity and small risk-sensitivity 18

S-H 2 D-H 2 + (noise variance term) H D-H 2 + (disturbance energy term) RS H 2 + H??? 19

Some limit results (state feedback) Assumptions b(x, u) = f(x) + g(x)u, where f : R n R n, g : R m R n and their first order derivatives Df, Dg are bounded and uniformly Lipschitz continuous. Control values: U R m is compact and convex. Disturbance values: D = R n. Φ : R n R is non negative; Φ and DΦ are bounded and uniformly Lipschitz continuous. 20

L : R n R m R is C 2, non negative; L(, u) and D x L(, u) are bounded and uniformly Lipschitz continuous, uniformly in u U; and D 2 ul(x, ) > 0, uniformly in x R n. There exists locally Lipschitz U (x, p) achieving maximum in min u U [p b(x, u) + L(x, u)] Use Elliott Kalton formulation for two player zero sum differential games (strategies, upper and lower values, etc) For later use, assume also L(x, u) = 1 2 u 2 + V (x), and the functions f, g, V and Φ are of class C and have compact support. U (x, p) is of class C 2. 21

Viscosity Solutions Viscosity solutions introduced by Crandall-Lions 1983. The definition for a fully nonlinear parabolic PDE v t + F(x, Dv, D2 v) = 0 in R n (0, T) v(x, T) = Φ(x) in R n. is as follows. 22

Definition An upper semicontinuous (u.s.c.) function v (resp. l.s.c. function v) is called a viscosity subsolution (resp. supersolution) if and v Φ (resp. v Φ) on R n {T } t φ(x, t) + F(x, Dφ(x, t), D2 φ(x, t)) 0 (resp. 0) for every smooth function φ and any local maximum (resp. minimum) (x, t) R n [0, T) of v φ (resp. v φ). A continuous function is called a viscosity solution if it is both a subsolution and a supersolution. 23

Comparison Theorem If v is a subsolution and if v is a supersolution, then v v in R n [0, T]. Thus any continuous viscosity solution is unique. 24

Dynamic Programming PDEs RS (ε > 0, µ > 0) S µ,ε t + ε 2 Sµ,ε + min u U [DS µ,ε b(x, u) + µ ε L(x, u)sµ,ε ] = 0 in R n (0, T) S µ,ε (x, T) = exp µ ε Φ(x) in Rn Logarithmic transformation: (Fleming, 1970 s) W µ,ε (x, t) = ε µ log Sµ,ε (x, t) 25

S-DG (ε > 0, µ > 0) W µ,ε t + ε 2 W µ,ε + min u U max d D [DW µ,ε (b(x, u) + d) + L(x, u) 1 2µ d 2 ] = 0 in R n (0, T) W µ,ε (x, T) = Φ(x) in R n. Note: min u U max d D [ p (b(x, u) + d) + L(x, u) 1 2µ d 2 ] = min u U [p b(x, u) + L(x, u)] + 1 2 µ p 2 26

Representation theorem Viscosity formulation of stochastic differential games developed by Fleming Souganidis (1989). Theorem (James, 1992; Fleming-McEneany 1992) The function W µ,ε defined by the logarithmic transformation is the optimal cost function of the stochastic differential game defined above. i.e. RS S-DG Proof : Methods of Fleming (1971) imply DW µ,ε (x, t) C for all (x, t) R n [0, T]. In view of this bound, the set of disturbance values D can be taken as bounded. The result now follows from Fleming Souganidis (1989). 27

Limit PDEs Obtained as ε 0 and/or µ 0. D-DG (ε 0, µ > 0) W µ t + min u U max d D [DW µ (b(x, u) + d) + L(x, u) 1 2µ d 2 ] = 0 in R n (0, T) W µ (x, T) = Φ(x) in R n. 28

S-H 2 (ε > 0, µ 0) W ε t + ε 2 W ε + min u U [DW ε b(x, u) +L(x, u)] = 0 in R n (0, T) W ε (x, T) = Φ(x) in R n. D-H 2 (ε 0, µ 0) W t + min u U [DW b(x, u) + L(x, u)] = 0 in R n (0, T) W(x, T) = Φ(x) in R n. 29

Limit theorems depend on general viscosity limit techniques developed by Evans Ishii (1984) Barles Perthame (1987) Limit theorems Theorem (James, 1992) We have lim W µ,ε (x, t) = W µ (x, t) ε 0 uniformly on compact subsets, where W µ C(R n [0, T]) is the unique bounded continuous viscosity solution of the corresponding PDE and is the optimal cost function of the deterministic differential game defined above. i.e. S-DG D-DG as ε 0. 30

Proof: Assumptions imply The function 0 W µ,ε (x, t) C for all (x, t) R n [0, T], µ, ε > 0 v(x, t) = is u.s.c. and a viscosity subsolution. The function v(x, t) = is l.s.c and a viscosity supersolution. lim sup W µ,ε (y, s). ε 0, y x, s t lim inf W µ,ε (y, s) ε 0, y x, s t By construction v v By the comparison theorem v v 31

Therefore lim ε 0 W µ,ε (x, t) = v = v = v From Evans Souganidis (1984) the value W µ of the D-DG is the unique viscosity solution, and hence W µ = v. Expansion of optimal cost for S-DG W aε,bε (x, t) = W(x, t) + ε (W g (x, t), W n (x, t)) a b +... [methods of Fleming, 1970 s] 32

Robust interpretation (Boel, Dupuis, James, Petersen 1998, 2000) We wish to interpret the finiteness of the risk-sensitive criterion for a nominal system G 0 in terms of a performance bound for a perturbed system G. nominal system w G 0 Z 33

Nominal system dynamics: dx t = G 0 : b(x t )dt + ε 1 2 σ(x t )dw t Z t = h(x t ) Risk-sensitive criterion: S(x, T). = E x [ exp 1 2γ 2 ε T 0 Z t 2 dt ] γ 2 = 1/µ 34

Key tool is the representation [Dupuis-Ellis 1997] [ ] γ 2 1 T ( ε log S(x, T) = sup E x Zt 2 γ 2 v t 2) dt v V T 2 where X t is given by perturbed system dynamics: d G : X t = b( X t )dt + σ( X t )v t dt + ε 1 2 σ( X t )dw t Z t = h( X t ) 0 35

perturbed system w G Z v X perturbation E x [ 1 2 T 0 Z t 2 dt ] γ 2 E x [ 1 2 T 0 v t 2 dt ] robustness inequality + γ 2 ε log S(x, T) (analogous to H inequality) 36

Partially observed (output feedback) case (James-Baras-Elliott, 1994) Dynamics dx t dỹ t = b(x t, u t ) dt + ε dw t = h(x t ) dt + ε dv t Objective γ 2 = 1/µ [ ( )] J(u) = E exp µ T L(x t, u t ) dt + Φ(x T ) ε 0 37

Information state σ µ, ε solution of RS t (x) not cond prob σ µ,ε t, η = E 0 [η(x ε t)z ε t exp ( µ ε t 0 ) L(x ε s, u s ) ds Y t ] for all η Dynamics dσ µ,ε t = ( A u t + µ ) ε Lu t σ µ,ε t dt + 1 ε hσµ,ε t dỹt, ε σ µ,ε 0 = ρ in L 1 (R n ), stochastic PDE Representation J(u) = E 0 [ σ µ,ε T, exp µ ε Φ ] Linear case: Bensoussan-van Schuppen 38

Dynamic programming Value function S µ,ε (σ, t) = inf u Ũt,T E 0 σ,t [ ] σ µ,ε T, e µ ε Φ. Dyn prog equation t Sµ,ε + 1 2ε D2 S µ,ε (hσ, hσ) { + inf u U DS µ,ε (A u σ + µ ε Lu σ )} = 0 S µ,ε (σ, T) = σ, e µ ε Φ Optimal controller u t(σ) = achieves min in d.p.e. infinite dim, nonlinear, 2nd-order information state feedback 39

Small Noise limit Duality max-plus large deviations σ, ν = R n σ(x)ν(x) dx (p, q) = sup x R n { p(x) + q(x) } lim ε 0 ε µ log e µ ε p, e µ ε q = (p, q) 40

Information state Value function lim ε 0 lim ε 0 ε µ log σµ,ε t (x) = p µ t (x), ε µ log Sµ,ε (e µ ε p, t) = W µ (p, t) limit limit Some consequences Can solve output feedback deterministic dynamic games (D-DG) using analogous, though deterministic, information state methods (cf. max-plus). Output feedback nonlinear H control problem can be solved using the D-DG methods. 41

Risk-sensitive control of quantum systems (James, 2004 - current) system (atom) H X bath (EM field) Γ db,db System: eg. atom, can be controlled, Hilbert space H Bath: e.g. electromagnetic field, continuously monitored, Hilbert space Γ (Fock space) 42

Dynamics on H Γ: [Hudson-Parthasarathy quantum stochastic DE, 1980 s] du(t) = { K(u(t))dt + LdB (t) L db(t) } U(t) with initial condition U(0) = I, where K(u) = i H(u) + 1 2 L L System operators evolve according to X(t) = j t (u, X) = U (t)x IU(t) i.e. dx(t) + (X(t)K(t) + K (t)x(t) L (t)x(t)l(t))dt = [X(t), L(t)]dB (t) [X(t), L (t)]db(t) where L(t) = j t (u, L), K(t) = j t (u, K(u(t))) 43

Measurements of real quadrature of field: Q t = B t + B t dy (t) = (L(t) + L (t))dt+ dq(t) output field input field Risk-sensitive cost: [James, 2004] J µ = π 0 vv, R (T)e µc2(t) R(T) where dr(t) dt [ A, B = tr(a B)] = µ 2 C 1(t)R(t), C 1 (t) = j t (u, C 1 (u(t))), C 2 (t) = j t (u, C 2 ) π 0 = initial system state vv = field vacuum state 44

Stochastic representation: J µ = E 0 [ σ µ T, eµc 2 ] where P 0 is standard Wiener measure, and the information state (an unnormalized density operator) satisfies dσ µ t +[Kσ µ t +σ µ t K Lσ t L ]dt = µ 1 2 [C 1σ µ t +σ µ t C 1 ]dt+[lσ µ t +σ µ t L ]dy(t) [modified stochastic master equation] π µ t = σµ t σ t, 1 [unnormalized state, new to physics] When µ 0 recover the Belavkin quantum filtering equation (stochastic master equation) (1980 s) [unnormalized state σ t, normalized state π t = σ t / σ t, 1 ]. 45

The dynamic programming equation is, formally, t Sµ (σ, t) + inf u U L µ;u S µ (σ, t) Discrete time (James 2004) = 0, 0 t < T S µ (σ, T) = σ, e µc 2 Some related current work Robustness interpretation, discrete time (James-Petersen 2004) Robustness, linear/quadratic case (Doherty, d Helon, James, Petersen, Wilson) 46

References [1] T. Basar and P. Bernhard. H Optimal Control and Related Minimax Design Problems. Birkhauser, Boston, 1995. [2] A. Bensoussan and J. H. van Schuppen. Optimal control of partially observable stochastic systems with an exponential-of-integral performance index. SIAM Journal on Control and Optimization, 23:599 613, 1985. [3] P. Dupuis and R.S. Ellis. A Weak Convergence Approach to the Theory of Large Deviations. Wiley, New York, 1997. [4] W.H. Fleming and W.M. McEneaney. Risk-sensitive control on an infinite time horizon. SIAM J. Control and Optimization, 33:1881 1915, 1995. [5] K. Glover and J. Doyle. State space formulea for all stabilizing controllers that statisfy an H norm bound and relations to risk-sensitivity. Systems and Control Letters, 11:167 172, 1988. [6] J.W. Helton and M.R. James. Extending H Control to Nonlinear Systems: Control of Nonlinear Systems to Achieve Performance Objectives, volume 1 of Advances in Design and Control. SIAM, Philadelphia, 1999. [7] D.H. Jacobson. Optimal stochastic linear systems with exponential performance criteria and their relation to deterministic differential games. IEEE Trans. Aut. Control, 18(2):124 131, 1973. [8] M. R. James, J.S. Baras, and R.J. Elliott. Risk-sensitive control and dynamic games for partially observed discrete-time nonlinear systems. IEEE Trans. Automatic Control, 39(4):780 792, 1994. [9] M.R. James. Asymptotic analysis of nonlinear stochastic risk-sensitive control and differential games. Mathematics of Control, Signals and Systems, 5(4):401 417, 1992. [10] M.R. James. Risk-sensitive optimal control of quantum systems. Phys. Rev. A, 69:032108, 2004. 47

[11] P. Dai Pra, L. Meneghini, and W.J. Runggaldier. Connections between stochastic control and dynamic games. Math. Control, Signals and Systems, 9(4):303 326, 1996. [12] A.J. van der Schaft. L 2 -Gain and Passivity Techniques in Nonlinear Control. Springer Verlag, New York, 1996. [13] P. Whittle. Risk-sensitive linear quadratic Gaussian control. Advances in Applied Probability, 13:764 777, 1981. [14] G. Zames. On the input-output stability of time-varying nonlinear feedback systems. part i: Conditions derived using concepts of loop gain, conicity, and positivity. part ii: Conditions involving circles in the frequency plane and sector nonlinearities. IEEE Trans. Aut. Control, 11:228 238, 465 476, 1966. [15] G. Zames. Feedback and optimal sensitivity: Model reference transformation, multiplicative seminorms and approximate inverses. IEEE Trans. Aut. Control, 26:301 320, 1981. [16] C.D. Charalambous and J.L. Hibey, Minimum principle for partially observable nonlinear risk-sensitive control problems using measure-valued decompositions, Stochastics and Stochastics Reports, vol. 57, no. 3+4, pp. 247-288, August 1996. [17] W.H. Fleming and M.R. James, The Risk-Sensitive Index and the H 2 and H Norms for Nonlinear Systems, Mathematics of Control, Signals, and Systems, 8(3), 1995, 199-221. [18] H. Nagai, Bellman Equation of Risk-Sensitive Control, SIAM J. Control Optim., 34(1), 1996, 74-101. [19] M.G. Crandall, L.C. Evans and P.L. Lions, Some Properties of Viscosity Solutions of Hamilton Jacobi Equations, Trans. AMS, 282 (2) (1984) 487 502. [20] M.G. Crandall, H. Ishii and P.L. Lions, User s Guide to Viscosity Solutions of Second Order Partial Differential Equations, CEREMADE Report No. 9039, (1990). [21] P. Dupuis, M.R. James, and I.R. Petersen, Robust Properties of Risk-Sensitive Control, Math. Control, Systems and Signals, 13, 318-332, 2000. [22] M. Bardi and I. Capuzzo-Dolcetta, Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations, Birkhauser, Boston, 1997. 48

[23] W.H. Fleming and H.M. Soner, Controlled Markov Processes and Viscosity Solutions, Springer, New York, 1993. 49