Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini

Similar documents
The HJB-POD approach for infinite dimensional control problems

Proper Orthogonal Decomposition in PDE-Constrained Optimization

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle

An adaptive RBF-based Semi-Lagrangian scheme for HJB equations

POD for Parametric PDEs and for Optimality Systems

Suboptimal feedback control of PDEs by solving Hamilton-Jacobi Bellman equations on sparse grids

Proper Orthogonal Decomposition for Optimal Control Problems with Mixed Control-State Constraints

arxiv: v1 [math.na] 29 Jul 2018

Proper Orthogonal Decomposition. POD for PDE Constrained Optimization. Stefan Volkwein

arxiv: v1 [math.na] 28 Dec 2018

Proper Orthogonal Decomposition (POD) for Nonlinear Dynamical Systems. Stefan Volkwein

Numerical Optimal Control Overview. Moritz Diehl

Suboptimal Open-loop Control Using POD. Stefan Volkwein

An optimal control problem for a parabolic PDE with control constraints

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations

Suboptimal feedback control of PDEs by solving HJB equations on adaptive sparse grids

Fast Marching Methods for Front Propagation Lecture 1/3

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

Filtered scheme and error estimate for first order Hamilton-Jacobi equations

ESC794: Special Topics: Model Predictive Control

Reduced-Order Greedy Controllability of Finite Dimensional Linear Systems. Giulia Fabrini Laura Iapichino Stefan Volkwein

Asymptotic Stability of POD based Model Predictive Control for a semilinear parabolic PDE

INTERPRETATION OF PROPER ORTHOGONAL DECOMPOSITION AS SINGULAR VALUE DECOMPOSITION AND HJB-BASED FEEDBACK DESIGN PAPER MSC-506

Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0

Model order reduction & PDE constrained optimization

pacman A classical arcade videogame powered by Hamilton-Jacobi equations Simone Cacace Universita degli Studi Roma Tre

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs

Eileen Kammann Fredi Tröltzsch Stefan Volkwein

Suboptimal feedback control of PDEs by solving HJB equations on adaptive sparse grids

Optimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.

Reduced basis method for efficient model order reduction of optimal control problems

Reduced-Order Model Development and Control Design

Proper Orthogonal Decomposition (POD)

Adaptive methods for control problems with finite-dimensional control space

Numerical methods for nonlinear optimal control problems

A Posteriori Estimates for Cost Functionals of Optimal Control Problems

Chapter 5. Pontryagin s Minimum Principle (Constrained OCP)

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

arxiv: v1 [math.na] 13 Sep 2014

Optimal Control Theory

Mean field games and related models

Deterministic Dynamic Programming

Computational Issues in Nonlinear Dynamics and Control

The Direct Transcription Method For Optimal Control. Part 2: Optimal Control

Solving Distributed Optimal Control Problems for the Unsteady Burgers Equation in COMSOL Multiphysics

minimize x subject to (x 2)(x 4) u,

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

Suboptimality of minmax MPC. Seungho Lee. ẋ(t) = f(x(t), u(t)), x(0) = x 0, t 0 (1)

Static and Dynamic Optimization (42111)

Control Theory and PDE s

Taylor expansions for the HJB equation associated with a bilinear control problem

Characterizing controllability probabilities of stochastic control systems via Zubov s method

ON CONVERGENCE OF A RECEDING HORIZON METHOD FOR PARABOLIC BOUNDARY CONTROL. fredi tröltzsch and daniel wachsmuth 1

Polynomial approximation of high-dimensional Hamilton-Jacobi-Bellman equations and applications to feedback control of semilinear parabolic PDEs

Multiobjective PDE-constrained optimization using the reduced basis method

Constrained Minimization and Multigrid

Solution of Stochastic Optimal Control Problems and Financial Applications

Bang bang control of elliptic and parabolic PDEs

MATH4406 (Control Theory) Unit 1: Introduction Prepared by Yoni Nazarathy, July 21, 2012

IEOR 265 Lecture 14 (Robust) Linear Tube MPC

Robotics. Control Theory. Marc Toussaint U Stuttgart

Matrix stabilization using differential equations.

HJ equations. Reachability analysis. Optimal control problems

Hamburger Beiträge zur Angewandten Mathematik

ACM/CMS 107 Linear Analysis & Applications Fall 2016 Assignment 4: Linear ODEs and Control Theory Due: 5th December 2016

POD/DEIM 4DVAR Data Assimilation of the Shallow Water Equation Model

Alberto Bressan. Department of Mathematics, Penn State University

Direct Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U.

L 2 -induced Gains of Switched Systems and Classes of Switching Signals

Priority Programme The Combination of POD Model Reduction with Adaptive Finite Element Methods in the Context of Phase Field Models

Newton s Method and Efficient, Robust Variants

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables

Game Theory Extra Lecture 1 (BoB)

Minimal time mean field games

Proper Orthogonal Decomposition Surrogate Models for Nonlinear Dynamical Systems: Error Estimates and Suboptimal Control

Stabilization of Heat Equation

Lipschitz continuity for solutions of Hamilton-Jacobi equation with Ornstein-Uhlenbeck operator

Snapshot location by error equilibration in proper orthogonal decomposition for linear and semilinear parabolic partial differential equations

UCLA Chemical Engineering. Process & Control Systems Engineering Laboratory

Inexact Newton Methods and Nonlinear Constrained Optimization

On the Stability of the Best Reply Map for Noncooperative Differential Games

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

Exam February h

Pontryagin s maximum principle

Numerical Optimization Algorithms

Reduced-Order Methods in PDE Constrained Optimization

Chapter 2 Optimal Control Problem

LECTURE 3: DISCRETE GRADIENT FLOWS

On differential games with long-time-average cost

CONVERGENT DIFFERENCE SCHEMES FOR DEGENERATE ELLIPTIC AND PARABOLIC EQUATIONS: HAMILTON JACOBI EQUATIONS AND FREE BOUNDARY PROBLEMS

RICE UNIVERSITY. Nonlinear Model Reduction via Discrete Empirical Interpolation by Saifon Chaturantabut

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

Implicit-explicit exponential integrators

Termination criteria for inexact fixed point methods

A model order reduction technique for speeding up computational homogenisation

A Second Order Discontinuous Galerkin Fast Sweeping Method for Eikonal Equations

Numerical algorithms for one and two target optimal controls

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

CHAPTER 5: Linear Multistep Methods

Calculating the domain of attraction: Zubov s method and extensions

Transcription:

Numerical approximation for optimal control problems via MPC and HJB Giulia Fabrini Konstanz Women In Mathematics 15 May, 2018 G. Fabrini (University of Konstanz) Numerical approximation for OCP 1 / 33

Outline Outline 1 Introduction and motivations 2 Hamilton-Jacobi-Bellman 3 Model Predictive Control 4 Numerical Tests G. Fabrini (University of Konstanz) Numerical approximation for OCP 2 / 33

Introduction and motivations Outline 1 Introduction and motivations 2 Hamilton-Jacobi-Bellman 3 Model Predictive Control 4 Numerical Tests G. Fabrini (University of Konstanz) Numerical approximation for OCP 3 / 33

Introduction and motivations Introduction and Motivations Some history The method is largely due to the work of Lev Pontryagin and Richard Bellman in the 1950s. The theory of control analyzes the properties of controlled systems, i.e. dynamical systems on which we can act through a control. Aim Bring the system from an initial state to a certain final state satisfying some criteria. Several applications: Robotics, aeronautics, electrical and aerospace engineering, biological and medical field. G. Fabrini (University of Konstanz) Numerical approximation for OCP 4 / 33

Introduction and motivations Controlled dynamical system { ẏ(t) = F(y(t), u(t)) (t > 0) y(0) = x (x R n ) Assumptions on the data u( ) U: control, where U := {u( ) : [0, + [ U measurable} F : R n U R n is the dynamics, which satisfies: F is continuous respect to (y, u) F is local bounded F is Lipschitz respect to y!y x (t, u), solution of the problem (Caratheodory Theorem). G. Fabrini (University of Konstanz) Numerical approximation for OCP 5 / 33

Introduction and motivations The infinite horizon problem Cost functional J(x, u) := 0 L(y x (t, u), u(t))e λt dt. where λ > 0 is the interest rate and L is the running cost. Goal We are interested in minimizing this cost functional. We want to find an optimal pair (y, u ) which minimizes the cost functional. G. Fabrini (University of Konstanz) Numerical approximation for OCP 6 / 33

Introduction and motivations Two possible approaches Open-loop control Control expressed as functions of time t (necessary condition - Pontryagin Minimum Principle or direct methods, e.g. gradient method). Remark: it cannot take into account errors in the real state of the system due to model errors or external disturbances. Feedback control Control expressed as functions of the state system (Dynamic Programming, Model Predictive Control) Remark: robust to external perturbations. G. Fabrini (University of Konstanz) Numerical approximation for OCP 7 / 33

Hamilton-Jacobi-Bellman Outline 1 Introduction and motivations 2 Hamilton-Jacobi-Bellman 3 Model Predictive Control 4 Numerical Tests G. Fabrini (University of Konstanz) Numerical approximation for OCP 8 / 33

Hamilton-Jacobi-Bellman Value Function v(y 0 ) := inf J(y 0, u( )) u( ) U The value function is the unique viscosity solution of the Bellman equation associated to the control problem via Dynamic Programming. Dynamic Programming Principle { t } v(x) = min L(y x (s), u(s))e λs ds + v(y x (t))e λt u U Hamilton-Jacobi-Bellman t 0 λv(x) + max{ f (x, u) v(x) L(x, u)} = 0, x Ω, u U u U G. Fabrini (University of Konstanz) Numerical approximation for OCP 9 / 33

Feedback Control Hamilton-Jacobi-Bellman Given v(x) for any x R n, we define u(yx (t))) = arg min[f(x, u) v(x) + L(x, u)] u U where y (t) is the solution of { ẏ (t) = F(y (t), u (t)), t (t 0, ] y(t 0 ) = x, Technical difficulties The bottleneck is the approximation of the value function v, however this remains the main goal since v allows to get back to feedback controls in a rather simple way. G. Fabrini (University of Konstanz) Numerical approximation for OCP 10 / 33

Hamilton-Jacobi-Bellman Numerical computation of the value function The bottleneck of the DP approach is the computation of the value function, since this requires to solve a non linear PDE in high-dimension. This is a challenging problem due to the huge number of nodes involved and to the singularities of the solution. Another important issue is the choice of the domain. Several numerical schemes: finite difference, finite volume, Semi-Lagrangian (obtained by a discrete version of the Dynamic Programming Principle ). G. Fabrini (University of Konstanz) Numerical approximation for OCP 11 / 33

Hamilton-Jacobi-Bellman Semi-Lagrangian discretization of HJB These schemes are based on the direct discretization of the directional derivative f (x, u) v(x). Continuous Version λv(x) = max{ F(x, u) Dv(x) L(x, u)} u U Semi-Discrete Approximation (Value Iteration) Making a discretization in time of the continuous control problem (e.g. using the Euler method): V k+1 = min u U {e λ t V k (x + t F (x, u)) + t L (x i, u)} G. Fabrini (University of Konstanz) Numerical approximation for OCP 12 / 33

Hamilton-Jacobi-Bellman Semi-Lagrangian discretization of HJB Fully discrete SL Value Iteration (VI) scheme ( T (V k ) V k+1 ) = T (V k ), for i = 1,..., N, with i min {βi 1[V k ](x i + t F(x i, u)) + L(x i, u)}. u U Fixed grid in Ω R n bounded, Steps x. Nodes: {x 1,..., x N } Stability for large time steps t. Error estimation: (Falcone/Ferretti 97) Slow convergence, since β = e λ t 1 when t 0 PROBLEM: Find a reasonable computational domain. G. Fabrini (University of Konstanz) Numerical approximation for OCP 13 / 33

Hamilton-Jacobi-Bellman Value Iteration for infinite horizon optimal control (VI) Require: Mesh G, t, initial guess V 0, tolerance ɛ. 1: while V k+1 V k ɛ do 2: for x i G do 3: V k+1 i 4: k = k + 1 5: end for 6: end while Remarks [ = min u U {e λ t I V k] (x i + t F (x i, u)) + t L (x i, u)} VI algorithm converges (slowly) for any initial guess V 0. We can provide an error estimate, G. Fabrini (University of Konstanz) Numerical approximation for OCP 14 / 33

Model Predictive Control Outline 1 Introduction and motivations 2 Hamilton-Jacobi-Bellman 3 Model Predictive Control 4 Numerical Tests G. Fabrini (University of Konstanz) Numerical approximation for OCP 15 / 33

Model Predictive Control Model Predictive Control Dynamics: { ẏ(t) = F(y(t), u(t)) y(0) = y 0 t > 0 Infinite horizon cost functional: J (u( )) = Finite horizon cost functional: J N (u( )) = t N 0 0 L(y(t; u, y 0 ))e λt dt t 0 L(y(t; u, y 0 ))e λt dt, t N 0 = t 0 + N t, N N G. Fabrini (University of Konstanz) Numerical approximation for OCP 16 / 33

Model Predictive Control MPC trajectories (in L. Grüne, J. Pannek, NMPC ) black=prediction (obtained with an open loop optimization) red= MPC closed loop, y(t n ) = y µ N (t n ) G. Fabrini (University of Konstanz) Numerical approximation for OCP 17 / 33

MPC METHOD Model Predictive Control solves infinite time horizon problem by means of iterative solution of finite horizon (N 2) optimal control problems. min J N (u) s.t. u U N FEEDBACK CONTROL: µ N (y(t)) = u (t) We obtain a closed loop representation by applying the map µ N ẏ = F ( ) y(t), µ N (y(t)) OPTIMAL... TRAJECTORIES: y (t 0 ),..., y (t N 0 ) CONTROLS: u (t 0 ),..., u (t N 0 ) G. Fabrini (University of Konstanz) Numerical approximation for OCP 18 / 33

Model Predictive Control Advantages and disadvantages HJB PRO 1. Valid for all problems in any dimension, a-priori error estimates in L. 2. SL schemes can work on structured and unstructured grids. 3. The computation of feedbacks is almost built in. HJB CONS 1. Curse of dimensionality 2. Computational cost and huge memory allocations. MPC PRO 1. Easy to implement,short computational time. 2. It can be applied to high dimensional problems. 3. Feedback controls. MPC CONS 1. Approximate feedback just along one trajectory. 3. With a short horizon sub-optimal trajectory. G. Fabrini (University of Konstanz) Numerical approximation for OCP 19 / 33

IDEA Model Predictive Control Try to combine the advantages of the two methods in order to obtain an efficient algorithm. The approximation of the HJB equation needs to restrict the computational domain Ω to a subset of R n. The choice of the domain is totally arbitrary. GOAL: To find a reasonable way to compute Ω QUESTION: How to compute the computational domain? SOLUTION: An inexact MPC solver may provide a reference trajectory for our optimal control problem. G. Fabrini (University of Konstanz) Numerical approximation for OCP 20 / 33

Model Predictive Control Localized DP Algorithm Algorithm Start: Inizialization Step 1: Solve MPC and compute yy MPC for a given initial condition y 0 Step 2: Compute the distance from yy MPC 0 via the Eikonal equation Step 3: Select the tube Ω ρ with distance ρ with respect to yy MPC 0 Step 4: Compute the constrained value function v tube in Ω ρ via HJB Step 5: Compute the optimal feedbacks and trajectory using v tube. End A posteriori error-estimate for the choice of ρ y (t) y MPC (t) C eλt σ ζ y MPC (t) where ζ is a perturbation which is independent of y, C is a computable constant. G. Fabrini (University of Konstanz) Numerical approximation for OCP 21 / 33 t

Numerical Tests Outline 1 Introduction and motivations 2 Hamilton-Jacobi-Bellman 3 Model Predictive Control 4 Numerical Tests G. Fabrini (University of Konstanz) Numerical approximation for OCP 22 / 33

Numerical Tests Heat equation y t (t, x) ν y(t, x) = 0, a.e. in Q, ν y n (t, s, ν) = m i=1 u i(t)b i (s), a.e. on Σ y(0, x, ν) = y (x), a.e. in Ω where Q = (0, ) Ω, Σ = (0, ) Γ, Γ = Ω, b i : Γ R, given shape functions, ν R fixed parameter. (1) Cost Functional { 1 min J(u) = min u U u U 2 0 e λt y(t) ȳ 2 L 2 (Ω) dt + 1 2 } m σ i u i 2 L 2 (0, ) i=1 where U := {u L 2 (0, + ; R m ) u a u(t) u b t)} Goal: find a control which steers the trajectories to a desired state. G. Fabrini (University of Konstanz) Numerical approximation for OCP 23 / 33

Numerical Tests Finite Element Formulation Considering the weak formulation and using a FE scheme of piecewise linear basis functions ϕ i, i = 1,..., N, we lead to: where My (t) = νay(t) + Bu(t) for t (0, T ], y(0) = y. (2) A = (( a(ϕ j, ϕ i ) )) 1 i,j N, B = (( b j, ϕ i L 2 (Γ) y(t) = ( yi N (t) ) 1 i N, y = ( y i N ) 1 i N. )) 1 i N,1 j M, Proper Ortoghonal Decomposition (Kunisch, Hinze, Volkwein,...) We have to solve an optimal control problem for a large system of ODEs POD decomposition allows to reduce the number of variables to approximate partial differential equations. G. Fabrini (University of Konstanz) Numerical approximation for OCP 24 / 33

Numerical Tests Proper Orthogonal Decomposition and SVD Given snapshots Y = [y 1,..., y n ] R N n with y i = y(t; ). We look for an orthonormal basis {ψ i } l i=1 in Rm with l min{n, m} s.t. m l 2 d J(ψ 1,..., ψ l ) = α j y j y j, ψ i ψ i = reaches a minimum where {α j } n j=1 R+. j=1 i=1 L 2 (Ω) i=l+1 σ 2 i min J(ψ 1,..., ψ l ) s.t. ψ i, ψ j = δ ij Singular Value Decomposition: Y = ΨΣV T. For l {1,..., d = rank(y )}, {ψ i } l i=1 are called POD basis of rank l. Ansatz: l y(x, t) yi l (t)ψ i(x). i=1 G. Fabrini (University of Konstanz) Numerical approximation for OCP 25 / 33

Numerical Tests Reduced dynamics The system can be expressed as: ẏ(t) = F ( y(t), u(t) ) for t > 0, y(0) = y 0, with F(y) = M 1 (Ay + Bu + f ) R l, for y R l, u R m. Reduced Cost Functional with: L(y, u) = 1 2 J l (u) = 0 e λt L ( y(t), u(t) ) dt ( ) (y ȳ) M(y ȳ) + σu u for (y, u) R l R m. G. Fabrini (University of Konstanz) Numerical approximation for OCP 26 / 33

Numerical Tests Test 1 Dynamics and Cost Functional y t (t, x) ν y(t, x) = 0, a.e. in Q, ν y n (t, s, ν) = m i=1 u i(t)b i (s), a.e. on Σ y(0, x, ν) = y (x), a.e. in Ω J(u) = 1 2 Parameters: 0 e λt y(t) ȳ 2 L 2 (Ω) dt + 1 2 m σ i u i 2 L 2 (0, ) Snapshots: Q = [0, 1] Ω, Ω = [0, 1] 2,ν = 1, µ = 0, m = 1, y (x) = 0χ [0,0.5] + 0.5χ [0.5,1] x = 0.5 and t = 0.02, u(t) 1, l = 2 (number of POD basis). MPC: t MPC = 0.02, N = 5, λ = 1, u(t) [ 4, 4], ȳ 1. HJB: t HJB = 0.05, t traj = 0.02, # contr value= 21, # contr traj= 81 i=1 G. Fabrini (University of Konstanz) Numerical approximation for OCP 27 / 33

Numerical Tests G. Fabrini (University of Konstanz) Numerical approximation for OCP 28 / 33

Numerical Tests Figure: optimal trajectories for different intial profiles. λ = 1 MPC N=5 HJB in Ω ρ HJB in Ω CPU time [s] 37 72 198 J(u) 0.05 0.05 0.05 y(t ) ȳ L 2 0.1 0.02 0.01 G. Fabrini (University of Konstanz) Numerical approximation for OCP 29 / 33

Numerical Tests Test 2 Parameters: same as Test 1, different initial profile: y (x) = sin(πx). G. Fabrini (University of Konstanz) Numerical approximation for OCP 30 / 33

Numerical Tests Figure: optimal trajectories for different intial profiles. λ = 1 MPC N=5 HJB in Ω ρ HJB in Ω CPU time [s] 72 88 205 J(u) 0.013 0.012 0.012 y(t ) ȳ L 2 0.02 0.003 0.003 G. Fabrini (University of Konstanz) Numerical approximation for OCP 31 / 33

Numerical Tests CONCLUSIONS Local version of dynamic programming approach for infinite horizon optimal control problems The coupling between MPC and DP methods can produce rather accurate results Computational speed-up FUTURE DIRECTIONS Advection term (using more POD basis) A posteriori-error estimate including the advection term in the equation. G. Fabrini (University of Konstanz) Numerical approximation for OCP 32 / 33

Numerical Tests References A. Alla, G. Fabrini, M. Falcone, Coupling MPC and DP methods for an efficient solution of optimal control problems, Conference Proceedings of IFIP 2015. M. Bardi, I. Capuzzo Dolcetta,Optimal control and viscosity solutions of Hamilton-Jacobi-Bellman equations, Birkhäuser, 1997. G. Fabrini, M. Falcone, S. Volkwein, Coupling MPC and HJB for the computation of POD-basis feedback laws, to appear in ENUMATH proceedings, 2017. M. Falcone, R. Ferretti, Semi-Lagrangian Approximation schemes for linear and Hamilton-Jacobi equations, SIAM 2013. L. Grüne, J. Pannek, Nonlinear Model Predictive Control, Springer London, 2011. V. Gaitsgory, L. Grüne, N. Thatcher, Stabilization with discounted optimal control, 2014. S. Volkwein, F. Trölzsch POD a-posteriori error estimates for linear-quadratic optimal control problems, Computational Optimization and Applications, 44:83-115, 2009. S. Volkwein. Model Reduction using Proper Orthogonal Decomposition, Lecture Notes, University of Konstanz, 2013. THANK YOU FOR YOUR ATTENTION G. Fabrini (University of Konstanz) Numerical approximation for OCP 33 / 33