Lecture 11 Optimal Control
|
|
- Angela Williams
- 6 years ago
- Views:
Transcription
1 Lecture 11 Optimal Control The Maximum Principle Revisited Examples Numerical methods/optimica Examples, Lab 3 Material Lecture slides, including material by J. Åkesson, Automatic Control LTH Glad & Ljung, part of Chapter 18 Giacomo Como Lecture 11, Optimal Control p. 1
2 Goal To be able to solve simple problems using the maximum principle formulate advanced problems for numerical solution Giacomo Como Lecture 11, Optimal Control p. 2
3 Outline The Maximum Principle Revisited Examples Numerical methods/optimica Example Double integrator Example Alfa Laval Plate Reactor Giacomo Como Lecture 11, Optimal Control p. 3
4 Problem Formulation (1) Standard form (1): Minimize tf ẋ(t)=f(x(t),u(t)) 0 Trajectory cost Final cost {}}{{}}{ L(x(t),u(t))dt+ φ(x(t f )) u(t) U, 0 t t f, t f given x(0)=x 0 x(t) R n,u(t) R m U control constraints Here we have a fixed end-timet f. Giacomo Como Lecture 11, Optimal Control p. 4
5 The Maximum Principle (18.2) Introduce the Hamiltonian H(x,u, λ)=l(x,u)+λ T (t)f(x,u). If problem (1) has a solution{u (t),x (t)}, then min u U H(x (t),u, λ(t))=h(x (t),u (t), λ(t)), 0 t t f, where λ(t) solves the adjoint equation dλ(t)/dt= H T x(x (t),u (t), λ(t)), with λ(t f )= φ T x(x (t f )) Giacomo Como Lecture 11, Optimal Control p. 5
6 Remarks The Maximum Principle gives necessary conditions A pair(u ( ),x ( )) is called extremal if the conditions of the Maximum Principle are satisfied. Many extremals can exist. The maximum principle gives all possible candidates. However, there might not exist a minimum! Example Minimizex(1) whenẋ(t)=u(t),x(0)=0andu(t) is free Giacomo Como Lecture 11, Optimal Control p. 6
7 Goddard s Rocket Problem revisited How to send a rocket as high up in the air as possible? d v u D h = m dt v m γu h m (v(0),h(0),m(0))=(0,0,m 0 ),,γ >0 u motor force,d=d(v,h) air resistance Constraints:0 u u max andm(t f )=m 1 (empty) Optimization criterion:max u h(t f ) Giacomo Como Lecture 11, Optimal Control p. 7
8 Problem Formulation (2) tf min t f 0 0 u:[0,t f ] U L(x(t),u(t))dt+ φ(x(t f )) ẋ(t)=f(x(t),u(t)), x(0)=x 0 ψ(x(t f ))=0 Note the differences compared to standard form: End constraints ψ(x(t f ))=0 t f free variable (i.e., not specified a priori) Giacomo Como Lecture 11, Optimal Control p. 8
9 The Maximum Principle General Case (18.4) Introduce the Hamiltonian H(x,u, λ,n 0 )=n 0 L(x,u)+λ T (t)f(x,u) If problem (2) has a solutionu (t),x (t), then there is a vector function λ(t), a numbern 0 0, and a vector µ R r such that [n 0 µ T ] =0and min u U H(x (t),u, λ(t),n 0 )=H(x (t),u (t), λ(t),n 0 ), 0 t t f, where λ(t)= H T x (x (t),u (t), λ(t),n 0 ) λ(t f )=n 0 φ T x (x (t f ))+Ψ T x (x (t f ))µ If the end timet f is free, thenh(x (t f ),u (t f ), λ(t f ),n 0 )=0. Giacomo Como Lecture 11, Optimal Control p. 9
10 Normal/abnormal cases Can scalen 0, µ, λ(t) by the same constant Can reduce to two cases n 0 =1(normal) n 0 =0(abnormal, sincel and φ don t matter) As we saw before (18.2): fixed timet f and no end constraints normal case Giacomo Como Lecture 11, Optimal Control p. 10
11 Hamilton function is constant H is constant along extremals(x,u ) Proof (in the case whenu (t) Int(U)): d dt H=H xẋ+h λ λ+h u u=h x f f T H T x +0=0 Giacomo Como Lecture 11, Optimal Control p. 11
12 Feedback or feed-forward? Example: J min =1is achieved for dx =u, x(0)=1 dt ( minimizej= x 2 +u 2) dt 0 u(t)= e t openloop (1) or u(t) = x(t) closed loop (2) (??)= stable system (??)= asympt. stable system Sensitivity for noise and disturbances differ!! Giacomo Como Lecture 11, Optimal Control p. 12
13 Reference generation using optimal control Note that the optimization problem makes no distinction between open loop controlu (t) and closed loop control u (t,x). Feedback is needed to take care of disturbances and model errors. Idea: Use the optimal open loop solutionu (t),x (t) as reference values to a linear regulator that keeps the system close to the wanted trajectory Efficient for large setpoint changes. x x Planned trajectoryx Giacomo Como Lecture 11, Optimal Control p. 13
14 Recall Linear Quadratic Control minimizex T (t f )Q N x(t f )+ where tf ẋ=ax+bu, y=cx 0 x T Q 11 Q 12 x u Q12 T Q 22 u Optimal solution ift f =,Q N =0, all matrices constant, and x measurable: u= Lx wherel=q 1 22 (Q 12+B T S) ands=s T >0solves SA+A T S+Q 11 (Q 12 +SB)Q 1 22 (Q 12+B T S)=0 Giacomo Como Lecture 11, Optimal Control p. 14
15 Second Variations ApproximatingJ(x,u) around(x,u ) to second order δ 2 J= 1 2 δxt φ xx δx+ 1 2 δẋ=f x δx+f u δu tf t 0 δx T H xx H xu δx dt δu H ux H uu δu wherej=j + δ 2 J+... is a Taylor expansion of the criterion and δ x =x x and δ u =u u. Treat this as a new optimization problem. Linear time-varying system and quadratic criterion. Gives optimal controller u u =L(t)(x x ) Giacomo Como Lecture 11, Optimal Control p. 15
16 Opt. Ref. Gen. x u Lin. Cont. + u Proc. y ˆx Obs. Take care of deviations with linear controller Giacomo Como Lecture 11, Optimal Control p. 16
17 Outline The Maximum Principle Revisited Examples Numerical methods/optimica Example Double integrator Example Alfa Laval Plate Reactor Giacomo Como Lecture 11, Optimal Control p. 17
18 Example: Optimal heating Minimize tf =1 0 P(t) dt when Ṫ=P T 0 P P max T(0)=0, T(1)=1 T temperature P heat effect Giacomo Como Lecture 11, Optimal Control p. 18
19 Solution Hamiltonian H=n 0 P+λP λt Adjoint equation λ T = H T = H T = λ λ(1)=µ At optimality λ(t)=µe t 1 H=(n 0 + µe t 1 ) P λt }{{} σ(t) P (t)= { 0, σ(t)>0 P max, σ(t)<0 Giacomo Como Lecture 11, Optimal Control p. 19
20 Solution µ=0 (n 0, µ)=(0,0) Not allowed! µ =0 ConstantPor just one switch! T(t) approaches one from below, sop =0neart=1. Hence { 0, P 0 t t1 (t)= P max, t 1 <t 1 { 0, 0 t t1 T(t)= 1 t 1 e (t τ) P max dτ= ( e (t 1) e (t t 1) ) P max, t 1 <t 1 Timet 1 is given byt(1)= ( 1 e (1 t 1) ) P max =1 Has solution0 t 1 1ifP max 1 1 e 1 Giacomo Como Lecture 11, Optimal Control p. 20
21 Example The Milk Race Move milk in minimum time without spilling! [M. Grundelius Methods for Control of Liquid Slosh] [movie] Giacomo Como Lecture 11, Optimal Control p. 21
22 Minimal Time Problem NOTE! Common trick to rewrite criterion into standard form"!! minimize t f =minimize tf 0 1dt Control constraints No spilling u(t) u max i Cx(t) h Optimal controller has been found for the milk race Minimal time problem for linear systemẋ=ax+bu,y=cx with control constraints u i (t) u max i. Often bang-bang control as solution Giacomo Como Lecture 11, Optimal Control p. 22
23 Results- milk race Maximum slosh φ max =0.63 Maximum acceleration = 10 m/s 2 Time optimal acceleration profile 15 Acceleration Slosh Optimal time = 375 ms, industrial = 540ms Giacomo Como Lecture 11, Optimal Control p. 23
24 Outline The Maximum Principle Revisited Examples Numerical methods/optimica Example Double integrator Example Alfa Laval Plate Reactor Giacomo Como Lecture 11, Optimal Control p. 24
25 Numerical Methods for Dynamic Optimization Many algorithms Applicability highly model-dependent (ODE, DAE, PDE, hybrid?) Calculus of variations Single/Multiple Shooting Simultaneous methods Simulation-based methods Analogy with different simulation algorithms (but larger diversity) Heavy programming burden to use numerical algorithms Fortran C Engineering need for high-level descriptions Giacomo Como Lecture 11, Optimal Control p. 25
26 Modelica A Modeling Language Modelica is increasingly used in industry Expert knowledge Capital investments Usage so far Simulation (mainly) Other usages emerge Sensitivity analysis Optimization Model reduction System identification Control design Giacomo Como Lecture 11, Optimal Control p. 26
27 Optimica and JModelica A Research Project Shift focus: from encoding to problem formulation Enable dynamic optimization of Modelica models State of the art numerical algorithms Develop a high level description for optimization problems Extension of the Modelica language Develop prototype tools JModelica and The Optimica Compiler Code generation Giacomo Como Lecture 11, Optimal Control p. 27
28 Outline The Maximum Principle Revisited Examples Numerical methods/optimica Example Double integrator Example Alfa Laval Plate Reactor Giacomo Como Lecture 11, Optimal Control p. 28
29 Optimica An Example tf min 1dt u(t) 0 subject to the dynamic constraint ẋ(t)=v(t), x(0)=0 v(t)=u(t), v(0)=0 and x(t f )=1 v(t f )=0 v(t) u(t) 1 Giacomo Como Lecture 11, Optimal Control p. 29
30 A Modelica Model for a Double Integrator A double integrator model model DoubleIntegrator Real x(start=0); Real v(start=0); input Real u; equation der(x)=v; der(v)=u; end DoubleIntegrator; Giacomo Como Lecture 11, Optimal Control p. 30
31 The Optimica Description Minimum time optimization problem optimization DIMinTime (objective=cost(finaltime), starttime=0, finaltime(free=true,initialguess=1)) Real cost; DoubleIntegrator di(u(free=true,initialguess=0.0)); equation der(cost) = 1; constraint finaltime>=0.5; finaltime<=10; di.x(finaltime)=1; di.v(finaltime)=0; di.v<=0.5; di.u>= 1; di.u<=1; end DIMinTime; Giacomo Como Lecture 11, Optimal Control p. 31
32 Optimal Double Integrator Profiles u x v t [s] t [s] Giacomo Como Lecture 11, Optimal Control p. 32
33 Outline The Maximum Principle Revisited Examples Numerical methods/optimica Example Double integrator Example Alfa Laval Plate Reactor Giacomo Como Lecture 11, Optimal Control p. 33
34 Optimal Start-up of a Plate Reactor Reactant B Reactant A HEX T f FC FC q B1 q B2 T T T T T T T T T T Reactor outlet T c HEX Cooling water Achieve safe start-upt r T max Maximize the conversion Minimize the start-up time Giacomo Como Lecture 11, Optimal Control p. 34
35 The Optimization Problem Reduce sensitivity of the nominal start-up trajectory by: - Introducing a constraint on the accumulated concentration of reactant B - Introducing high frequency penalties on the control inputs min u tf 0 α A c 2 A,out + α Bc 2 B,out + α B1q 2 B1,f + α B2q 2 B2,f + α T1 Ṫf 2 + α T 2 Ṫc 2 dt subjectto ẋ=f(x,u) T r,i 155, i=1..n c B,1 600, c B, q B1 0.7, 0 q B Ṫf 2, 1.5 Ṫc T f 80, 20 T c 80 Giacomo Como Lecture 11, Optimal Control p. 35
36 The Optimization Problem Optimica Robust optimization formulation optimization PlateReactorOptimization (objective=cost(finaltime), starttime=0, finaltime=150) PlateReactor pr(u_t_cool_setpoint(free=true), u_tfeeda_setpoint(free=true), u_b1_setpoint(free=true), u_b2_setpoint(free=true)); parameter Real sc_u = 670/50 "Scaling factor"; parameter Real sc_c = 2392/50 "Scaling factor"; Real cost(start=0); equation der(cost) = 0.1*pr.cA[30]^2*sc_c^ *pr.cB[30]^2*sc_c^2 + 1*pr.u_B1_setpoint_f^2 + 1*pr.u_B2_setpoint_f^2 + 1*der(pr.u_T_cool_setpoint)^2*sc_u^2 + 1*der(pr.u_TfeedA_setpoint)^2*sc_u^2; constraint pr.tr/u_sc<=( )*ones(30); pr.cb[1]<=200/sc_c; pr.cb[16]<=400/sc_c; pr.u_b1_setpoint>=0; pr.u_b1_setpoint<=0.7; pr.u_b2_setpoint>=0; pr.u_b2_setpoint<=0.7; pr.u_t_cool_setpoint>=(15+273)/sc_u; pr.u_t_cool_setpoint<=(80+273)/sc_u; pr.u_tfeeda_setpoint>=(30+273)/sc_u; pr.u_tfeeda_setpoint<=(80+273)/sc_u; der(pr.u_t_cool_setpoint)>= 1.5/sc_u; der(pr.u_t_cool_setpoint)<=0.7/sc_u; der(pr.u_tfeeda_setpoint)>= 1.5/sc_u; der(pr.u_tfeeda_setpoint)<=2/sc_u; end PlateReactorOptimization; Giacomo Como Lecture 11, Optimal Control p. 36
37 1500 c B,1 [mol/m 3 ] c B,2 [mol/m 3 ] T 1 [ C] T 2 [ C] q B1 [-] q B2 [-] T f [ C] T c [ C] Time [s] Time [s] Time [s] Almost as fast, but more robust with lowerc B -constraints Time [s] Giacomo Como Lecture 11, Optimal Control p. 37
38 Summary The Maximum Principle Revisited Examples Numerical methods/optimica Example Double integrator Example Alfa Laval Plate Reactor Giacomo Como Lecture 11, Optimal Control p. 38
Deterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationLecture 9 Nonlinear Control Design
Lecture 9 Nonlinear Control Design Exact-linearization Lyapunov-based design Lab 2 Adaptive control Sliding modes control Literature: [Khalil, ch.s 13, 14.1,14.2] and [Glad-Ljung,ch.17] Course Outline
More informationCHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67
CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Chapter2 p. 1/67 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Main Purpose: Introduce the maximum principle as a necessary condition to be satisfied by any optimal
More informationStochastic and Adaptive Optimal Control
Stochastic and Adaptive Optimal Control Robert Stengel Optimal Control and Estimation, MAE 546 Princeton University, 2018! Nonlinear systems with random inputs and perfect measurements! Stochastic neighboring-optimal
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationProblem 1 Cost of an Infinite Horizon LQR
THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 5243 INTRODUCTION TO CYBER-PHYSICAL SYSTEMS H O M E W O R K # 5 Ahmad F. Taha October 12, 215 Homework Instructions: 1. Type your solutions in the LATEX homework
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationPontryagin s Minimum Principle 1
ECE 680 Fall 2013 Pontryagin s Minimum Principle 1 In this handout, we provide a derivation of the minimum principle of Pontryagin, which is a generalization of the Euler-Lagrange equations that also includes
More informationTheory and Applications of Constrained Optimal Control Proble
Theory and Applications of Constrained Optimal Control Problems with Delays PART 1 : Mixed Control State Constraints Helmut Maurer 1, Laurenz Göllmann 2 1 Institut für Numerische und Angewandte Mathematik,
More informationTheoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems.
ES22 Lecture Notes #11 Theoretical Justification for LQ problems: Sufficiency condition: LQ problem is the second order expansion of nonlinear optimal control problems. J = φ(x( ) + L(x,u,t)dt ; x= f(x,u,t)
More informationControl Systems I. Lecture 2: Modeling. Suggested Readings: Åström & Murray Ch. 2-3, Guzzella Ch Emilio Frazzoli
Control Systems I Lecture 2: Modeling Suggested Readings: Åström & Murray Ch. 2-3, Guzzella Ch. 2-3 Emilio Frazzoli Institute for Dynamic Systems and Control D-MAVT ETH Zürich September 29, 2017 E. Frazzoli
More informationOptimal Control. Lecture 3. Optimal Control of Discrete Time Dynamical Systems. John T. Wen. January 22, 2004
Optimal Control Lecture 3 Optimal Control of Discrete Time Dynamical Systems John T. Wen January, 004 Outline optimization of a general multi-stage discrete time dynamical systems special case: discrete
More informationControl Design of a Distributed Parameter. Fixed-Bed Reactor
Workshop on Control of Distributed Parameter Systems p. 1/1 Control Design of a Distributed Parameter Fixed-Bed Reactor Ilyasse Aksikas and J. Fraser Forbes Department of Chemical and Material Engineering
More informationTutorial on Control and State Constrained Optimal Control Problems
Tutorial on Control and State Constrained Optimal Control Problems To cite this version:. blems. SADCO Summer School 211 - Optimal Control, Sep 211, London, United Kingdom. HAL Id: inria-629518
More informationOptimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:
Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic
More informationOPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28
OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from
More informationAM 205: lecture 14. Last time: Boundary value problems Today: Numerical solution of PDEs
AM 205: lecture 14 Last time: Boundary value problems Today: Numerical solution of PDEs ODE BVPs A more general approach is to formulate a coupled system of equations for the BVP based on a finite difference
More informationOptimizing Economic Performance using Model Predictive Control
Optimizing Economic Performance using Model Predictive Control James B. Rawlings Department of Chemical and Biological Engineering Second Workshop on Computational Issues in Nonlinear Control Monterey,
More informationLecture 9 Nonlinear Control Design. Course Outline. Exact linearization: example [one-link robot] Exact Feedback Linearization
Lecture 9 Nonlinear Control Design Course Outline Eact-linearization Lyapunov-based design Lab Adaptive control Sliding modes control Literature: [Khalil, ch.s 13, 14.1,14.] and [Glad-Ljung,ch.17] Lecture
More informationProject Optimal Control: Optimal Control with State Variable Inequality Constraints (SVIC)
Project Optimal Control: Optimal Control with State Variable Inequality Constraints (SVIC) Mahdi Ghazaei, Meike Stemmann Automatic Control LTH Lund University Problem Formulation Find the maximum of the
More informationOptimization via the Hamilton-Jacobi-Bellman Method: Theory and Applications
Optimization via the Hamilton-Jacobi-Bellman Method: Theory and Applications Navin Khaneja lectre notes taken by Christiane Koch Jne 24, 29 1 Variation yields a classical Hamiltonian system Sppose that
More informationDynamic Consistency for Stochastic Optimal Control Problems
Dynamic Consistency for Stochastic Optimal Control Problems Cadarache Summer School CEA/EDF/INRIA 2012 Pierre Carpentier Jean-Philippe Chancelier Michel De Lara SOWG June 2012 Lecture outline Introduction
More informationSolution of Stochastic Optimal Control Problems and Financial Applications
Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty
More informationOrdinary differential equations. Phys 420/580 Lecture 8
Ordinary differential equations Phys 420/580 Lecture 8 Most physical laws are expressed as differential equations These come in three flavours: initial-value problems boundary-value problems eigenvalue
More informationDirect and indirect methods for optimal control problems and applications in engineering
Direct and indirect methods for optimal control problems and applications in engineering Matthias Gerdts Computational Optimisation Group School of Mathematics The University of Birmingham gerdtsm@maths.bham.ac.uk
More informationMath 211. Lecture #6. Linear Equations. September 9, 2002
1 Math 211 Lecture #6 Linear Equations September 9, 2002 2 Air Resistance 2 Air Resistance Acts in the direction opposite to the velocity. 2 Air Resistance Acts in the direction opposite to the velocity.
More information7 OPTIMAL CONTROL 7.1 EXERCISE 1. Solve the following optimal control problem. max. (x u 2 )dt x = u x(0) = Solution with the first variation
7 OPTIMAL CONTROL 7. EXERCISE Solve the following optimal control problem max 7.. Solution with the first variation The Lagrangian is L(x, u, λ, μ) = (x u )dt x = u x() =. [(x u ) λ(x u)]dt μ(x() ). Now
More informationControl Systems I. Lecture 2: Modeling and Linearization. Suggested Readings: Åström & Murray Ch Jacopo Tani
Control Systems I Lecture 2: Modeling and Linearization Suggested Readings: Åström & Murray Ch. 2-3 Jacopo Tani Institute for Dynamic Systems and Control D-MAVT ETH Zürich September 28, 2018 J. Tani, E.
More informationFormula Sheet for Optimal Control
Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming
More informationLecture 1: Pragmatic Introduction to Stochastic Differential Equations
Lecture 1: Pragmatic Introduction to Stochastic Differential Equations Simo Särkkä Aalto University, Finland (visiting at Oxford University, UK) November 13, 2013 Simo Särkkä (Aalto) Lecture 1: Pragmatic
More informationOptimization of a Pendulum System using Optimica and Modelica
Optimization of a Pendulum System using Optimica and Modelica Pontus Giselsson a Johan Åkesson a,b Anders Robertsson a a) Dept. of Automatic Control, Lund University, Sweden b) Modelon AB, Sweden Abstract
More informationGame Theory Extra Lecture 1 (BoB)
Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder,
More informationNumerical Methods for Constrained Optimal Control Problems
Numerical Methods for Constrained Optimal Control Problems Hartono Hartono A Thesis submitted for the degree of Doctor of Philosophy School of Mathematics and Statistics University of Western Australia
More informationModeling and Analysis of Dynamic Systems
Modeling and Analysis of Dynamic Systems Dr. Guillaume Ducard Fall 2017 Institute for Dynamic Systems and Control ETH Zurich, Switzerland G. Ducard c 1 / 57 Outline 1 Lecture 13: Linear System - Stability
More informationLQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin
LQR, Kalman Filter, and LQG Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin May 2015 Linear Quadratic Regulator (LQR) Consider a linear system
More informationLecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004
MER42 Advanced Control Lecture 9 Introduction to Kalman Filtering Linear Quadratic Gaussian Control (LQG) G. Hovland 24 Announcement No tutorials on hursday mornings 8-9am I will be present in all practical
More informationMATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem
MATH4406 (Control Theory) Unit 6: The Linear Quadratic Regulator (LQR) and Model Predictive Control (MPC) Prepared by Yoni Nazarathy, Artem Pulemotov, September 12, 2012 Unit Outline Goal 1: Outline linear
More informationHJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011
Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance
More informationIntroduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley
Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley Outline Introduction to optimal control Reachability as an optimal control problem Various shades of reachability Goal of This
More information1. The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ = Ax + Bu is
ECE 55, Fall 2007 Problem Set #4 Solution The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ Ax + Bu is x(t) e A(t ) x( ) + e A(t τ) Bu(τ)dτ () This formula is extremely important
More informationLecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations
Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations Simo Särkkä Aalto University Tampere University of Technology Lappeenranta University of Technology Finland November
More informationPoisson Jumps in Credit Risk Modeling: a Partial Integro-differential Equation Formulation
Poisson Jumps in Credit Risk Modeling: a Partial Integro-differential Equation Formulation Jingyi Zhu Department of Mathematics University of Utah zhu@math.utah.edu Collaborator: Marco Avellaneda (Courant
More informationControl Systems Design
ELEC4410 Control Systems Design Lecture 18: State Feedback Tracking and State Estimation Julio H. Braslavsky julio@ee.newcastle.edu.au School of Electrical Engineering and Computer Science Lecture 18:
More informationReflected Brownian Motion
Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide
More informationWeak Convergence of Numerical Methods for Dynamical Systems and Optimal Control, and a relation with Large Deviations for Stochastic Equations
Weak Convergence of Numerical Methods for Dynamical Systems and, and a relation with Large Deviations for Stochastic Equations Mattias Sandberg KTH CSC 2010-10-21 Outline The error representation for weak
More informationAutomatic Control II Computer exercise 3. LQG Design
Uppsala University Information Technology Systems and Control HN,FS,KN 2000-10 Last revised by HR August 16, 2017 Automatic Control II Computer exercise 3 LQG Design Preparations: Read Chapters 5 and 9
More informationSubject: Optimal Control Assignment-1 (Related to Lecture notes 1-10)
Subject: Optimal Control Assignment- (Related to Lecture notes -). Design a oil mug, shown in fig., to hold as much oil possible. The height and radius of the mug should not be more than 6cm. The mug must
More information(q 1)t. Control theory lends itself well such unification, as the structure and behavior of discrete control
My general research area is the study of differential and difference equations. Currently I am working in an emerging field in dynamical systems. I would describe my work as a cross between the theoretical
More informationI. D. Landau, A. Karimi: A Course on Adaptive Control Adaptive Control. Part 9: Adaptive Control with Multiple Models and Switching
I. D. Landau, A. Karimi: A Course on Adaptive Control - 5 1 Adaptive Control Part 9: Adaptive Control with Multiple Models and Switching I. D. Landau, A. Karimi: A Course on Adaptive Control - 5 2 Outline
More informationTwo-Point Boundary Value Problem and Optimal Feedback Control based on Differential Algebra
Two-Point Boundary Value Problem and Optimal Feedback Control based on Differential Algebra Politecnico di Milano Department of Aerospace Engineering Milan, Italy Taylor Methods and Computer Assisted Proofs
More informationModule 03 Linear Systems Theory: Necessary Background
Module 03 Linear Systems Theory: Necessary Background Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html September
More informationModule 05 Introduction to Optimal Control
Module 05 Introduction to Optimal Control Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html October 8, 2015
More informationTime-Invariant Linear Quadratic Regulators!
Time-Invariant Linear Quadratic Regulators Robert Stengel Optimal Control and Estimation MAE 546 Princeton University, 17 Asymptotic approach from time-varying to constant gains Elimination of cross weighting
More informationGame Theoretic Continuous Time Differential Dynamic Programming
Game heoretic Continuous ime Differential Dynamic Programming Wei Sun, Evangelos A. heodorou and Panagiotis siotras 3 Abstract In this work, we derive a Game heoretic Differential Dynamic Programming G-DDP
More informationStatic and Dynamic Optimization (42111)
Static and Dynamic Optimization (421) Niels Kjølstad Poulsen Build. 0b, room 01 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark Email:
More informationLinear Parameter Varying and Time-Varying Model Predictive Control
Linear Parameter Varying and Time-Varying Model Predictive Control Alberto Bemporad - Model Predictive Control course - Academic year 016/17 0-1 Linear Parameter-Varying (LPV) MPC LTI prediction model
More informationSynthesis via State Space Methods
Chapter 18 Synthesis via State Space Methods Here, we will give a state space interpretation to many of the results described earlier. In a sense, this will duplicate the earlier work. Our reason for doing
More informationChapter 3 Convolution Representation
Chapter 3 Convolution Representation DT Unit-Impulse Response Consider the DT SISO system: xn [ ] System yn [ ] xn [ ] = δ[ n] If the input signal is and the system has no energy at n = 0, the output yn
More informationQuasi-linear first order equations. Consider the nonlinear transport equation
Quasi-linear first order equations Consider the nonlinear transport equation u t + c(u)u x = 0, u(x, 0) = f (x) < x < Quasi-linear first order equations Consider the nonlinear transport equation u t +
More informationA posteriori analysis of a discontinuous Galerkin scheme for a diffuse interface model
A posteriori analysis of a discontinuous Galerkin scheme for a diffuse interface model Jan Giesselmann joint work with Ch. Makridakis (Univ. of Sussex), T. Pryer (Univ. of Reading) 9th DFG-CNRS WORKSHOP
More informationReview: control, feedback, etc. Today s topic: state-space models of systems; linearization
Plan of the Lecture Review: control, feedback, etc Today s topic: state-space models of systems; linearization Goal: a general framework that encompasses all examples of interest Once we have mastered
More informationNumerical Optimal Control Part 3: Function space methods
Numerical Optimal Control Part 3: Function space methods SADCO Summer School and Workshop on Optimal and Model Predictive Control OMPC 2013, Bayreuth Institute of Mathematics and Applied Computing Department
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationOptimal Control. McGill COMP 765 Oct 3 rd, 2017
Optimal Control McGill COMP 765 Oct 3 rd, 2017 Classical Control Quiz Question 1: Can a PID controller be used to balance an inverted pendulum: A) That starts upright? B) That must be swung-up (perhaps
More informationEuler s Method applied to the control of switched systems
Euler s Method applied to the control of switched systems FORMATS 2017 - Berlin Laurent Fribourg 1 September 6, 2017 1 LSV - CNRS & ENS Cachan L. Fribourg Euler s method and switched systems September
More information(Refer Slide Time: 00:32)
Nonlinear Dynamical Systems Prof. Madhu. N. Belur and Prof. Harish. K. Pillai Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 12 Scilab simulation of Lotka Volterra
More informationPontryagin s maximum principle
Pontryagin s maximum principle Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2012 Emo Todorov (UW) AMATH/CSE 579, Winter 2012 Lecture 5 1 / 9 Pontryagin
More informationIterative Learning Control (ILC)
Department of Automatic Control LTH, Lund University ILC ILC - the main idea Time Domain ILC approaches Stability Analysis Example: The Milk Race Frequency Domain ILC Example: Marine Vibrator Material:
More informationChapter Introduction. A. Bensoussan
Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688
More informationStochastic Differential Dynamic Programming
Stochastic Differential Dynamic Programming Evangelos heodorou, Yuval assa & Emo odorov Abstract Although there has been a significant amount of work in the area of stochastic optimal control theory towards
More informationMath 2a Prac Lectures on Differential Equations
Math 2a Prac Lectures on Differential Equations Prof. Dinakar Ramakrishnan 272 Sloan, 253-37 Caltech Office Hours: Fridays 4 5 PM Based on notes taken in class by Stephanie Laga, with a few added comments
More informationControl. CSC752: Autonomous Robotic Systems. Ubbo Visser. March 9, Department of Computer Science University of Miami
Control CSC752: Autonomous Robotic Systems Ubbo Visser Department of Computer Science University of Miami March 9, 2017 Outline 1 Control system 2 Controller Images from http://en.wikipedia.org/wiki/feed-forward
More informationMATH4406 (Control Theory) Unit 1: Introduction Prepared by Yoni Nazarathy, July 21, 2012
MATH4406 (Control Theory) Unit 1: Introduction Prepared by Yoni Nazarathy, July 21, 2012 Unit Outline Introduction to the course: Course goals, assessment, etc... What is Control Theory A bit of jargon,
More informationFinite Element Clifford Algebra: A New Toolkit for Evolution Problems
Finite Element Clifford Algebra: A New Toolkit for Evolution Problems Andrew Gillette joint work with Michael Holst Department of Mathematics University of California, San Diego http://ccom.ucsd.edu/ agillette/
More informationA review of stability and dynamical behaviors of differential equations:
A review of stability and dynamical behaviors of differential equations: scalar ODE: u t = f(u), system of ODEs: u t = f(u, v), v t = g(u, v), reaction-diffusion equation: u t = D u + f(u), x Ω, with boundary
More informationFAULT-TOLERANT CONTROL OF CHEMICAL PROCESS SYSTEMS USING COMMUNICATION NETWORKS. Nael H. El-Farra, Adiwinata Gani & Panagiotis D.
FAULT-TOLERANT CONTROL OF CHEMICAL PROCESS SYSTEMS USING COMMUNICATION NETWORKS Nael H. El-Farra, Adiwinata Gani & Panagiotis D. Christofides Department of Chemical Engineering University of California,
More informationMATH 220: Problem Set 3 Solutions
MATH 220: Problem Set 3 Solutions Problem 1. Let ψ C() be given by: 0, x < 1, 1 + x, 1 < x < 0, ψ(x) = 1 x, 0 < x < 1, 0, x > 1, so that it verifies ψ 0, ψ(x) = 0 if x 1 and ψ(x)dx = 1. Consider (ψ j )
More informationRecent Advances in State Constrained Optimal Control
Recent Advances in State Constrained Optimal Control Richard B. Vinter Imperial College London Control for Energy and Sustainability Seminar Cambridge University, 6th November 2009 Outline of the talk
More informationPrinciples of Optimal Control Spring 2008
MIT OpenCourseWare http://ocw.it.edu 16.323 Principles of Optial Control Spring 2008 For inforation about citing these aterials or our Ters of Use, visit: http://ocw.it.edu/ters. 16.323 Lecture 10 Singular
More informationNew Tools for Analysing Power System Dynamics
1 New Tools for Analysing Power System Dynamics Ian A. Hiskens Department of Electrical and Computer Engineering University of Wisconsin - Madison (With support from many talented people.) PSerc Research
More informationLQR and H 2 control problems
LQR and H 2 control problems Domenico Prattichizzo DII, University of Siena, Italy MTNS - July 5-9, 2010 D. Prattichizzo (Siena, Italy) MTNS - July 5-9, 2010 1 / 23 Geometric Control Theory for Linear
More informationRobotics. Control Theory. Marc Toussaint U Stuttgart
Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,
More informationDepartment of Aerospace Engineering and Mechanics University of Minnesota Written Preliminary Examination: Control Systems Friday, April 9, 2010
Department of Aerospace Engineering and Mechanics University of Minnesota Written Preliminary Examination: Control Systems Friday, April 9, 2010 Problem 1: Control of Short Period Dynamics Consider the
More informationOPTIMAL CONTROL CHAPTER INTRODUCTION
CHAPTER 3 OPTIMAL CONTROL What is now proved was once only imagined. William Blake. 3.1 INTRODUCTION After more than three hundred years of evolution, optimal control theory has been formulated as an extension
More informationUCLA Chemical Engineering. Process & Control Systems Engineering Laboratory
Constrained Innite-time Optimal Control Donald J. Chmielewski Chemical Engineering Department University of California Los Angeles February 23, 2000 Stochastic Formulation - Min Max Formulation - UCLA
More informationPhysics 351 Wednesday, April 22, 2015
Physics 351 Wednesday, April 22, 2015 HW13 due Friday. The last one! You read Taylor s Chapter 16 this week (waves, stress, strain, fluids), most of which is Phys 230 review. Next weekend, you ll read
More informationInfinite-dimensional nonlinear predictive controller design for open-channel hydraulic systems
Infinite-dimensional nonlinear predictive controller design for open-channel hydraulic systems D. Georges, Control Systems Dept - Gipsa-lab, Grenoble INP Workshop on Irrigation Channels and Related Problems,
More informationAutomatic Control 2. Nonlinear systems. Prof. Alberto Bemporad. University of Trento. Academic year
Automatic Control 2 Nonlinear systems Prof. Alberto Bemporad University of Trento Academic year 2010-2011 Prof. Alberto Bemporad (University of Trento) Automatic Control 2 Academic year 2010-2011 1 / 18
More informationInverse Optimality Design for Biological Movement Systems
Inverse Optimality Design for Biological Movement Systems Weiwei Li Emanuel Todorov Dan Liu Nordson Asymtek Carlsbad CA 921 USA e-mail: wwli@ieee.org. University of Washington Seattle WA 98195 USA Google
More informationLecture 5 Input output stability
Lecture 5 Input output stability or How to make a circle out of the point 1+0i, and different ways to stay away from it... k 2 yf(y) r G(s) y k 1 y y 1 k 1 1 k 2 f( ) G(iω) Course Outline Lecture 1-3 Lecture
More informationLecture 6 Quantum Mechanical Systems and Measurements
Lecture 6 Quantum Mechanical Systems and Measurements Today s Program: 1. Simple Harmonic Oscillator (SHO). Principle of spectral decomposition. 3. Predicting the results of measurements, fourth postulate
More informationEE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games
EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games S. S. Sastry REVISED March 29th There exist two main approaches to optimal control and dynamic games: 1. via the Calculus of Variations
More informationSystems and Control Theory Lecture Notes. Laura Giarré
Systems and Control Theory Lecture Notes Laura Giarré L. Giarré 2017-2018 Lesson 5: State Space Systems State Dimension Infinite-Dimensional systems State-space model (nonlinear) LTI State Space model
More information4F3 - Predictive Control
4F3 Predictive Control - Lecture 2 p 1/23 4F3 - Predictive Control Lecture 2 - Unconstrained Predictive Control Jan Maciejowski jmm@engcamacuk 4F3 Predictive Control - Lecture 2 p 2/23 References Predictive
More informationPrashant Mhaskar, Nael H. El-Farra & Panagiotis D. Christofides. Department of Chemical Engineering University of California, Los Angeles
HYBRID PREDICTIVE OUTPUT FEEDBACK STABILIZATION OF CONSTRAINED LINEAR SYSTEMS Prashant Mhaskar, Nael H. El-Farra & Panagiotis D. Christofides Department of Chemical Engineering University of California,
More informationA Robust Controller for Scalar Autonomous Optimal Control Problems
A Robust Controller for Scalar Autonomous Optimal Control Problems S. H. Lam 1 Department of Mechanical and Aerospace Engineering Princeton University, Princeton, NJ 08544 lam@princeton.edu Abstract Is
More informationSpacecraft Trajectory Optimization. Edited. Bruce A. Conway. Published by Cambridge University Press 2010
Spacecraft Trajectory Optimization Edited by Bruce A. Conway Published by Cambridge University Press 2010 Cambridge University Press 2010 ISBN 978-0-521-51850-5 2 Primer Vector Theory and Applications
More informationTutorial on Control and State Constrained Optimal Control Pro. Control Problems and Applications Part 3 : Pure State Constraints
Tutorial on Control and State Constrained Optimal Control Problems and Applications Part 3 : Pure State Constraints University of Münster Institute of Computational and Applied Mathematics SADCO Summer
More information1.5 First Order PDEs and Method of Characteristics
1.5. FIRST ORDER PDES AND METHOD OF CHARACTERISTICS 35 1.5 First Order PDEs and Method of Characteristics We finish this introductory chapter by discussing the solutions of some first order PDEs, more
More information