Control of Dynamical System
|
|
- Meryl Stephens
- 5 years ago
- Views:
Transcription
1 Control of Dynamical System Yuan Gao Applied Mathematics University of Washington Spring / 21
2 Simplified Model for a Car To start, let s consider the simple example of controlling a car. A extremely simplified model follows directly from Newton s law of motion F = ma. We take F (t) = cu(t), where u(t) is the control (pressing the paddle/break). The dynamics of a car is then Possible goals: d 2 x dt 2 = c m u. Drive the car to a specific speed. Drive the car to a desired location. 2 / 21
3 Controlling the Speed The differential equation concerning speed is dv dt = c m u. Define e(t) = v d v(t) as the error at time t, where v d is the desired speed. The goal is to let e(t) 0. If v(0) = v d, what should u(0) be? If v(0) > v d, what sign should u(0) take? If v(0) < v d, what sign should u(0) take? 3 / 21
4 Bang-Bang Control u max if e(t) > 0, u(t) := u max if e(t) < 0, 0 if e(t) = 0. 4 / 21
5 Bang-Bang Control / 21
6 Difference Equation dv dt = c m u. In many situations, control happens at discrete time. or v(t + t) v(t) t = c m u(t), v(t + t) = v(t) + c m u(t) t. Drawbacks off Bang-Bang control: suffers from overshooting overacts to small errors bad for actuators 6 / 21
7 PE Control PE control, or proportional to error control, is we choose control that is proportional to the error: u(t) = ke(t) / 21
8 Analysis of PE Control In terms of PE control, Suppose v(0) = 0, the solution is dv dt = c m u(t) = c m k(v d v). v(t) = v d (1 e c m kt ). Does PE control suffer from overshooting? v(t + t) = v(t) + c m k(v d v(t)) t. The required condition for stability is c k t < 2. m 8 / 21
9 Adding Air Resistance A more realistic model... dv(t) = c u(t) γv(t). dt m / 21
10 Failure of PE Control Plug in u(t) = ke(t), dv(t) dt = c m k(v d v(t)) γv(t). The equilibrium velocity is ck ck+mγ v d. Thinking in this way, we have another way of controlling the system by setting its equilibrium the point we want! dv(t) dt = c m u(t) γv(t) = 0. = u(t) = m c γv d. Drawback: it requires a lot more information the mass m, coefficient c, and frictional constant γ not robust parameters changing across different situations 10 / 21
11 PID Control P: Proportional, I: Integral, D: Derivative. u(t) = k p e(t) + k i t 0 e(τ)dτ + k d d dt e(t) / 21
12 Controlling Position Goal: x x d, v 0. The system of equations is In matrix form: d dt dx dt = v. dv dt = c m u γv. [ ] [ x 0 1 = v 0 γ ] [ ] [ ] x 0 + c u. v m 12 / 21
13 Comparison of Two Controllers k(x x d ) k(x x d + v v d ) / 21
14 Optimal Control Theory x: state u(x): the control at state x l(x, u): cost of applying control u at state x The goal is to minimize the total cost: J(x, u) = where x k+1 = next(x k, u k ). n l(x k, u k ), k=0 14 / 21
15 Bellman s Principle of Optimality An optimal policy has the property that whatever the initial state and initial decision are, the remaining decisions must constitute an optimal policy with regard to the state resulting from the first decision. Introduce the optimal value function (optimal cost-to-go): v(x) = minimal total cost for completing the task starting from state x Consider all possible choices of u at current state x, add l(x, u) to v(next(x, u)), and choose the u for which the sum l(x, u) + v(next(x, u)) is minimal. 15 / 21
16 Bellman Equation Dynamic Programming Finite horizon formulation: v(x ) = 0. The Bellman Equation: v(x) = min{l(x, u) + v(next(x, u))}. u Infinite horizon discounted cost formulation: minimize The Bellman Equation: J(x, u) = lim n k=0 n α k l(x k, u k ). v(x) = min{l(x, u) + αv(next(x, u))}. u 16 / 21
17 Continuous Optimal Control Assume the dynamics The goal is to minimize dx dt = f (x, u). J(x, u) = T 0 l(x(t), u(t), t)dt. Choosing a time step of t, we have n J(x, u) = l(x k, u k, k t) t, k=0 and next(x k, u k ) = x k + f (x k, u k ) t. 17 / 21
18 Continuous Optimal Control Apply the bellman equation, v(x, k t) = min u {l(x, u, k t) t + v (x + f (x, u) t, (k + 1) t)}. Taylor expansion for v in the first variable x gives v (x + f (x, u) t, (k + 1) t) v(x, (k + 1) t) + v x (x, (k + 1) t)f (x, u) t. This gives us (v(x, k t) v(x, (k + 1) t)) / t = min u {l(x, u, k t) + f (x, u)v x (x, (k + 1) t)}. Now take the limit as t 0, v t (x, t) = min u {l(x, u, t) + f (x, u)v x (x, t)}. The Hamiltion-Jacobi-Bellman (HJB) equation! 18 / 21
19 Infinite Horizon Formulation The total cost becomes J(x, u) = 0 e αt l(x(t), u(t))dt. Choosing a time step of t, we have J(x, u) = e αk t l(x k, u k ) t. k=0 Apply the infinite horizon discounted cost Bellman Equation, } v(x) = min u {l(x, u) t + e α t v (x + f (x, u) t). Using similar techniques we get the infinite horizon HJB equation: { αv(x) = min u l(x, u) + f (x, u)v (x) }. 19 / 21
20 A Simple Example Back to the simplest example of controlling the speed of a car dx dt = c m u, here x represents the velocity, to be consistent with previous notations. We define the cost function as l(x, u) = 1 2 (x x d) u2. Assume the value function has the form v(x) = 1 2 V (x x d) 2. Plug into the infinite horizon HJB equation and solve for V, we get ( ) V = m2 c 2 α + α c2 m 2 and u = c m V (x x d). 20 / 21
21 Varying α α = 0.1 control velocity α = 1 control velocity / 21
Hamilton-Jacobi-Bellman Equation Feb 25, 2008
Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS
More informationOptimal Control. McGill COMP 765 Oct 3 rd, 2017
Optimal Control McGill COMP 765 Oct 3 rd, 2017 Classical Control Quiz Question 1: Can a PID controller be used to balance an inverted pendulum: A) That starts upright? B) That must be swung-up (perhaps
More informationControlled Diffusions and Hamilton-Jacobi Bellman Equations
Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationEN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018
EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal
More informationOptimal Control Theory
Optimal Control Theory The theory Optimal control theory is a mature mathematical discipline which provides algorithms to solve various control problems The elaborate mathematical machinery behind optimal
More informationInteractions. Yuan Gao. Spring Applied Mathematics University of Washington
Interactions Yuan Gao Applied Mathematics University of Washington yuangao@uw.edu Spring 2015 1 / 27 Nonlinear System Consider the following coupled ODEs: dx = f (x, y). dt dy = g(x, y). dt In general,
More informationRobust control and applications in economic theory
Robust control and applications in economic theory In honour of Professor Emeritus Grigoris Kalogeropoulos on the occasion of his retirement A. N. Yannacopoulos Department of Statistics AUEB 24 May 2013
More informationMotion Along a Straight Line
Chapter 2 Motion Along a Straight Line 2.1 Displacement, Time, and Average Velocity 1D motion. Very often it is convenient to model an object whose motion you analyze e.g. car, runner, stone, etc.) as
More informationKINEMATICS & DYNAMICS
KINEMATICS & DYNAMICS BY ADVANCED DIFFERENTIAL EQUATIONS Question (**+) In this question take g = 0 ms. A particle of mass M kg is released from rest from a height H m, and allowed to fall down through
More informationA path integral approach to the Langevin equation
A path integral approach to the Langevin equation - Ashok Das Reference: A path integral approach to the Langevin equation, A. Das, S. Panda and J. R. L. Santos, arxiv:1411.0256 (to be published in Int.
More informationRobotics. Control Theory. Marc Toussaint U Stuttgart
Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,
More informationVibrational Motion. Chapter 5. P. J. Grandinetti. Sep. 13, Chem P. J. Grandinetti (Chem. 4300) Vibrational Motion Sep.
Vibrational Motion Chapter 5 P. J. Grandinetti Chem. 4300 Sep. 13, 2017 P. J. Grandinetti (Chem. 4300) Vibrational Motion Sep. 13, 2017 1 / 20 Simple Harmonic Oscillator Simplest model for harmonic oscillator
More informationOPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28
OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from
More informationHJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011
Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationReinforcement Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina
Reinforcement Learning Introduction Introduction Unsupervised learning has no outcome (no feedback). Supervised learning has outcome so we know what to predict. Reinforcement learning is in between it
More informationOptimal Control. Quadratic Functions. Single variable quadratic function: Multi-variable quadratic function:
Optimal Control Control design based on pole-placement has non unique solutions Best locations for eigenvalues are sometimes difficult to determine Linear Quadratic LQ) Optimal control minimizes a quadratic
More informationIntroduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley
Introduction to Reachability Somil Bansal Hybrid Systems Lab, UC Berkeley Outline Introduction to optimal control Reachability as an optimal control problem Various shades of reachability Goal of This
More informationECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming
ECE7850 Lecture 7 Discrete Time Optimal Control and Dynamic Programming Discrete Time Optimal control Problems Short Introduction to Dynamic Programming Connection to Stabilization Problems 1 DT nonlinear
More informationExistence and uniqueness: Picard s theorem
Existence and uniqueness: Picard s theorem First-order equations Consider the equation y = f(x, y) (not necessarily linear). The equation dictates a value of y at each point (x, y), so one would expect
More informationOptimal Control. Lecture 18. Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen. March 29, Ref: Bryson & Ho Chapter 4.
Optimal Control Lecture 18 Hamilton-Jacobi-Bellman Equation, Cont. John T. Wen Ref: Bryson & Ho Chapter 4. March 29, 2004 Outline Hamilton-Jacobi-Bellman (HJB) Equation Iterative solution of HJB Equation
More informationGame Theory Extra Lecture 1 (BoB)
Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder,
More informationMATH 4B Differential Equations, Fall 2016 Final Exam Study Guide
MATH 4B Differential Equations, Fall 2016 Final Exam Study Guide GENERAL INFORMATION AND FINAL EXAM RULES The exam will have a duration of 3 hours. No extra time will be given. Failing to submit your solutions
More informationTime Evolution of Speed
ENERGY & MOMENTUM Newton's 2 nd Law: Time Evolution of Speed Describes how force affects velocity of the CM How do forces affect speed of the CM? Depends on relative direction of force and velocity: F
More informationMIT Manufacturing Systems Analysis Lectures 19 21
MIT 2.852 Manufacturing Systems Analysis Lectures 19 21 Scheduling: Real-Time Control of Manufacturing Systems Stanley B. Gershwin Spring, 2007 Copyright c 2007 Stanley B. Gershwin. Definitions Events
More informationMATH 2250 Final Exam Solutions
MATH 225 Final Exam Solutions Tuesday, April 29, 28, 6: 8:PM Write your name and ID number at the top of this page. Show all your work. You may refer to one double-sided sheet of notes during the exam
More informationSolutions to Homework 1, Introduction to Differential Equations, 3450: , Dr. Montero, Spring y(x) = ce 2x + e x
Solutions to Homewor 1, Introduction to Differential Equations, 3450:335-003, Dr. Montero, Spring 2009 problem 2. The problem says that the function yx = ce 2x + e x solves the ODE y + 2y = e x, and ass
More informationThermodynamic limit and phase transitions in non-cooperative games: some mean-field examples
Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples Conference on the occasion of Giovanni Colombo and Franco Ra... https://events.math.unipd.it/oscgc18/ Optimization,
More informationUCLA Chemical Engineering. Process & Control Systems Engineering Laboratory
Constrained Innite-Time Nonlinear Quadratic Optimal Control V. Manousiouthakis D. Chmielewski Chemical Engineering Department UCLA 1998 AIChE Annual Meeting Outline Unconstrained Innite-Time Nonlinear
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationEXAM. Exam #1. Math 3350 Summer II, July 21, 2000 ANSWERS
EXAM Exam #1 Math 3350 Summer II, 2000 July 21, 2000 ANSWERS i 100 pts. Problem 1. 1. In each part, find the general solution of the differential equation. dx = x2 e y We use the following sequence of
More informationDYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION
DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION UNIVERSITY OF MARYLAND: ECON 600. Alternative Methods of Discrete Time Intertemporal Optimization We will start by solving a discrete time intertemporal
More informationAn Application to Growth Theory
An Application to Growth Theory First let s review the concepts of solution function and value function for a maximization problem. Suppose we have the problem max F (x, α) subject to G(x, β) 0, (P) x
More informationCHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME. Chapter2 p. 1/67
CHAPTER 2 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Chapter2 p. 1/67 THE MAXIMUM PRINCIPLE: CONTINUOUS TIME Main Purpose: Introduce the maximum principle as a necessary condition to be satisfied by any optimal
More informationME Thermodynamics I
HW-03 (25 points) i) Given: for writing Given, Find, Basic equations Rigid tank containing nitrogen gas in two sections initially separated by a membrane. Find: Initial density (kg/m3) of nitrogen gas
More informationNumerical Optimal Control Overview. Moritz Diehl
Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize
More informationIntroduction to optimal control theory in continuos time (with economic applications) Salvatore Federico
Introduction to optimal control theory in continuos time (with economic applications) Salvatore Federico June 26, 2017 2 Contents 1 Introduction to optimal control problems in continuous time 5 1.1 From
More informationPredicting the future with Newton s Second Law
Predicting the future with Newton s Second Law To represent the motion of an object (ignoring rotations for now), we need three functions x(t), y(t), and z(t), which describe the spatial coordinates of
More informationIntroduction to First Order Equations Sections
A B I L E N E C H R I S T I A N U N I V E R S I T Y Department of Mathematics Introduction to First Order Equations Sections 2.1-2.3 Dr. John Ehrke Department of Mathematics Fall 2012 Course Goals The
More informationChapter 5. Pontryagin s Minimum Principle (Constrained OCP)
Chapter 5 Pontryagin s Minimum Principle (Constrained OCP) 1 Pontryagin s Minimum Principle Plant: (5-1) u () t U PI: (5-2) Boundary condition: The goal is to find Optimal Control. 2 Pontryagin s Minimum
More informationLinear conic optimization for nonlinear optimal control
Linear conic optimization for nonlinear optimal control Didier Henrion 1,2,3, Edouard Pauwels 1,2 Draft of July 15, 2014 Abstract Infinite-dimensional linear conic formulations are described for nonlinear
More informationChristopher Watkins and Peter Dayan. Noga Zaslavsky. The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015
Q-Learning Christopher Watkins and Peter Dayan Noga Zaslavsky The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015 Noga Zaslavsky Q-Learning (Watkins & Dayan, 1992)
More informationEE C128 / ME C134 Feedback Control Systems
EE C128 / ME C134 Feedback Control Systems Lecture Additional Material Introduction to Model Predictive Control Maximilian Balandat Department of Electrical Engineering & Computer Science University of
More informationECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko
ECON 582: An Introduction to the Theory of Optimal Control (Chapter 7, Acemoglu) Instructor: Dmytro Hryshko Continuous-time optimization involves maximization wrt to an innite dimensional object, an entire
More informationMA 102 Mathematics II Lecture Feb, 2015
MA 102 Mathematics II Lecture 1 20 Feb, 2015 Differential Equations An equation containing derivatives is called a differential equation. The origin of differential equations Many of the laws of nature
More informationMATH 18.01, FALL PROBLEM SET #5 SOLUTIONS (PART II)
MATH 8, FALL 7 - PROBLEM SET #5 SOLUTIONS (PART II (Oct ; Antiderivatives; + + 3 7 points Recall that in pset 3A, you showed that (d/dx tanh x x Here, tanh (x denotes the inverse to the hyperbolic tangent
More information1 Simple Harmonic Oscillator
Physics 1a Waves Lecture 3 Caltech, 10/09/18 1 Simple Harmonic Oscillator 1.4 General properties of Simple Harmonic Oscillator 1.4.4 Superposition of two independent SHO Suppose we have two SHOs described
More informationBayesian Decision Theory in Sensorimotor Control
Bayesian Decision Theory in Sensorimotor Control Matthias Freiberger, Martin Öttl Signal Processing and Speech Communication Laboratory Advanced Signal Processing Matthias Freiberger, Martin Öttl Advanced
More informationModeling Mechanical Systems
Modeling Mechanical Systems Mechanical systems can be either translational or rotational. Although the fundamental relationships for both types are derived from Newton s law, they are different enough
More informationIntegration by Parts. MAT 126, Week 2, Thursday class. Xuntao Hu
MAT 126, Week 2, Thursday class Xuntao Hu Recall that the substitution rule is a combination of the FTC and the chain rule. We can also combine the FTC and the product rule: d dx [f (x)g(x)] = f (x)g (x)
More informationThe Method Of Direction Fields With Illustrative Examples
The Method Of Direction Fields With Illustrative Examples By Lei Zeng Abstract. When the exact solution of a differential equation is impossible to find, the study of its direction field can provide valuable
More informationSection Taylor and Maclaurin Series
Section.0 Taylor and Maclaurin Series Ruipeng Shen Feb 5 Taylor and Maclaurin Series Main Goal: How to find a power series representation for a smooth function us assume that a smooth function has a power
More informationECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko
ECON 582: Dynamic Programming (Chapter 6, Acemoglu) Instructor: Dmytro Hryshko Indirect Utility Recall: static consumer theory; J goods, p j is the price of good j (j = 1; : : : ; J), c j is consumption
More informationLecture 6.1 Work and Energy During previous lectures we have considered many examples, which can be solved using Newtonian approach, in particular,
Lecture 6. Work and Energy During previous lectures we have considered many examples, which can be solved using Newtonian approach, in particular, Newton's second law. However, this is not always the most
More informationMaximum Process Problems in Optimal Control Theory
J. Appl. Math. Stochastic Anal. Vol. 25, No., 25, (77-88) Research Report No. 423, 2, Dept. Theoret. Statist. Aarhus (2 pp) Maximum Process Problems in Optimal Control Theory GORAN PESKIR 3 Given a standard
More informationES 111 Mathematical Methods in the Earth Sciences Lecture Outline 15 - Tues 20th Nov 2018 First and Higher Order Differential Equations
ES 111 Mathematical Methods in the Earth Sciences Lecture Outline 15 - Tues 20th Nov 2018 First and Higher Order Differential Equations Integrating Factor Here is a powerful technique which will work (only!)
More informationUncertainty quantification and systemic risk
Uncertainty quantification and systemic risk Josselin Garnier (Université Paris Diderot) with George Papanicolaou and Tzu-Wei Yang (Stanford University) February 3, 2016 Modeling systemic risk We consider
More informationPHYSICS 110A : CLASSICAL MECHANICS HW 1 SOLUTIONS. r = R vt
PHYSICS 11A : CLASSICAL MECHANICS HW 1 SOLUTIONS 2) Taylor 1.46 a) The equations of motion for the puck are: r = R vt φ = Assuming the puck is launched from the position φ =. Technically with the polar
More informationStrauss PDEs 2e: Section Exercise 4 Page 1 of 6
Strauss PDEs 2e: Section 5.3 - Exercise 4 Page of 6 Exercise 4 Consider the problem u t = ku xx for < x < l, with the boundary conditions u(, t) = U, u x (l, t) =, and the initial condition u(x, ) =, where
More information2t t dt.. So the distance is (t2 +6) 3/2
Math 8, Solutions to Review for the Final Exam Question : The distance is 5 t t + dt To work that out, integrate by parts with u t +, so that t dt du The integral is t t + dt u du u 3/ (t +) 3/ So the
More informationhttp://geocities.com/kenahn7/ Today in this class Chap.2, Sec.1-7 Motion along a straight line 1. Position and displacement 2. 3. Acceleration Example: Motion with a constant acceleration Position and
More informationMath 2214 Solution Test 1D Spring 2015
Math 2214 Solution Test 1D Spring 2015 Problem 1: A 600 gallon open top tank initially holds 300 gallons of fresh water. At t = 0, a brine solution containing 3 lbs of salt per gallon is poured into the
More informationMAS153/MAS159. MAS153/MAS159 1 Turn Over SCHOOL OF MATHEMATICS AND STATISTICS hours. Mathematics (Materials) Mathematics For Chemists
Data provided: Formula sheet MAS53/MAS59 SCHOOL OF MATHEMATICS AND STATISTICS Mathematics (Materials Mathematics For Chemists Spring Semester 203 204 3 hours All questions are compulsory. The marks awarded
More informationXXIX Applications of Differential Equations
MATHEMATICS 01-BNK-05 Advanced Calculus Martin Huard Winter 015 1. Suppose that the rate at which a population of size yt at time t changes is proportional to the amount present. This gives rise to the
More informationRobotics: Science & Systems [Topic 6: Control] Prof. Sethu Vijayakumar Course webpage:
Robotics: Science & Systems [Topic 6: Control] Prof. Sethu Vijayakumar Course webpage: http://wcms.inf.ed.ac.uk/ipab/rss Control Theory Concerns controlled systems of the form: and a controller of the
More information28. Pendulum phase portrait Draw the phase portrait for the pendulum (supported by an inextensible rod)
28. Pendulum phase portrait Draw the phase portrait for the pendulum (supported by an inextensible rod) θ + ω 2 sin θ = 0. Indicate the stable equilibrium points as well as the unstable equilibrium points.
More informationModule 05 Introduction to Optimal Control
Module 05 Introduction to Optimal Control Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/ taha/index.html October 8, 2015
More informationMATH : Calculus II (42809) SYLLABUS, Spring 2010 MW 4-5:50PM, JB- 138
MATH -: Calculus II (489) SYLLABUS, Spring MW 4-5:5PM, JB- 38 John Sarli, JB-36 O ce Hours: MTW 3-4PM, and by appointment (99) 537-5374 jsarli@csusb.edu Text: Calculus of a Single Variable, Larson/Hostetler/Edwards
More informationWork and kinetic energy. LA info session today at 5pm in UMC235 CAPA homework due tomorrow night.
Work and kinetic energy LA info session today at 5pm in UMC235 CAPA homework due tomorrow night. 1 Work I apply a force of 2N in the x direction to an object that moves 5m in x. How much work have I done
More informationMath 211. Lecture #6. Linear Equations. September 9, 2002
1 Math 211 Lecture #6 Linear Equations September 9, 2002 2 Air Resistance 2 Air Resistance Acts in the direction opposite to the velocity. 2 Air Resistance Acts in the direction opposite to the velocity.
More informationPhysics 9 Spring 2012 Midterm 1 Solutions
Physics 9 Spring 22 NAME: TA: Physics 9 Spring 22 Midterm s For the midterm, you may use one sheet of notes with whatever you want to put on it, front and back. Please sit every other seat, and please
More informationLecture 3: Hamilton-Jacobi-Bellman Equations. Distributional Macroeconomics. Benjamin Moll. Part II of ECON Harvard University, Spring
Lecture 3: Hamilton-Jacobi-Bellman Equations Distributional Macroeconomics Part II of ECON 2149 Benjamin Moll Harvard University, Spring 2018 1 Outline 1. Hamilton-Jacobi-Bellman equations in deterministic
More informationODEs. September 7, Consider the following system of two coupled first-order ordinary differential equations (ODEs): A =
ODEs September 7, 2017 In [1]: using Interact, PyPlot 1 Exponential growth and decay Consider the following system of two coupled first-order ordinary differential equations (ODEs): d x/dt = A x for the
More informationAn efficient approach to stochastic optimal control. Bert Kappen SNN Radboud University Nijmegen the Netherlands
An efficient approach to stochastic optimal control Bert Kappen SNN Radboud University Nijmegen the Netherlands Bert Kappen Examples of control tasks Motor control Bert Kappen Pascal workshop, 27-29 May
More informationC. Show your answer in part B agrees with your answer in part A in the limit that the constant c 0.
Problem #1 A. A projectile of mass m is shot vertically in the gravitational field. Its initial velocity is v o. Assuming there is no air resistance, how high does m go? B. Now assume the projectile is
More informationWeek 9 Generators, duality, change of measure
Week 9 Generators, duality, change of measure Jonathan Goodman November 18, 013 1 Generators This section describes a common abstract way to describe many of the differential equations related to Markov
More informationMATH 1242 FINAL EXAM Spring,
MATH 242 FINAL EXAM Spring, 200 Part I (MULTIPLE CHOICE, NO CALCULATORS).. Find 2 4x3 dx. (a) 28 (b) 5 (c) 0 (d) 36 (e) 7 2. Find 2 cos t dt. (a) 2 sin t + C (b) 2 sin t + C (c) 2 cos t + C (d) 2 cos t
More informationPhysics 2101 S c e t c i cti n o 3 n 3 March 31st Announcements: Quiz today about Ch. 14 Class Website:
Physics 2101 Section 3 March 31 st Announcements: Quiz today about Ch. 14 Class Website: http://www.phys.lsu.edu/classes/spring2010/phys2101 3/ http://www.phys.lsu.edu/~jzhang/teaching.html Simple Harmonic
More informationVELOCITY. If you have a graph of position and you take the derivative, what would the derivative represent? Position. Time
VELOCITY If you have a graph of position and you take the derivative, what would the derivative represent? Position Time Average rate of Change What is the average rate of change of temperature over the
More informationMotion in One Dimension
Motion in One Dimension Much of the physics we ll learn this semester will deal with the motion of objects We start with the simple case of one-dimensional motion Or, motion in x: As always, we begin by
More informationRisk-Sensitive and Robust Mean Field Games
Risk-Sensitive and Robust Mean Field Games Tamer Başar Coordinated Science Laboratory Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign Urbana, IL - 6181 IPAM
More informationSimple Harmonic Motion
Simple Harmonic Motion (FIZ 101E - Summer 2018) July 29, 2018 Contents 1 Introduction 2 2 The Spring-Mass System 2 3 The Energy in SHM 5 4 The Simple Pendulum 6 5 The Physical Pendulum 8 6 The Damped Oscillations
More informationAPPLICATIONS OF INTEGRATION
6 APPLICATIONS OF INTEGRATION APPLICATIONS OF INTEGRATION 6.5 Average Value of a Function In this section, we will learn about: Applying integration to find out the average value of a function. AVERAGE
More informationSlides II - Dynamic Programming
Slides II - Dynamic Programming Julio Garín University of Georgia Macroeconomic Theory II (Ph.D.) Spring 2017 Macroeconomic Theory II Slides II - Dynamic Programming Spring 2017 1 / 32 Outline 1. Lagrangian
More informationDispersion relations, linearization and linearized dynamics in PDE models
Dispersion relations, linearization and linearized dynamics in PDE models 1 Dispersion relations Suppose that u(x, t) is a function with domain { < x 0}, and it satisfies a linear, constant coefficient
More informationDEPARTMENT OF MATHEMATICS AND STATISTICS UNIVERSITY OF MASSACHUSETTS. MATH 233 SOME SOLUTIONS TO EXAM 2 Fall 2018
DEPARTMENT OF MATHEMATICS AND STATISTICS UNIVERSITY OF MASSACHUSETTS MATH 233 SOME SOLUTIONS TO EXAM 2 Fall 208 Version A refers to the regular exam and Version B to the make-up. Version A. A particle
More informationOverview of the Seminar Topic
Overview of the Seminar Topic Simo Särkkä Laboratory of Computational Engineering Helsinki University of Technology September 17, 2007 Contents 1 What is Control Theory? 2 History
More informationReview: control, feedback, etc. Today s topic: state-space models of systems; linearization
Plan of the Lecture Review: control, feedback, etc Today s topic: state-space models of systems; linearization Goal: a general framework that encompasses all examples of interest Once we have mastered
More informationIntroduction to Continuous-Time Dynamic Optimization: Optimal Control Theory
Econ 85/Chatterjee Introduction to Continuous-ime Dynamic Optimization: Optimal Control heory 1 States and Controls he concept of a state in mathematical modeling typically refers to a specification of
More informationThese will be no tutorials for Math on Tuesday April 26.
Worksheet The purpose of this worksheet is 1. To understand how the differential equation describing simple harmonic motion is derived. 2. To explore how to predict what the solution to this differential
More informationLecture 9: Kinetic Energy and Work 1
Lecture 9: Kinetic Energy and Work 1 CHAPTER 6: Work and Kinetic Energy The concept of WORK has a very precise definition in physics. Work is a physical quantity produced when a Force moves an object through
More informationToulouse School of Economics, M2 Macroeconomics 1 Professor Franck Portier. Exam Solution
Toulouse School of Economics, 2013-2014 M2 Macroeconomics 1 Professor Franck Portier Exam Solution This is a 3 hours exam. Class slides and any handwritten material are allowed. You must write legibly.
More informationSurface x(u, v) and curve α(t) on it given by u(t) & v(t). Math 4140/5530: Differential Geometry
Surface x(u, v) and curve α(t) on it given by u(t) & v(t). α du dv (t) x u dt + x v dt Surface x(u, v) and curve α(t) on it given by u(t) & v(t). α du dv (t) x u dt + x v dt ( ds dt )2 Surface x(u, v)
More informationPart A: From (14.19) in your text (see notes Stability1 ), we have
Prob14.4 14.4 Consider a round-rotor generator delivering a steady-state power P G =.5 to an infinite bus through a transmission line with reactance X L =.4. Assume that E a =1.8, V =1/_, H=5sec, and X
More informationMultiple Choice Review Problems
Multiple Choice Review Problems 1. (NC) Which graph best represents the position of a particle, st ( ), as a function of time, if the particle's velocity is negative and the particle's acceleration is
More informationSolution of Stochastic Optimal Control Problems and Financial Applications
Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty
More informationODE Math 3331 (Summer 2014) June 16, 2014
Page 1 of 12 Please go to the next page... Sample Midterm 1 ODE Math 3331 (Summer 2014) June 16, 2014 50 points 1. Find the solution of the following initial-value problem 1. Solution (S.O.V) dt = ty2,
More information[A + 1 ] + (1 ) v: : (b) Show: the derivative of T at v = v 0 < 0 is: = (v 0 ) (1 ) ; [A + 1 ]
Homework #2 Economics 4- Due Wednesday, October 5 Christiano. This question is designed to illustrate Blackwell's Theorem, Theorem 3.3 on page 54 of S-L. That theorem represents a set of conditions that
More informationMath 162: Calculus IIA
Math 62: Calculus IIA Final Exam ANSWERS December 9, 26 Part A. (5 points) Evaluate the integral x 4 x 2 dx Substitute x 2 cos θ: x 8 cos dx θ ( 2 sin θ) dθ 4 x 2 2 sin θ 8 cos θ dθ 8 cos 2 θ cos θ dθ
More informationA Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley
A Tour of Reinforcement Learning The View from Continuous Control Benjamin Recht University of California, Berkeley trustable, scalable, predictable Control Theory! Reinforcement Learning is the study
More information