Lecture 5: Control Over Lossy Networks
|
|
- Henry Eugene Mosley
- 5 years ago
- Views:
Transcription
1 Lecture 5: Control Over Lossy Networks Yilin Mo July 2, Classical LQG Control The system: x k+1 = Ax k + Bu k + w k, y k = Cx k + v k x 0 N (0, Σ), w k N (0, Q), v k N (0, R). Information available for the controller at time k: Y k = (y 0,..., y k ). The control at time k is a function of the information Y k : u k (Y k ). The goal of a finite horizon LQG problem is to find a controller that minimizes the following quadratic cost: J(N) = min E u 0,...,u N N ( x T k W x k + u T ) k Uu k. k=0 1.1 Optimal Estimator Design Since the system is linear, the following Kalman filtering equations holds: 1. Initialization: 2. Prediction: 3. Correction: ˆx 0 1 = 0, P 0 1 = Σ. (1) ˆx k+1 k = Aˆx k + Bu k, P k+1 k = AP k A T + Q. (2) ˆx k+1 = ˆx k+1 k + P k+1 k C T (CP k+1 k C T + R) 1 (y k+1 C ˆx k+1 k ), (3) P k+1 = P k+1 k P k+1 k C T (CP k+1 k C T + R) 1 CP k+1 k. (4) And we have that E(x k Y k ) = ˆx k, Cov(x k Y k ) = P k. One important thing to notice: the P k is independent from u k. This is because the system is linear and hence we can subtract u k. 1
2 1.2 Optimal Controller Design and Separation Principle Now we can try to solve the optimal control design problem by dynamical programming. Define the value function [ N ( V (t) = min E x T k W x k + u T ) ] k Uu k u t,...,u N Clearly and Now, by Bellman equation We will guess that k=t J(N) = V (0), V (N) = Ex T NW x N. V (t) = min u t E ( x T t W x t + u T t Uu t + V (t + 1) ) (5) V (t) = Ex T t S t x t + c t (6) To prove this, we will use induction. Clearly S N = W and c N = 0. Now suppose (6) holds for t + 1, and we look at the following quantity: E ( x T t W x t + u T t Uu t + V (t + 1) ) = Ex T (W + A T S t+1 A)x + tr(s t+1 Q) + c t+1 + E [ u T t (U + B T S t+1 B)u t + x T t A T S t+1 Bu t + u T t B T S t+1 Ax t ] Notice that the controller do not know x t. Hence, let us rewrite x t as x t = ˆx t + x t ˆx t = ˆx t + e t. Theorem e k is independent of Y k and hence ˆx k and u k. 2. Proof. 1. Notice that Eˆx T k Sˆx k = Ex T k Sx k tr(sp k ) E(e k Y k ) = E(x k Y k ) E(ˆx k Y k ) = ˆx k ˆx k = 0. Hence, e k is linearly independent from Y k. Since e k and Y k are jointly Gaussian, e k and Y k are independent. 2. Since Ex T k Sx k = Eˆx T k Sˆx k + Eˆx T k Se k + Ee T k Sˆx k + Ee T k Se k = Eˆx T k Sˆx k tr(sp k ) 2
3 Now let us look at E [ u T t (U + B T S t+1 B)u t + x T t A T S t+1 Bu t + u T t B T ] S t+1 Ax t = E [ u T t (U + B T S t+1 B)u t + ˆx T t A T S t+1 Bu t + u T t B T ] S t+1 Aˆx t = E [ (u t u t ) T (U + B T S t+1 B)(u t u t ) ˆx T t A T S t+1 B(U + B T S t+1 B) 1 B T ] S t+1 Aˆx t where u t = (U + B T S t+1 B) 1 B T S t+1 Aˆx t. Hence V (t) = Ex T t (W + A T S t+1 A A T S t+1 B(U + B T S t+1 B) 1 B T S t+1 A)x t + c t+1 + tr(a T S t+1 B(U + B T S t+1 B) 1 B T S t+1 AP k ) + tr(s t+1 Q) Therefore S t = W + A T S t+1 A A T S t+1 B(U + B T S t+1 B) 1 B T S t+1 A, (7) and c t = c t+1 + tr(a T S t+1 B(U + B T S t+1 B) 1 B T S t+1 AP k ) + tr(s t+1 Q). Thus, J(N) = E(x T 0 S 0 x 0 ) + c 0 = tr(s 0 Σ) + c Infinite Horizon LQG problem Define J as J(N) J = lim N N. We consider the problem of finding a controller that minimizes the infinite horizon cost J. Notice that (7) is a Riccati equation. Hence, if N, then S k converges to S, which is the fixed solution of S = W + A T SA A T SB(U + B T SB) 1 B T SA, (8) The optimal controller is given by u k = (U + B T SB) 1 B T SAˆx k. 2 Witsenhausen s Counterexample Consider x 0 N (0, σ 2 ). 1. The first player knows x 0 and he computes an which is a function of x 0. x 1 = f(x 0 ), 3
4 2. The first player sends x 1 to the second player though a noisy channel. Therefore, the second player receives where v N (0, 1). y 2 = x 1 + v, 3. The second player then computes x 2 = g(y 2 ). The goal is to minimize the following cost function J = min f,g E k2 (x 0 x 1 ) 2 + (x 1 x 2 ) 2 Alternatively, one can consider the following equivalent scheme: 1. The controller knows x 0 and it computes an control u which is a function of x 0. u = f(x 0 ), 2. The state of the system satisfies the following update equation: x 1 = x 0 + u. 3. The second player observe the system via a noisy sensor: where v N (0, 1). y 2 = x 1 + v, 4. The second player then computes the state estimate x 2 = g(y 2 ). The goal is to minimize the following cost function 2.1 Optimal linear strategy J = min f,g E k2 u 2 + (x 1 x 2 ) 2 We adopt the first setting. Consider that both f(x) = λx and g(x) = µx are linear, then J = min λ,µ E k2 (1 λ) 2 x (λx 0 µ(λx 0 + v)) 2 Therefore, and λ = arg min λ µ = λ2 σ λ 2 σ 2, If k 2 σ 2 = 1 and k 0, then λ 1 and J 1. k 2 σ 2 (1 λ) 2 + λ2 σ λ 2 σ 2. 4
5 3 Nonlinear strategy One can prove that for small k and k 2 σ 2 = 0, the following design is better than the linear design: f(x) = σsgn(x), g(x) = σ 1 e 2σx 1 + e 2σx. 4 Control Over Lossy Networks The system: x k+1 = Ax k + ν k Bu k + w k, y k = Cx k + v k x 0 N (0, Σ), w k N (0, Q), v k N (0, R). ν k is an i.i.d. Bernouli process with P (ν k = 1) = λ. The goal of a finite horizon LQG problem is to find a controller that minimizes the following quadratic cost: J = 4.1 TCP case min u 0,...,u N lim N 1 N N E k=0 Information available for the controller at time k: ( x T k W x k + ν k u T k Uu k ). I k = (y 0,..., y k, ν 0,..., ν k ). J is finite if and only if the following Riccati equation has a positive semidefinite solution: Optimal Filter: S = W + A T SA λa T SB(U + B T SB) 1 B T SA, (9) 1. Initialization: ˆx 0 1 = 0, P 0 1 = Σ. (10) 2. Prediction: ˆx k+1 k = Aˆx k + ν k Bu k, P k+1 k = AP k A T + Q. (11) 3. Correction: ˆx k+1 = ˆx k+1 k + P k+1 k C T (CP k+1 k C T + R) 1 (y k+1 C ˆx k+1 k ), (12) P k+1 = P k+1 k P k+1 k C T (CP k+1 k C T + R) 1 CP k+1 k. (13) 5
6 Optimal Control: The optimal controller is given by where S is the solution of (9). 4.2 UDP case u k = (U + B T SB) 1 B T SAˆx k, Information available for the controller at time k: Y k = (y 0,..., y k ). We do not know whether u k has been applied to the system or not. The control actually affect the estimation performance. The optimal control law and the stability of the system is unknown. 6
Lecture 3: Functions of Symmetric Matrices
Lecture 3: Functions of Symmetric Matrices Yilin Mo July 2, 2015 1 Recap 1 Bayes Estimator: (a Initialization: (b Correction: f(x 0 Y 1 = f(x 0 f(x k Y k = αf(y k x k f(x k Y k 1, where ( 1 α = f(y k x
More informationLQG CONTROL WITH MISSING OBSERVATION AND CONTROL PACKETS. Bruno Sinopoli, Luca Schenato, Massimo Franceschetti, Kameshwar Poolla, Shankar Sastry
LQG CONTROL WITH MISSING OBSERVATION AND CONTROL PACKETS Bruno Sinopoli, Luca Schenato, Massimo Franceschetti, Kameshwar Poolla, Shankar Sastry Department of Electrical Engineering and Computer Sciences
More informationFalse Data Injection Attacks in Control Systems
False Data Injection Attacks in Control Systems Yilin Mo, Bruno Sinopoli Department of Electrical and Computer Engineering, Carnegie Mellon University First Workshop on Secure Control Systems Bruno Sinopoli
More informationLecture 10 Linear Quadratic Stochastic Control with Partial State Observation
EE363 Winter 2008-09 Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation partially observed linear-quadratic stochastic control problem estimation-control separation principle
More informationSecure Control Against Replay Attacks
Secure Control Against Replay Attacks Bruno Sinopoli, Yilin Mo Department of Electrical and Computer Engineering, Carnegie Mellon Trust Autumn 2009 Conference Bruno Sinopoli (Carnegie Mellon) Secure Control
More informationTime Varying Optimal Control with Packet Losses.
Time Varying Optimal Control with Packet Losses. Bruno Sinopoli, Luca Schenato, Massimo Franceschetti, Kameshwar Poolla, Shankar S. Sastry Department of Electrical Engineering and Computer Sciences University
More informationNetworked Control Systems: Estimation and Control over Lossy Networks
Noname manuscript No. (will be inserted by the editor) Networked Control Systems: Estimation and Control over Lossy Networks João P. Hespanha Alexandre R. Mesquita the date of receipt and acceptance should
More informationSteady State Kalman Filter
Steady State Kalman Filter Infinite Horizon LQ Control: ẋ = Ax + Bu R positive definite, Q = Q T 2Q 1 2. (A, B) stabilizable, (A, Q 1 2) detectable. Solve for the positive (semi-) definite P in the ARE:
More informationLinear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters
Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 204 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationOptimal control and estimation
Automatic Control 2 Optimal control and estimation Prof. Alberto Bemporad University of Trento Academic year 2010-2011 Prof. Alberto Bemporad (University of Trento) Automatic Control 2 Academic year 2010-2011
More information6.241 Dynamic Systems and Control
6.241 Dynamic Systems and Control Lecture 24: H2 Synthesis Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology May 4, 2011 E. Frazzoli (MIT) Lecture 24: H 2 Synthesis May
More informationCALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems. CDS 110b
CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems CDS 110b R. M. Murray Kalman Filters 25 January 2006 Reading: This set of lectures provides a brief introduction to Kalman filtering, following
More informationA Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition
A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition Junfeng Wu, Karl Henrik Johansson and Ling Shi E-mail: jfwu@ust.hk Stockholm, 9th, January 2014 1 / 19 Outline Outline
More informationGaussians. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics
Gaussians Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Outline Univariate Gaussian Multivariate Gaussian Law of Total Probability Conditioning
More informationFactor Analysis and Kalman Filtering (11/2/04)
CS281A/Stat241A: Statistical Learning Theory Factor Analysis and Kalman Filtering (11/2/04) Lecturer: Michael I. Jordan Scribes: Byung-Gon Chun and Sunghoon Kim 1 Factor Analysis Factor analysis is used
More informationJoint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:
Joint Distributions Joint Distributions A bivariate normal distribution generalizes the concept of normal distribution to bivariate random variables It requires a matrix formulation of quadratic forms,
More informationRECURSIVE ESTIMATION AND KALMAN FILTERING
Chapter 3 RECURSIVE ESTIMATION AND KALMAN FILTERING 3. The Discrete Time Kalman Filter Consider the following estimation problem. Given the stochastic system with x k+ = Ax k + Gw k (3.) y k = Cx k + Hv
More informationLecture 19 Observability and state estimation
EE263 Autumn 2007-08 Stephen Boyd Lecture 19 Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time
More informationCALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems. CDS 110b
CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems CDS 110b R. M. Murray Kalman Filters 14 January 2007 Reading: This set of lectures provides a brief introduction to Kalman filtering, following
More informationContents lecture 5. Automatic Control III. Summary of lecture 4 (II/II) Summary of lecture 4 (I/II) u y F r. Lecture 5 H 2 and H loop shaping
Contents lecture 5 Automatic Control III Lecture 5 H 2 and H loop shaping Thomas Schön Division of Systems and Control Department of Information Technology Uppsala University. Email: thomas.schon@it.uu.se,
More informationLecture 3. Optimization Problems and Iterative Algorithms
Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex
More informationLQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin
LQR, Kalman Filter, and LQG Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin May 2015 Linear Quadratic Regulator (LQR) Consider a linear system
More information26. Filtering. ECE 830, Spring 2014
26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem
More informationMath 24 Spring 2012 Questions (mostly) from the Textbook
Math 24 Spring 2012 Questions (mostly) from the Textbook 1. TRUE OR FALSE? (a) The zero vector space has no basis. (F) (b) Every vector space that is generated by a finite set has a basis. (c) Every vector
More informationLecture 5 Linear Quadratic Stochastic Control
EE363 Winter 2008-09 Lecture 5 Linear Quadratic Stochastic Control linear-quadratic stochastic control problem solution via dynamic programming 5 1 Linear stochastic system linear dynamical system, over
More informationOptimization-Based Control
Optimization-Based Control Richard M. Murray Control and Dynamical Systems California Institute of Technology DRAFT v1.7a, 19 February 2008 c California Institute of Technology All rights reserved. This
More informationKalman Filtering. Namrata Vaswani. March 29, Kalman Filter as a causal MMSE estimator
Kalman Filtering Namrata Vaswani March 29, 2018 Notes are based on Vincent Poor s book. 1 Kalman Filter as a causal MMSE estimator Consider the following state space model (signal and observation model).
More information1 Kalman Filter Introduction
1 Kalman Filter Introduction You should first read Chapter 1 of Stochastic models, estimation, and control: Volume 1 by Peter S. Maybec (available here). 1.1 Explanation of Equations (1-3) and (1-4) Equation
More informationA Study of Covariances within Basic and Extended Kalman Filters
A Study of Covariances within Basic and Extended Kalman Filters David Wheeler Kyle Ingersoll December 2, 2013 Abstract This paper explores the role of covariance in the context of Kalman filters. The underlying
More informationECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form
ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER KRISTOFFER P. NIMARK The Kalman Filter We will be concerned with state space systems of the form X t = A t X t 1 + C t u t 0.1 Z t
More information9 Controller Discretization
9 Controller Discretization In most applications, a control system is implemented in a digital fashion on a computer. This implies that the measurements that are supplied to the control system must be
More informationKalman filtering with intermittent heavy tailed observations
Kalman filtering with intermittent heavy tailed observations Sabina Zejnilović Abstract In large wireless sensor networks, data can experience loss and significant delay which from the aspect of control
More informationBayesian Decision Theory in Sensorimotor Control
Bayesian Decision Theory in Sensorimotor Control Matthias Freiberger, Martin Öttl Signal Processing and Speech Communication Laboratory Advanced Signal Processing Matthias Freiberger, Martin Öttl Advanced
More information6 OUTPUT FEEDBACK DESIGN
6 OUTPUT FEEDBACK DESIGN When the whole sate vector is not available for feedback, i.e, we can measure only y = Cx. 6.1 Review of observer design Recall from the first class in linear systems that a simple
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationObservability and state estimation
EE263 Autumn 2015 S Boyd and S Lall Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time observability
More informationNetworked Sensing, Estimation and Control Systems
Networked Sensing, Estimation and Control Systems Vijay Gupta University of Notre Dame Richard M. Murray California Institute of echnology Ling Shi Hong Kong University of Science and echnology Bruno Sinopoli
More informationKalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q
Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I
More informationTopics. Review of lecture 2/11 Error, Residual and Condition Number. Review of lecture 2/16 Backward Error Analysis The General Case 1 / 22
Topics Review of lecture 2/ Error, Residual and Condition Number Review of lecture 2/6 Backward Error Analysis The General Case / 22 Theorem (Calculation of 2 norm of a symmetric matrix) If A = A t is
More informationLecture 4: Least Squares (LS) Estimation
ME 233, UC Berkeley, Spring 2014 Xu Chen Lecture 4: Least Squares (LS) Estimation Background and general solution Solution in the Gaussian case Properties Example Big picture general least squares estimation:
More informationHamilton-Jacobi-Bellman Equation Feb 25, 2008
Hamilton-Jacobi-Bellman Equation Feb 25, 2008 What is it? The Hamilton-Jacobi-Bellman (HJB) equation is the continuous-time analog to the discrete deterministic dynamic programming algorithm Discrete VS
More informationLinear Encoder-Decoder-Controller Design over Channels with Packet Loss and Quantization Noise
Linear Encoder-Decoder-Controller Design over Channels with Packet Loss and Quantization Noise S. Dey, A. Chiuso and L. Schenato Abstract In this paper we consider the problem of designing coding and decoding
More informationTropical Polynomials
1 Tropical Arithmetic Tropical Polynomials Los Angeles Math Circle, May 15, 2016 Bryant Mathews, Azusa Pacific University In tropical arithmetic, we define new addition and multiplication operations on
More information4 Derivations of the Discrete-Time Kalman Filter
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time
More informationx(n + 1) = Ax(n) and y(n) = Cx(n) + 2v(n) and C = x(0) = ξ 1 ξ 2 Ex(0)x(0) = I
A-AE 567 Final Homework Spring 213 You will need Matlab and Simulink. You work must be neat and easy to read. Clearly, identify your answers in a box. You will loose points for poorly written work. You
More informationProbability Background
Probability Background Namrata Vaswani, Iowa State University August 24, 2015 Probability recap 1: EE 322 notes Quick test of concepts: Given random variables X 1, X 2,... X n. Compute the PDF of the second
More informationStatistical Filtering and Control for AI and Robotics. Part II. Linear methods for regression & Kalman filtering
Statistical Filtering and Control for AI and Robotics Part II. Linear methods for regression & Kalman filtering Riccardo Muradore 1 / 66 Outline Linear Methods for Regression Gaussian filter Stochastic
More informationLinear-Quadratic Optimal Control: Full-State Feedback
Chapter 4 Linear-Quadratic Optimal Control: Full-State Feedback 1 Linear quadratic optimization is a basic method for designing controllers for linear (and often nonlinear) dynamical systems and is actually
More informationAP Calculus Chapter 3 Testbank (Mr. Surowski)
AP Calculus Chapter 3 Testbank (Mr. Surowski) Part I. Multiple-Choice Questions (5 points each; please circle the correct answer.). If f(x) = 0x 4 3 + x, then f (8) = (A) (B) 4 3 (C) 83 3 (D) 2 3 (E) 2
More informationMachine Learning 4771
Machine Learning 4771 Instructor: ony Jebara Kalman Filtering Linear Dynamical Systems and Kalman Filtering Structure from Motion Linear Dynamical Systems Audio: x=pitch y=acoustic waveform Vision: x=object
More informationEE363 homework 2 solutions
EE363 Prof. S. Boyd EE363 homework 2 solutions. Derivative of matrix inverse. Suppose that X : R R n n, and that X(t is invertible. Show that ( d d dt X(t = X(t dt X(t X(t. Hint: differentiate X(tX(t =
More informationMath 651 Introduction to Numerical Analysis I Fall SOLUTIONS: Homework Set 1
ath 651 Introduction to Numerical Analysis I Fall 2010 SOLUTIONS: Homework Set 1 1. Consider the polynomial f(x) = x 2 x 2. (a) Find P 1 (x), P 2 (x) and P 3 (x) for f(x) about x 0 = 0. What is the relation
More informationA Tour of Reinforcement Learning The View from Continuous Control. Benjamin Recht University of California, Berkeley
A Tour of Reinforcement Learning The View from Continuous Control Benjamin Recht University of California, Berkeley trustable, scalable, predictable Control Theory! Reinforcement Learning is the study
More informationSECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS
SECTION C: CONTINUOUS OPTIMISATION LECTURE : THE METHOD OF LAGRANGE MULTIPLIERS HONOUR SCHOOL OF MATHEMATICS OXFORD UNIVERSITY HILARY TERM 005 DR RAPHAEL HAUSER. Examples. In this lecture we will take
More informationPartial Fraction Decomposition
Partial Fraction Decomposition As algebra students we have learned how to add and subtract fractions such as the one show below, but we probably have not been taught how to break the answer back apart
More informationState estimation and the Kalman filter
State estimation and the Kalman filter PhD, David Di Ruscio Telemark university college Department of Technology Systems and Control Engineering N-3914 Porsgrunn, Norway Fax: +47 35 57 52 50 Tel: +47 35
More informationKalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University
Kalman Filter 16-385 Computer Vision (Kris Kitani) Carnegie Mellon University Examples up to now have been discrete (binary) random variables Kalman filtering can be seen as a special case of a temporal
More informationFeedback Control of Turbulent Wall Flows
Feedback Control of Turbulent Wall Flows Dipartimento di Ingegneria Aerospaziale Politecnico di Milano Outline Introduction Standard approach Wiener-Hopf approach Conclusions Drag reduction A model problem:
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More informationECE7850 Lecture 7. Discrete Time Optimal Control and Dynamic Programming
ECE7850 Lecture 7 Discrete Time Optimal Control and Dynamic Programming Discrete Time Optimal control Problems Short Introduction to Dynamic Programming Connection to Stabilization Problems 1 DT nonlinear
More informationEvent-based State Estimation of Linear Dynamical Systems: Communication Rate Analysis
2014 American Control Conference Estimation of Linear Dynamical Systems: Dawei Shi, Department of Electrical and Computer Engineering, University of Alberta, Edmonton, AB, Canada Department of Electronic
More information= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt +
ECE 553, Spring 8 Posted: May nd, 8 Problem Set #7 Solution Solutions: 1. The optimal controller is still the one given in the solution to the Problem 6 in Homework #5: u (x, t) = p(t)x k(t), t. The minimum
More informationLimits: An Intuitive Approach
Limits: An Intuitive Approach SUGGESTED REFERENCE MATERIAL: As you work through the problems listed below, you should reference Chapter. of the recommended textbook (or the equivalent chapter in your alternative
More informationMiscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier
Miscellaneous Regarding reading materials Reading materials will be provided as needed If no assigned reading, it means I think the material from class is sufficient Should be enough for you to do your
More informationImplicit sampling for particle filters. Alexandre Chorin, Mathias Morzfeld, Xuemin Tu, Ethan Atkins
0/20 Implicit sampling for particle filters Alexandre Chorin, Mathias Morzfeld, Xuemin Tu, Ethan Atkins University of California at Berkeley 2/20 Example: Try to find people in a boat in the middle of
More informationCS281A/Stat241A Lecture 17
CS281A/Stat241A Lecture 17 p. 1/4 CS281A/Stat241A Lecture 17 Factor Analysis and State Space Models Peter Bartlett CS281A/Stat241A Lecture 17 p. 2/4 Key ideas of this lecture Factor Analysis. Recall: Gaussian
More information13. Nonlinear least squares
L. Vandenberghe ECE133A (Fall 2018) 13. Nonlinear least squares definition and examples derivatives and optimality condition Gauss Newton method Levenberg Marquardt method 13.1 Nonlinear least squares
More informationSolution to EE 617 Mid-Term Exam, Fall November 2, 2017
Solution to EE 67 Mid-erm Exam, Fall 207 November 2, 207 EE 67 Solution to Mid-erm Exam - Page 2 of 2 November 2, 207 (4 points) Convex sets (a) (2 points) Consider the set { } a R k p(0) =, p(t) for t
More information= 0 otherwise. Eu(n) = 0 and Eu(n)u(m) = δ n m
A-AE 567 Final Homework Spring 212 You will need Matlab and Simulink. You work must be neat and easy to read. Clearly, identify your answers in a box. You will loose points for poorly written work. You
More informationSimple Iteration, cont d
Jim Lambers MAT 772 Fall Semester 2010-11 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Simple Iteration, cont d In general, nonlinear equations cannot be solved in a finite sequence
More informationand lim lim 6. The Squeeze Theorem
Limits (day 3) Things we ll go over today 1. Limits of the form 0 0 (continued) 2. Limits of piecewise functions 3. Limits involving absolute values 4. Limits of compositions of functions 5. Limits similar
More informationOn Transmitter Design in Power Constrained LQG Control
On Transmitter Design in Power Constrained LQG Control Peter Breun and Wolfgang Utschic American Control Conference June 2008 c 2008 IEEE. Personal use of this material is permitted. However, permission
More informationTime and Event-based Sensor Scheduling for Networks with Limited Communication Resources
Proceedings of the 18th World Congress The International Federation of Automatic Control Time and Event-based Sensor Scheduling for Networks with Limited Communication Resources Ling Shi Karl Henrik Johansson
More informationHomework Solution # 3
ECSE 644 Optimal Control Feb, 4 Due: Feb 17, 4 (Tuesday) Homework Solution # 3 1 (5%) Consider the discrete nonlinear control system in Homework # For the optimal control and trajectory that you have found
More informationDual Proximal Gradient Method
Dual Proximal Gradient Method http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/19 1 proximal gradient method
More informationMATH 614 Dynamical Systems and Chaos Lecture 2: Periodic points. Hyperbolicity.
MATH 614 Dynamical Systems and Chaos Lecture 2: Periodic points. Hyperbolicity. Orbit Let f : X X be a map defining a discrete dynamical system. We use notation f n for the n-th iteration of f defined
More informationFrom Bayes to Extended Kalman Filter
From Bayes to Extended Kalman Filter Michal Reinštein Czech Technical University in Prague Faculty of Electrical Engineering, Department of Cybernetics Center for Machine Perception http://cmp.felk.cvut.cz/
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture State space models, 1st part: Model: Sec. 10.1 The
More informationThere are none. Abstract for Gauranteed Margins for LQG Regulators, John Doyle, 1978 [Doy78].
Chapter 7 Output Feedback There are none. Abstract for Gauranteed Margins for LQG Regulators, John Doyle, 1978 [Doy78]. In the last chapter we considered the use of state feedback to modify the dynamics
More informationEE363 homework 5 solutions
EE363 Prof. S. Boyd EE363 homework 5 solutions 1. One-step ahead prediction of an autoregressive time series. We consider the following autoregressive (AR) system p t+1 = αp t + βp t 1 + γp t 2 + w t,
More informationSYSTEMTEORI - KALMAN FILTER VS LQ CONTROL
SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL 1. Optimal regulator with noisy measurement Consider the following system: ẋ = Ax + Bu + w, x(0) = x 0 where w(t) is white noise with Ew(t) = 0, and x 0 is a stochastic
More informationA Crash Course on Kalman Filtering
A Crash Course on Kalman Filtering Dan Simon Cleveland State University Fall 2014 1 / 64 Outline Linear Systems Probability State Means and Covariances Least Squares Estimation The Kalman Filter Unknown
More informationDiscrete Math, Second Problem Set (June 24)
Discrete Math, Second Problem Set (June 24) REU 2003 Instructor: Laszlo Babai Scribe: D Jeremy Copeland 1 Number Theory Remark 11 For an arithmetic progression, a 0, a 1 = a 0 +d, a 2 = a 0 +2d, to have
More informationRobust Control 5 Nominal Controller Design Continued
Robust Control 5 Nominal Controller Design Continued Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University 4/14/2003 Outline he LQR Problem A Generalization to LQR Min-Max
More informationMath 489AB Exercises for Chapter 1 Fall Section 1.0
Math 489AB Exercises for Chapter 1 Fall 2008 Section 1.0 1.0.2 We want to maximize x T Ax subject to the condition x T x = 1. We use the method of Lagrange multipliers. Let f(x) = x T Ax and g(x) = x T
More informationKalman Filter. Man-Wai MAK
Kalman Filter Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: S. Gannot and A. Yeredor,
More informationMATH Topics in Applied Mathematics Lecture 2-6: Isomorphism. Linear independence (revisited).
MATH 311-504 Topics in Applied Mathematics Lecture 2-6: Isomorphism. Linear independence (revisited). Definition. A mapping f : V 1 V 2 is one-to-one if it maps different elements from V 1 to different
More informationarxiv: v1 [cs.sy] 30 Sep 2015
Optimal Sensor Scheduling and Remote Estimation over an Additive Noise Channel Xiaobin Gao, Emrah Akyol, and Tamer Başar arxiv:1510.00064v1 cs.sy 30 Sep 015 Abstract We consider a sensor scheduling and
More informationFormula Sheet for Optimal Control
Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming
More informationControlled Diffusions and Hamilton-Jacobi Bellman Equations
Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter
More informationQ-Learning and Stochastic Approximation
MS&E338 Reinforcement Learning Lecture 4-04.11.018 Q-Learning and Stochastic Approximation Lecturer: Ben Van Roy Scribe: Christopher Lazarus Javier Sagastuy In this lecture we study the convergence of
More informationSuppose that we have a specific single stage dynamic system governed by the following equation:
Dynamic Optimisation Discrete Dynamic Systems A single stage example Suppose that we have a specific single stage dynamic system governed by the following equation: x 1 = ax 0 + bu 0, x 0 = x i (1) where
More informationA Matrix Theoretic Derivation of the Kalman Filter
A Matrix Theoretic Derivation of the Kalman Filter 4 September 2008 Abstract This paper presents a matrix-theoretic derivation of the Kalman filter that is accessible to students with a strong grounding
More informationLinear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.
Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the
More informationMATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018
MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S
More informationMarkov Chain Monte Carlo Methods for Stochastic Optimization
Markov Chain Monte Carlo Methods for Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U of Toronto, MIE,
More informationSF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 10: Interior methods. Anders Forsgren. 1. Try to solve theory question 7.
SF2822 Applied Nonlinear Optimization Lecture 10: Interior methods Anders Forsgren SF2822 Applied Nonlinear Optimization, KTH 1 / 24 Lecture 10, 2017/2018 Preparatory question 1. Try to solve theory question
More informationStatistical Techniques in Robotics (16-831, F12) Lecture#17 (Wednesday October 31) Kalman Filters. Lecturer: Drew Bagnell Scribe:Greydon Foil 1
Statistical Techniques in Robotics (16-831, F12) Lecture#17 (Wednesday October 31) Kalman Filters Lecturer: Drew Bagnell Scribe:Greydon Foil 1 1 Gauss Markov Model Consider X 1, X 2,...X t, X t+1 to be
More informationParticle Filtering Approaches for Dynamic Stochastic Optimization
Particle Filtering Approaches for Dynamic Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge I-Sim Workshop,
More informationStable adaptive control with online learning
Stable adaptive control with online learning Andrew Y. Ng Stanford University Stanford, CA 9435, USA H. Jin Kim Seoul National University Seoul, Korea Abstract Learning algorithms have enjoyed numerous
More information