Observability and state estimation

Similar documents
Lecture 19 Observability and state estimation

Linear dynamical systems with inputs & outputs

EL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1)

Lecture 5 Least-squares

Observability. It was the property in Lyapunov stability which allowed us to resolve that

ME 234, Lyapunov and Riccati Problems. 1. This problem is to recall some facts and formulae you already know. e Aτ BB e A τ dτ

1 Continuous-time Systems

Control Systems Design, SC4026. SC4026 Fall 2009, dr. A. Abate, DCSC, TU Delft

14 - Gaussian Stochastic Processes

Linear algebra review

Control Systems Design, SC4026. SC4026 Fall 2010, dr. A. Abate, DCSC, TU Delft

Module 08 Observability and State Estimator Design of Dynamical LTI Systems

Dynamical systems: basic concepts

6.241 Dynamic Systems and Control

Linear System Theory

LMIs for Observability and Observer Design

16.31 Fall 2005 Lecture Presentation Mon 31-Oct-05 ver 1.1

Kalman Decomposition B 2. z = T 1 x, where C = ( C. z + u (7) T 1, and. where B = T, and

Intro. Computer Control Systems: F9

Intro. Computer Control Systems: F8

Controllability, Observability, Full State Feedback, Observer Based Control

Control Systems. Laplace domain analysis

Lecture 2. Linear Systems

Modern Control Systems

Raktim Bhattacharya. . AERO 632: Design of Advance Flight Control System. Norms for Signals and Systems

Grammians. Matthew M. Peet. Lecture 20: Grammians. Illinois Institute of Technology

Linear Systems. Linear systems?!? (Roughly) Systems which obey properties of superposition Input u(t) output

Control Systems. Frequency domain analysis. L. Lanari

Lecture 4 and 5 Controllability and Observability: Kalman decompositions

Control Systems I. Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback. Readings: Emilio Frazzoli

Lecture 5 Linear Quadratic Stochastic Control

Topic # Feedback Control. State-Space Systems Closed-loop control using estimators and regulators. Dynamics output feedback

ECEEN 5448 Fall 2011 Homework #4 Solutions

EE263: Introduction to Linear Dynamical Systems Review Session 6

Deterministic Dynamic Programming

Identification Methods for Structural Systems

Control Systems. Time response. L. Lanari

Chapter 3. LQ, LQG and Control System Design. Dutch Institute of Systems and Control

Nonlinear Control Systems

3 Gramians and Balanced Realizations

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science : Dynamic Systems Spring 2011

The norms can also be characterized in terms of Riccati inequalities.

Control Systems. Time response

Fall 線性系統 Linear Systems. Chapter 08 State Feedback & State Estimators (SISO) Feng-Li Lian. NTU-EE Sep07 Jan08

Module 03 Linear Systems Theory: Necessary Background

EE363 homework 7 solutions

CONTROL DESIGN FOR SET POINT TRACKING

Nonlinear Observers. Jaime A. Moreno. Eléctrica y Computación Instituto de Ingeniería Universidad Nacional Autónoma de México

EE363 Review Session 1: LQR, Controllability and Observability

SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL

DESIGN OF OBSERVERS FOR SYSTEMS WITH SLOW AND FAST MODES

Control Systems Design

CDS Solutions to the Midterm Exam

Observability and Constructability

Control Systems Design

to have roots with negative real parts, the necessary and sufficient conditions are that:

Solution via Laplace transform and matrix exponential

Observability Methods in Sensor Scheduling. Utku Ilkturk

Lecture 4 Continuous time linear quadratic regulator

Nonlinear Control Systems

5. Observer-based Controller Design

Dynamic interpretation of eigenvectors

EE363 homework 8 solutions

1 Some Facts on Symmetric Matrices

Min-Max Output Integral Sliding Mode Control for Multiplant Linear Uncertain Systems

Problem 1 Cost of an Infinite Horizon LQR

POLE PLACEMENT. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 19

MAE143 A - Signals and Systems - Winter 11 Midterm, February 2nd

ECEEN 5448 Fall 2011 Homework #5 Solutions

ECEN 605 LINEAR SYSTEMS. Lecture 8 Invariant Subspaces 1/26

6 OUTPUT FEEDBACK DESIGN

Linear Systems. Chapter Basic Definitions

ẋ n = f n (x 1,...,x n,u 1,...,u m ) (5) y 1 = g 1 (x 1,...,x n,u 1,...,u m ) (6) y p = g p (x 1,...,x n,u 1,...,u m ) (7)

A Tutorial on Recursive methods in Linear Least Squares Problems

Module 07 Controllability and Controller Design of Dynamical LTI Systems

Final Exam Solutions

2 Introduction of Discrete-Time Systems

Analog Signals and Systems and their properties

Controllability of Linear Systems with Input and State Constraints

x(t) = t[u(t 1) u(t 2)] + 1[u(t 2) u(t 3)]

State Variable Analysis of Linear Dynamical Systems

Introduction to Modern Control MT 2016

On Controllability of Linear Systems 1

EE221A Linear System Theory Final Exam

Prashant Mhaskar, Nael H. El-Farra & Panagiotis D. Christofides. Department of Chemical Engineering University of California, Los Angeles

Stochastic and Adaptive Optimal Control

Lecture 2 and 3: Controllability of DT-LTI systems

EE Control Systems LECTURE 9

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Lecture 6. Regularized least-squares and minimum-norm methods 6 1

Approximation of the Linearized Boussinesq Equations

Steady State Kalman Filter

ECEN 605 LINEAR SYSTEMS. Lecture 7 Solution of State Equations 1/77

Balanced Truncation 1

Controllability. Chapter Reachable States. This chapter develops the fundamental results about controllability and pole assignment.

Time Series Analysis

Pole placement control: state space and polynomial approaches Lecture 2

There are none. Abstract for Gauranteed Margins for LQG Regulators, John Doyle, 1978 [Doy78].

Multivariable Control. Lecture 03. Description of Linear Time Invariant Systems. John T. Wen. September 7, 2006

Krylov Subspace Methods for Nonlinear Model Reduction

Transcription:

EE263 Autumn 2015 S Boyd and S Lall Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time observability least-squares observers example 1

State estimation set up we consider the discrete-time system x(t + 1) = Ax(t) + Bu(t) + w(t), y(t) = Cx(t) + Du(t) + v(t) w is state disturbance or noise v is sensor noise or error A, B, C, and D are known u and y are observed over time interval [0, t 1] w and v are not known, but can be described statistically, or assumed small (eg, in RMS value) 2

State estimation problem state estimation problem: estimate x(s) from s = 0: estimate initial state s = t 1: estimate current state u(0),, u(t 1), y(0),, y(t 1) s = t: estimate (ie, predict) next state an algorithm or system that yields an estimate ˆx(s) is called an observer or state estimator ˆx(s) is denoted ˆx(s t 1) to show what information estimate is based on (read, ˆx(s) given t 1 ) 3

Noiseless case let s look at finding x(0), with no state or measurement noise: x(t + 1) = Ax(t) + Bu(t), y(t) = Cx(t) + Du(t) with x(t) R n, u(t) R m, y(t) R p then we have where O t = C CA CA t 1 y(0) y(t 1) = O tx(0) + T t, Tt = u(0) u(t 1) D 0 CB D 0 CA t 2 B CA t 3 B CB D O t maps initials state into resulting output over [0, t 1] T t maps input to output over [0, t 1] 4

hence we have y(0) O tx(0) = y(t 1) T t u(0) u(t 1) RHS is known, x(0) is to be determined 5

hence: can uniquely determine x(0) if and only if null(o t) = {0} null(o t) gives ambiguity in determining x(0) if x(0) null(o t) and u = 0, output is zero over interval [0, t 1] input u does not affect ability to determine x(0); its effect can be subtracted out 6

Observability matrix by C-H theorem, each A k is linear combination of A 0,, A n 1 hence for t n, null(o t) = null(o) where is called the observability matrix O = O n = C CA CA n 1 if x(0) can be deduced from u and y over [0, t 1] for any t, then x(0) can be deduced from u and y over [0, n 1] null(o) is called unobservable subspace; describes ambiguity in determining state from input and output system is called observable if null(o) = {0}, ie, Rank(O) = n 7

Observability controllability duality let (Ã, B, C, D) be dual of system (A, B, C, D), ie, controllability matrix of dual system is à = A T, B = C T, C = B T, D = D T C = [ B à B Ãn 1 B] = [C T A T C T (A T ) n 1 C T ] = O T, transpose of observability matrix similarly we have Õ = CT thus, system is observable (controllable) if and only if dual system is controllable (observable) in fact, null(o) = range(o T ) = range( C) ie, unobservable subspace is orthogonal complement of controllable subspace of dual 8

Observers for noiseless case suppose Rank(O t) = n (ie, system is observable) and let F be any left inverse of O t, ie, F O t = I then we have the observer y(0) u(0) x(0) = F y(t 1) T t u(t 1) which deduces x(0) (exactly) from u, y over [0, t 1] in fact we have y(τ t + 1) u(τ t + 1) x(τ t + 1) = F y(τ) T t u(τ) ie, our observer estimates what state was t 1 epochs ago, given past t 1 inputs & outputs observer is (multi-input, multi-output) finite impulse response (FIR) filter, with inputs u and y, and output ˆx 9

Invariance of unobservable set fact: the unobservable subspace null(o) is invariant, ie, if z null(o), then Az null(o) proof: suppose z null(o), ie, CA k z = 0 for k = 0,, n 1 evidently CA k (Az) = 0 for k = 0,, n 2; (by C-H) where n 1 CA n 1 (Az) = CA n z = α ica i z = 0 i=0 det(si A) = s n + α n 1s n 1 + + α 0 10

Continuous-time observability continuous-time system with no sensor or state noise: ẋ = Ax + Bu, y = Cx + Du can we deduce state x from u and y? let s look at derivatives of y: y = Cx + Du ẏ = Cẋ + D u = CAx + CBu + D u ÿ = CA 2 x + CABu + CB u + Dü and so on 11

Continuous-time observability hence we have y ẏ y (n 1) where O is the observability matrix and T = = Ox + T u u u (n 1) D 0 CB D 0 CA n 2 B CA n 3 B CB D (same matrices we encountered in discrete-time case!) 12

rewrite as Ox = y ẏ y (n 1) T u u u (n 1) RHS is known; x is to be determined hence if null(o) = {0} we can deduce x(t) from derivatives of u(t), y(t) up to order n 1 in this case we say system is observable 13

can construct an observer using any left inverse F of O: y u ẏ u x = F y (n 1) T u (n 1) reconstructs x(t) (exactly and instantaneously) from u(t),, u (n 1) (t), y(t),, y (n 1) (t) derivative-based state reconstruction is dual of state transfer using impulsive inputs 14

A converse suppose z null(o) (the unobservable subspace), and u is any input, with x, y the corresponding state and output, ie, ẋ = Ax + Bu, y = Cx + Du then state trajectory x = x + e ta z satisfies x = A x + Bu, y = C x + Du ie, input/output signals u, y consistent with both state trajectories x, x hence if system is unobservable, no signal processing of any kind applied to u and y can deduce x unobservable subspace null(o) gives fundamental ambiguity in deducing x from u, y 15

Least-squares observers discrete-time system, with sensor noise: x(t + 1) = Ax(t) + Bu(t), y(t) = Cx(t) + Du(t) + v(t) we assume Rank(O t) = n (hence, system is observable) least-squares observer uses pseudo-inverse: y(0) u(0) ˆx(0) = O t y(t 1) T t u(t 1) where O t = ( O T t O t ) 1 O T t 16

interpretation: ˆx ls (0) minimizes discrepancy between output ŷ that would be observed, with input u and initial state x(0) (and no sensor noise), and output y that was observed, t 1 measured as ŷ(τ) y(τ) 2 τ=0 can express least-squares initial state estimate as ˆx ls (0) = ( t 1 ) 1 t 1 (A T ) τ C T CA τ (A T ) τ C T ỹ(τ) τ=0 where ỹ is observed output with portion due to input subtracted: ỹ = y h u where h is impulse response τ=0 17

Least-squares observer uncertainty ellipsoid since O t O t = I, we have x(0) = ˆx ls (0) x(0) = O t where x(0) is the estimation error of the initial state in particular, ˆx ls (0) = x(0) if sensor noise is zero (ie, observer recovers exact state in noiseless case) v(0) v(t 1) now assume sensor noise is unknown, but has RMS value α, 1 t 1 v(τ) 2 α 2 t τ=0 18

set of possible estimation errors is ellipsoid v(0) x(0) E unc = O t v(t 1) 1 t 1 t τ=0 v(τ) 2 α 2 E unc is uncertainty ellipsoid for x(0) (least-square gives best E unc) shape of uncertainty ellipsoid determined by matrix ( ) t 1 ) 1 1 (Ot T O t = (A T ) τ C T CA τ τ=0 maximum norm of error is ˆx ls (0) x(0) α t O t 19

Infinite horizon uncertainty ellipsoid the matrix P = lim t ( t 1 ) 1 (A T ) τ C T CA τ τ=0 always exists, and gives the limiting uncertainty in estimating x(0) from u, y over longer and longer periods: if A is stable, P > 0 ie, can t estimate initial state perfectly even with infinite number of measurements u(t), y(t), t = 0, (since memory of x(0) fades ) if A is not stable, then P can have nonzero nullspace ie, initial state estimation error gets arbitrarily small (at least in some directions) as more and more of signals u and y are observed 20

Example particle in R 2 moves with uniform velocity (linear, noisy) range measurements from directions 15, 0, 20, 30, once per second range noises IID N (0, 1); can assume RMS value of v is not much more than 2 no assumptions about initial position & velocity particle range sensors problem: estimate initial position & velocity from range measurements 21

express as linear system 1 0 1 0 x(t + 1) = 0 1 0 1 0 0 1 0 x(t), y(t) = 0 0 0 1 (x 1(t), x 2(t)) is position of particle (x 3(t), x 4(t)) is velocity of particle can assume RMS value of v is around 2 k i is unit vector from sensor i to origin k T 1 k T 4 true initial position & velocities: x(0) = (1 3 004 003) x(t) + v(t) 22

range measurements (& noiseless versions): measurements from sensors 1 4 5 4 3 2 1 range 0 1 2 3 4 5 0 20 40 60 80 100 120 t 23

estimate based on (y(0),, y(t)) is ˆx(0 t) actual RMS position error is (ˆx1(0 t) x 1(0)) 2 + (ˆx 2(0 t) x 2(0)) 2 (similarly for actual RMS velocity error) RMS position error 15 1 05 0 10 20 30 40 50 60 70 80 90 100 110 120 RMS velocity error 1 08 06 04 02 0 10 20 30 40 50 60 70 80 90 100 110 120 t 24

Continuous-time least-squares state estimation assume ẋ = Ax + Bu, y = Cx + Du + v is observable least-squares estimate of initial state x(0), given u(τ), y(τ), 0 τ t: choose ˆx ls (0) to minimize integral square residual J = t 0 ỹ(τ) Ce τa x(0) where ỹ = y h u is observed output minus part due to input let s expand as J = x(0) T Qx(0) + 2r T x(0) + s, Q = t 0 e τat C T Ce τa dτ, r = s = t 0 t 0 ỹ(τ) T ỹ(τ) dτ 2 dτ e τat C T ỹ(τ) dτ, 25

setting x(0) J to zero, we obtain the least-squares observer ( t ˆx ls (0) = Q 1 r = estimation error is ( t x(0) = ˆx ls (0) x(0) = therefore if v = 0 then ˆx ls (0) = x(0) 0 ) 1 t e τat C T Ce τa dτ e ATτ C T ỹ(τ) dτ 0 0 ) 1 t e τat C T Ce τa dτ e τat C T v(τ) dτ 0 26