Subspace Identification

Similar documents
RECURSIVE SUBSPACE IDENTIFICATION IN THE LEAST SQUARES FRAMEWORK

Subspace Identification Methods

Identification of MIMO linear models: introduction to subspace methods

Closed and Open Loop Subspace System Identification of the Kalman Filter

ELEC system identification workshop. Subspace methods

Adaptive Channel Modeling for MIMO Wireless Communications

Properties of Zero-Free Spectral Matrices Brian D. O. Anderson, Life Fellow, IEEE, and Manfred Deistler, Fellow, IEEE

EE731 Lecture Notes: Matrix Computations for Signal Processing

Subspace Identification A Markov Parameter Approach

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

DS-GA 1002 Lecture notes 10 November 23, Linear models

Lecture 4 and 5 Controllability and Observability: Kalman decompositions

Sparsity in system identification and data-driven control

Preliminary Examination in Numerical Analysis

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

Empirical Gramians and Balanced Truncation for Model Reduction of Nonlinear Systems

4 Derivations of the Discrete-Time Kalman Filter

SUBSPACE SYSTEM IDENTIFICATION Theory and applications

Notes on Time Series Modeling

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Lessons in Estimation Theory for Signal Processing, Communications, and Control

The Essentials of Linear State-Space Systems

Review of some mathematical tools

CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM

14 - Gaussian Stochastic Processes

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Linear Regression and Its Applications

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Cross Directional Control

STAT 100C: Linear models

Linear Algebra and Robot Modeling

arxiv: v1 [math.oc] 11 Jun 2009

CME 345: MODEL REDUCTION

ENGG5781 Matrix Analysis and Computations Lecture 8: QR Decomposition

Observability and Identifiability of Jump Linear Systems

14 Singular Value Decomposition

sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU)

15 Singular Value Decomposition

MATHEMATICS FOR COMPUTER VISION WEEK 2 LINEAR SYSTEMS. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

Manning & Schuetze, FSNLP, (c)

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St

Stochastic Realization of Binary Exchangeable Processes

SIMON FRASER UNIVERSITY School of Engineering Science

Learning a Linear Dynamical System Model for Spatiotemporal

LECTURE 7. Least Squares and Variants. Optimization Models EE 127 / EE 227AT. Outline. Least Squares. Notes. Notes. Notes. Notes.

A New Subspace Identification Method for Open and Closed Loop Data

CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Manning & Schuetze, FSNLP (c) 1999,2000

Gaussian Filtering Strategies for Nonlinear Systems

Lecture 9: Numerical Linear Algebra Primer (February 11st)

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

The following definition is fundamental.

6.241 Dynamic Systems and Control

DS-GA 1002 Lecture notes 12 Fall Linear regression

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides

Full-State Feedback Design for a Multi-Input System

X t = a t + r t, (7.1)

Chapter 3. Tohru Katayama

Continuum Limit of Forward Kolmogorov Equation Friday, March 06, :04 PM

Statistical Geometry Processing Winter Semester 2011/2012

October 25, 2013 INNER PRODUCT SPACES

Nonlinear Observers. Jaime A. Moreno. Eléctrica y Computación Instituto de Ingeniería Universidad Nacional Autónoma de México

Time-Varying Systems and Computations Lecture 3

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

Principal Component Analysis

Background on Chevalley Groups Constructed from a Root System

26 : Spectral GMs. Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G.

Designing Information Devices and Systems I Spring 2018 Lecture Notes Note Introduction to Linear Algebra the EECS Way

Linear System Theory

FIR Filters for Stationary State Space Signal Models

VISA: Versatile Impulse Structure Approximation for Time-Domain Linear Macromodeling

Normed & Inner Product Vector Spaces

ADAPTIVE FILTER THEORY

RECURSIVE ESTIMATION AND KALMAN FILTERING

7. Symmetric Matrices and Quadratic Forms

Markov decision processes and interval Markov chains: exploiting the connection

Balanced Truncation 1

A subspace fitting method based on finite elements for identification and localization of damages in mechanical systems

7 Principal Component Analysis

DEVELOPMENT AND USE OF OPERATIONAL MODAL ANALYSIS

Linear Systems. Linear systems?!? (Roughly) Systems which obey properties of superposition Input u(t) output

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

Working with Block Structured Matrices

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved

VARIANCE COMPUTATION OF MODAL PARAMETER ES- TIMATES FROM UPC SUBSPACE IDENTIFICATION

Vector Spaces, Orthogonality, and Linear Least Squares

Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models

L26: Advanced dimensionality reduction

Robustness of Principal Components

Properties of Matrices and Operations on Matrices

Designing Information Devices and Systems I Fall 2018 Lecture Notes Note Introduction to Linear Algebra the EECS Way

MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2

HST.582J/6.555J/16.456J

Learning the Linear Dynamical System with ASOS ( Approximated Second-Order Statistics )

Computational Techniques Prof. Sreenivas Jayanthi. Department of Chemical Engineering Indian institute of Technology, Madras

Latent semantic indexing

Transcription:

Chapter 10 Subspace Identification Given observations of m 1 input signals, and p 1 signals resulting from those when fed into a dynamical system under study, can we estimate the internal dynamics regulating that system? Subspace techniques encode the notion of the state as bottleneck between past and future using a series of geometric operations on input-output sequences They should be contrasted to the ideas governing PEM approaches as described in Chapter 5 Subspace algorithms make extensive use of the observability and controllability matrices and of their structure 101 Deterministic Subspace Identification The class of subspace algorithms found their root in the methods for converting known sequences of impulse response matrices (IRs), Transfer Function matrices (TFs) or covariance matrices into a corresponding state-space system Since those state-spaces satisfy the characteristics as encoded in IRs, TFs or covariance matrices, they are referred to as realizations (not to be confused with Given: An input signal {u t } n t=1 R m and output signal {y t } n t=1 R p, both of length n, and satisfying an (unkown) deterministic state-space of order d, or ( x t+1 = Ax t + Bu t (101) y = Cx t + Du t where t =1,,n and x 0 is fixed Problem: Recover (a) The order d of the unknown system (b) The unknown system matrices (A, B, C, D) Figure 101: The problem a deterministic Subspace Identification algorithms aims to solve 1

101 DETERMINISTIC SUBSPACE IDENTIFICATION realizations of random variables as described in Chapter 5) The naming convention originates from times where people sought quite intensively to physical realizations of mathematically described electrical systems 1011 Deterministic Realization: IRSS Given a sequence of IR matrices {G } 0 with P 0 kgk < 1, consider the problem of finding a realization S =(A, B, C, D) Since we know that G 0 = D, we make life easier by focussing on the problem of recovering the matrices (A, B, C) from given sequence of IR matrices {G } 1 Let us start by defining the Block-Hankel matrix H d 0 R d0 p d 0m of degree d 0 which will play a central role in the derivation: G 1 G G G d 0 G G G d 0 +1 H d 0 = G Gd0 + R d0 p d 0m (10) 5 G d 0 G d 0 1 Then the input-signal y and its corresponding output-signal u are related as follows Let u t = (u t, u t 1, u t,) T R 1 be the reverse of the input-signal, and let y + t =(y t, y t+1, y t+,) T R 1 be the forward output signal taking values from instant t on Then we can write y t + = y t y t+1 y t+ G 1 G G 5 = H G G 1u t = G 5 u t u t 1 u t 5 (10) where the limit is taken for d 0!1(represented by the open ended in the matrix) Hence the interpretation of the matrix H 1 is that when presented the system with an input which drops to 0 when going beyond t, then the block-hankel matrix H 1 computes the output of the system after t That is, it gives the system output when the system is in free mode Lemma 10 (Factorization) Given a State-space system S =(A, B, C) (with D =0), then H d = O d TT 1 C d, (10) for any nonsingular transformation T R d d Moreover, we have the following rank conditions rank(h d ) apple max(rank(o d ),rank(c d )) apple d (105) Lemma 11 (Minimal Realization) Given a State-space system S =(A, B, C) (with D =0) with minimal state-dimension d, then we have for all d 0 d that rank(h d 0)=d (10) 18

101 DETERMINISTIC SUBSPACE IDENTIFICATION This result is directly seen by construction However the implications are important: it says that the rank of the block-hankel matrix is equal to the minimal dimension of the states in a state-space model which can exhibit the given behavior That means that in using this factorization, we are not only given the system-matrices, but as well the minimal dimension Recall that the problem of order estimation in PEM methods is often considered as a separate model selection problem, and required in a senses the use of ad-hoc tools The idea of the Ho-Kalman realization algorithm is then that in case the block-hankel matrix H d 0 can be formed for large enough d 0, the Singular Value Decomposition of this matrix let us recover the observability matrix O d 0 and the controllability matrix C d 0 (up to a transformation T) These can in turn be used to extract the system matrices A, B, C (up to a transformation T) In order to perform this last step, we need the following straightforward idea: Proposition (Recovering System Matrices) The observability matrix O d satisfies the following recursive relations C CA apple CA O d = CA d 1 = 5 The controlability matrix C d satisfies the following recursive relations C d = B AB A B A d 1 B C O d 1 A (10) apple B = (108) AC d 1 5 That means that once the matrix C d and O d are known, the system matrices (A, B, C) can be recovered straightforwardly by selecting appropriate parts Optimality of the SVD then ensures that we will recover a minimal state space realizing the sequence of IR matrices {G } >0 Infull, the algorithm becomes as follows Algorithm 1 (Ho-Kalman) Given d 0 up to within a similarity transform T d and the IRs {G } d0 =0, findarealizations =(A, B, C, D) 1 Set D = G 0 Decompose Find A Find B and C 19

101 DETERMINISTIC SUBSPACE IDENTIFICATION 101 NSID Now we turn to the question how one may use the ideas of the Kalman-Ho algorithm in order to find a realization directly from input-output data, rather than from given IRs Again, this can be done by performing a sequence of projections under the assumption that the given data obeys a state-space model with unknown (but fixed) system matrices S =(A, B, C, D) This technique is quite di erent from the PEM approach were the estimation problem is turned into an optimization problem To make the di erence explicit, we refer to the projection-based algorithms as Subspace Identification (SID) algorithms The first studied SID algorithm was (arguably) NSID (niftily pronounced as a californian plate as enforce it ) The abbreviation stands for Numerical algorithm For Subspace IDentification The central insight is the following expression of the future output of the system in terms of (i) the unknown states at that time, and (ii) the future input signals Formally, we define the matrices representing past as u 1 u d 0 y 1 y d 0 u u d0 +1 y y d0 +1 U p = u t+1 u t+d 0, Y p = y t+1 y t+d 0, (109) 5 5 u n d0 +1 u n d 0 y n d0 +1 y n d 0 and equivalently we define the matrices representing future as u d0 +1 u d 0 y d0 +1 y d 0 u d0 + u d0 +1 y d0 + y d0 +1 U f = u t+1 u t+d 0, Y f = y t+1 y t+d 0 (1010) 5 5 u n d0 +1 u n y n d0 +1 y n Now we are all set to give the main factorization from which the NSID algorithm will follow Lemma 1 (NSID) Assuming that a system S =(A, B, C, D) underlying the signals exists, than there exist matrices F R md0 pd 0 and F R md0 pd 0 such that Y f = X f O d0 T + U f G d0 T (1011) = U p F + Y p F 0 + U f G d0 T (101) This really follows from working out the terms, schematically (again empty entries denote blocks 150

101 DETERMINISTIC SUBSPACE IDENTIFICATION of zeros): y d 0 +1 y d 0 y d 0 + y d 0 +1 Y f = y t+1 y t+d 0 5 y n d 0 +1 y n x d 0 +1 u d 0 +1 u d 0 x d0 + T C u d0 + u d0 +1 G 0 G 1 G d 0 1 = CA x t+1 5 + G 0 u t+1 u t+d 0 5 5 CA d0 1 5 G 0 x n d0 +1 u n d0 +1 u n = X d0 f O d0 T + U f G d0 T (101) We see that the sequence of states {x} t d 0 can be written as a (finite) linear combination of the matrices U p and Y f Hereto, we interprete a similar linear relation y 1 y d 0 y y d 0 +1 Y p = y t+1 y t+d 0 5 y n d 0 +1 y n d 0 x 1 u 1 u d 0 x T C u u d0 +1 G 0 G 1 G d 0 1 = CA x t+1 5 + G 0 u t+1 u t+d 0 5 5 CA d0 1 5 G 0 x n d0 +1 u n d 0 +1 u n d 0 = X d0 p O d0 T + U p G d0 T (101) Note that this step needs some care to let the indices match properly, ie we can only use the signals between iteration d 0,,n d 0 As such the factorization follows readily This factorization gives us an equation connecting the past and future Moreover, note that eq (101) looks like a model which is linear in the unknown F, F 0 and G d0, and we know by now how to solve this one since we can construct the matrices Y f and U p, Y p, U f Indeed, using an OLS estimator - or equivalent the orthogonal projection - lets us recover the unknown matrices, that is, under appropriate rank conditions in order to guarantee that only a single solution exists We 151

10 STOCHASTIC SUBSPACE TECHNIQUES proceed however by defining a shortcut to solving this problem in order to arrive at less expensive implementations What is left to us is to decompose the term O = X T d0 f Od0 in the state variables and the observability matrix This however, we learned to do from Kalman-Ho s algorithm (previous Subsection) Indeed, an SVD decomposition of the matrix O gives us the rank d of the observability matrix, together with a reduced rank decomposition from which {x t } and O d can be recovered (up to a similarity transform T) Now it is not too di cult to retrieve the system matrices S =(A, B, C, D) A more robust way goes as follows: given the input and output signals, as well as the recovered states, we know that those satisfy the state-space system and as such Z f, x y 1 x y x t+1 x n y n 1 y t x 1 u 1 x u apple apple = A C A C x t u t B C, Z p B C, (1015) 5 5 x n 1 u n 1 from which one can recover directly the matrices (A, B, C, D), for instance by solving a LS problem or an Orthogonal projection as Z pz f The NSID algorithm is summarized as follows: Algorithm (NSID) Given d 0 d and the signals u R pn and y R mn,findarealization S =(A, B, C, D) up to within a similarity transform T 1 Find O d Find the state sequence X Find A, B, C, D 101 Variations on the Theme Since its inception, almost decades of research yielded many important improvements of the algorithm Here we will review some of those, starting with a description of a family of related methods going under the name of MOESP subspace identification algorithms MOESP Intersection Projection 10 Stochastic Subspace Techniques 101 Stochastic Realization: CovSS Given a sequence of covariance matrices { = E[Y t Y t+ ]} 0 with P 0 k k < 1, can we find a stochastic MIMO system S =(A, C) realizing those covariance matrices? More specifically, if the realized system S =(A, C) were driven by zero-mean colored Gaussian noise {W t } t and {V t } t, 15

10 STOCHASTIC SUBSPACE TECHNIQUES Given: An input signal {y t } n t=1 R p of length n, and satisfying an (unkown) deterministic state-space of order d, or ( x t+1 = Ax t + v t (101) y = Cx t + w t where t =1,,n and x 0 is fixed Problem: Recover (a) The order d of the unknown system (b) The unknown system matrices (A, C) Figure 10: The problem a stochastic Subspace Identification algorithms aims to solve the covariance matrices of the outputs are to match { = E[Y t Y t+ ]} 0 Ideas go along the line as set out in the previous subsection Such series of covariance matrices might be given equivalently as a spectral density function (z) Here we use the following relation generalizing the z-transform and its inverse (z) = = Z 1 1X = 1 1 z, (101) (z)z dz (1018) The key once more is to consider the stochastic processes representing past and future Specifically, given a process Y =(,Y t 1,Y t,y t+1,) T taking values as an infinite vector, define Y t + and Y t as follows ( Y t + =(Y t,y t+1,y t+ ) T Y t =(Y t,y t 1,Y t,) T (1019), both taking values as one-sided infinite vectors The next idea is to build up the covariance matrices associated to those stochastic processes The devil is in the details here! The following (infinite dimensional) matrices are block Toeplitz: 0 T 1 T L + = E Y t + Y t +T 1 0 T 1 = 1 (100) 5 and L = E Y t Yt T = 0 1 T 1 0 1 T T 1 5 15 (101)

10 FURTHER WORK ON SUBSPACE IDENTIFICATION The (infinite dimensional) block Hankel matrix H is here defined as 1 H = E Y t Y t +T = E Y + t Yt T = 5 (10) Now again we can factorize the matrix H into an observability part, and a controllability part That is define the infinite observability matrix corresponding with a stochastic MIMO system S =(A, C) as C CA O = CA 5, (10) and let the infinite controllability matrix of S =(A, C) be given as Then we have again C = C AC A C (10) Lemma 1 (Factorisation) Let H, C, O be infinite matrices as before associated with a stochastic MIMO system S =(A, B) If the realization were driven by white noise we have for all 1 that ( 0 =0 = E[Y t Y t+ ]= CA 1 C (105) 1 Moreover H = OC (10) 10 Further Work on Subspace Identification 10 Implementations of Subspace Identification 15

Chapter 11 Design of Experiments How to set up a data-collection experiment so as too guarantee accurately identified models? 155

15