CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Prediction Error Methods - Torsten Söderström

Similar documents
Expressions for the covariance matrix of covariance data

EL1820 Modeling of Dynamical Systems

On Identification of Cascade Systems 1

12. Prediction Error Methods (PEM)

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Recursive Algorithms - Han-Fu Chen

EECE Adaptive Control

On Input Design for System Identification

EECE Adaptive Control

6.435, System Identification

Sign-Perturbed Sums (SPS): A Method for Constructing Exact Finite-Sample Confidence Regions for General Linear Systems

IDENTIFICATION OF A TWO-INPUT SYSTEM: VARIANCE ANALYSIS

Order Selection for Vector Autoregressive Models

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION - Vol. XII - Lyapunov Stability - Hassan K. Khalil

Elements of Multivariate Time Series Analysis

Further Results on Model Structure Validation for Closed Loop System Identification

Improving performance and stability of MRI methods in closed-loop

Refined Instrumental Variable Methods for Identifying Hammerstein Models Operating in Closed Loop

OPTIMAL EXPERIMENT DESIGN IN CLOSED LOOP. KTH, Signals, Sensors and Systems, S Stockholm, Sweden.

A unified framework for EIV identification methods in the presence of mutually correlated noises

Accuracy Analysis of Time-domain Maximum Likelihood Method and Sample Maximum Likelihood Method for Errors-in-Variables Identification

Model Identification and Validation for a Heating System using MATLAB System Identification Toolbox

A Mathematica Toolbox for Signals, Models and Identification

Exploring Granger Causality for Time series via Wald Test on Estimated Models with Guaranteed Stability

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.

A New Subspace Identification Method for Open and Closed Loop Data

REGLERTEKNIK AUTOMATIC CONTROL LINKÖPING

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION - Vol. XIII - Nonlinear Observers - A. J. Krener

RECURSIVE SUBSPACE IDENTIFICATION IN THE LEAST SQUARES FRAMEWORK

Digital Signal Processing

Model structure. Lecture Note #3 (Chap.6) Identification of time series model. ARMAX Models and Difference Equations

What can regularization offer for estimation of dynamical systems?

Matlab software tools for model identification and data analysis 11/12/2015 Prof. Marcello Farina

LTI Approximations of Slightly Nonlinear Systems: Some Intriguing Examples

Time Series: Theory and Methods

EL1820 Modeling of Dynamical Systems

Curriculum Vitae Wenxiao Zhao

Advanced Process Control Tutorial Problem Set 2 Development of Control Relevant Models through System Identification

Cramér-Rao Bounds for Estimation of Linear System Noise Covariances

Errors-in-variables identification through covariance matching: Analysis of a colored measurement noise case

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Frequency Domain System Identification - J. Schoukens and R. Pintelon

Identification of a Chemical Process for Fault Detection Application

Continuous-Time Model Identification and State Estimation Using Non-Uniformly Sampled Data

14 th IFAC Symposium on System Identification, Newcastle, Australia, 2006

NON-LINEAR CONTROL OF OUTPUT PROBABILITY DENSITY FUNCTION FOR LINEAR ARMAX SYSTEMS

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION - Vol. VII - System Characteristics: Stability, Controllability, Observability - Jerzy Klamka

PERFORMANCE ANALYSIS OF CLOSED LOOP SYSTEM WITH A TAILOR MADE PARAMETERIZATION. Jianhong Wang, Hong Jiang and Yonghong Zhu

COMPLEX SIGNALS are used in various areas of signal

PARAMETER ESTIMATION AND ORDER SELECTION FOR LINEAR REGRESSION PROBLEMS. Yngve Selén and Erik G. Larsson

An Exponentially Weighted Moving Average Method for Identification and Monitoring of Stochastic Systems

sine wave fit algorithm

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

On the convergence of the iterative solution of the likelihood equations

On Moving Average Parameter Estimation

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

Dynamic Time Series Regression: A Panacea for Spurious Correlations

DISCRETE-TIME ADAPTIVE LQG/LTR CONTROL

Outline 2(42) Sysid Course VT An Overview. Data from Gripen 4(42) An Introductory Example 2,530 3(42)

Short-Term Load Forecasting Using ARIMA Model For Karnataka State Electrical Load

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

Continuous-Time Model Identification and State Estimation Using Non-Uniformly Sampled Data

Estimation of white-box model parameters via artificial data generation: a two-stage approach

Estimation of electromechanical modes in power systems using system identification techniques

NEW RESULTS ON THE IDENTIFICATION OF INTERVAL PREDICTOR MODELS

Sound Listener s perception

Matlab software tools for model identification and data analysis 10/11/2017 Prof. Marcello Farina

Bounding the parameters of linear systems with stability constraints

EECE Adaptive Control

Finite-time experiment design with multisines

f-domain expression for the limit model Combine: 5.12 Approximate Modelling What can be said about H(q, θ) G(q, θ ) H(q, θ ) with

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION Vol. VI - Identification of NARMAX and Related Models - Stephen A. Billings and Daniel Coca

ADAPTIVE INVERSE CONTROL BASED ON NONLINEAR ADAPTIVE FILTERING. Information Systems Lab., EE Dep., Stanford University

UTILIZING PRIOR KNOWLEDGE IN ROBUST OPTIMAL EXPERIMENT DESIGN. EE & CS, The University of Newcastle, Australia EE, Technion, Israel.

Computer Exercise 1 Estimation and Model Validation

Published in: Proceedings of the 16th IFAC Symposium on System Identification (SYSID 2012), July , Brussels, Belgium

THIS paper deals with robust control in the setup associated

Optimal Polynomial Control for Discrete-Time Systems

THERE are two types of configurations [1] in the

Identification of ARX, OE, FIR models with the least squares method

Control Systems Lab - SC4070 System Identification and Linearization

Closed-Loop Identification using Routine Operating Data: the Effect of Time Delay

MOOSE: Model Based Optimal Input Design Toolbox for Matlab

System Modeling and Identification CHBE 702 Korea University Prof. Dae Ryook Yang

CONTROLLABILITY AND OBSERVABILITY OF 2-D SYSTEMS. Klamka J. Institute of Automatic Control, Technical University, Gliwice, Poland

LTI Systems, Additive Noise, and Order Estimation

REGLERTEKNIK AUTOMATIC CONTROL LINKÖPING

Comparison of Prediction-Error-Modelling Criteria

Closed-loop Identification of Hammerstein Systems Using Iterative Instrumental Variables

Virtual Reference Feedback Tuning for non-linear systems

Identification of LPV Output-Error and Box-Jenkins Models via Optimal Refined Instrumental Variable Methods

Ross Bettinger, Analytical Consultant, Seattle, WA

Nonlinear Identification of Backlash in Robot Transmissions

Virtual Reference Feedback Tuning for Non Minimum Phase Plants

Using Neural Networks for Identification and Control of Systems

Title without the persistently exciting c. works must be obtained from the IEE

basis of dierent closed-loop transfer functions. Here the multivariable situation is considered

Estimating Moving Average Processes with an improved version of Durbin s Method

Study of Time Series and Development of System Identification Model for Agarwada Raingauge Station

Maximum Likelihood Modeling Of Orbits Of Nonlinear ODEs

L 2 Model Reduction and Variance Reduction

Prediction-based adaptive control of a class of discrete-time nonlinear systems with nonlinear growth rate

Transcription:

PREDICTIO ERROR METHODS Torsten Söderström Department of Systems and Control, Information Technology, Uppsala University, Uppsala, Sweden Keywords: prediction error method, optimal prediction, identifiability, ARMAX model, consistency Contents. Description.. Introduction.2. General Linear Dynamic Models.2.. Introduction.2.2. ARMAX Models.2.3. State Space Models.3. Optimal Prediction.4. Interpretations.5. Implementation Aspects.5.. Optimization.5.2. Evaluation of Gradients.6. Extensions.6.. Prefiltering of Data.6.2. Modified Criterion Function.6.3. Using Multistep Prediction Errors 2. Properties 2.. Identifiability 2.2. Convergence and Consistency 2.3. Asymptotic Accuracy and Distribution 2.4. Model Approximation Acknowledgement Glossary Bibliography Biographical Sketch Summary A general linear stochastic model is introduced. It will be described how it covers various typical special cases, like ARMAX models and other black box input-output model, state space models that can be parameterized either by black box considerations or by physical insight. The general linear model can be used also to discuss identifiability. Prediction error methods are introduced as a general methodology for estimating the parameters in a general linear model. The parameter estimates are obtained by minimization of the

sample prediction error variance. The estimates are consistent and have a low asymptotic covariance matrix under weak conditions.. Description.. Introduction In this subsection, the class of prediction error methods (PEM s) will be described. The description is confined to the off-line (or batch case), where the parameter vector in a general linear model of the following form is estimated: M ( θ): yt ( ) = Gq ( ; θ) ut ( ) + H( q ; θ) et ( ) T Ee() t e () s =Λ( θδ ) ts, Here, yt () is the ny -dimensional output at time t and () Further, et () is a sequence of independent and identically distributed (iid) random variables with zero mean, which is referred to as white noise. Further, Gq ( ; θ ) and Hq ( ; θ ) are filters of finite order (i.e. rational functions of q ). They have dimensions ( ny nu ) and ( ny ny ), respectively. () ut the nu -dimensional input. The parameter vector θ is to be estimated from the available input-output data y(), u(),..., y( ), u( ). In practice, one does not use the model Eq. () as such, but a special case such as an ARMAX model or some suitably parameterized state space model. To keep the description general, confine it here to the model Eq. () which covers deliberately all possible cases. A model obtained by identification can be used in many ways, depending on the purpose of modeling. In many applications the model is used for prediction. ote that this is often inherent when the model is to be used as a basis for control system synthesis. Most systems are stochastic, which means that the output at time t cannot be determined exactly from data available at time t. It is thus valuable to know, forecast or predict at time t what the output of the process is likely to be at time t, in order to take an appropriate control action, i.e. to determine the input ut ( ). It, therefore, makes sense to determine the model parameter vector θ so that the prediction error ε(, t θ) = y() t yˆ ( t t ; θ) (2) is small. In Eq. (2), yt ˆ( t; θ ) denotes a prediction of yt () given the data up to and including time t (i.e. yt ( ), ut ( ), yt ( 2), ut ( 2),...) and based on the model parameter vector θ.

ow formalize this idea and consider the general model structure introduced in Eq. (). Assume that G(0, θ ) = 0, i.e. that the model has at least one pure delay from input to output. The optimal mean square predictor can conceptually be written as 2 yt ˆ( t ; θ) = L( q ; θ)() yt + L( q ; θ)() ut (3) which is a function of past data only if the predictor filter L 2 ( q, θ ) are constrained by L ( q ; θ ) and L (0; θ ) = 0, L (0; θ ) = 0. (4) 2 Once the model and the predictor are given, the prediction errors are computed as in Eq. (2). The parameter estimate ˆ θ is then chosen to make the prediction errors ε(, θ),..., ε(, θ) small. To define a prediction error method the user has to make the following choices: Choice of model structure. This concerns the parameterization of Gq ( ; θ ), Hq ( ; θ ) and Λ( θ ) in Eq. () as functions of θ. Choice of criterion. This concerns a scalar-valued function of all the prediction errors ε(, θ),..., ε(, θ ), which will assess the performance of the predictor used; this criterion is to be minimized with respect to θ to choose the best predictor in the class considered. The most common choice of criterion is (for the single-output case) the sample variance of the prediction errors V 2 ( θ ) ε ( t, θ). (5) t = Several modifications exist, though, see subsection.6. For multivariable models, the criterion can be modified, for example, as ( ny > ) V 2 ( θ) = ε( t, θ), (6) t = but several other alternatives exist as well. The prediction error estimate of θ is now defined as the minimizing element of the criterion Eq. (5), i.e. ˆ = arg minv ( θ ) (7) θ θ

To complete the description, it remains to describe how the prediction errors are to be computed for the general linear model, Eq. (). This is done in subsection.5..2. General Linear Dynamic Models.2.. Introduction In this subsection, general linear models of the form Eq. () will be discussed. The model will be parameterized by a vector θ, which is to be estimated. The general form of model structure that will be used is the following M ( θ): yt ( ) = Gq ( ; θ) ut ( ) + Hq ( ; θ) et ( ), T Ee() t e () s =Λ ( θ ) δ ts,. (8) The filters Gq ( ; θ ) and Hq ( ; θ ) as well as the noise covariance matrix Λ ( θ ) are functions of the parameter vector θ. Often θ (which is assumed to be nθ - n dimensional) is restricted to lie in a subset of R θ. This set is given by D = { θ H ( q ; θ) and H ( q ; θ) G( q ; θ) are asymptotically stable is nonnegative definite G(0; θ) = 0, H(0; θ) = I, Λ ( θ) }. (9) The reasons for these restrictions in the definition of D will become clear in the following, where it will be shown that when θ belongs to D, there is a simple form for the optimal one step prediction of yt () given past data yt ( ), ut ( ), yt ( 2), ut ( 2),... For stationary disturbances with rational spectral densities it is a consequence of the spectral factorization theorem, that they can be modeled within the restrictions given by Eq. (9). Eq. (8) describes a general linear model. The following examples describe typical model structures by specifying the parameterization. That is to say, they specify how Gq ( ; θ ), Hq ( ; θ ) and Λ ( θ ) depend on the parameter vector θ..2.2. ARMAX Models Let yt () and ut () be scalar signals and consider the model structure A( q ) y( t) = B( q ) u( t) + C( q ) e( t) (0) where

na = + + + na Aq ( ) aq a q nb = + + nb Bq ( ) bq b q nc = + + + nc. Cq ( ) cq c q () The parameter vector is taken as θ = ( a... a b... b c... c ) T. (2) na nb nc The model Eq. (0) is called an ARMAX model, which is short for an ARMA model (AutoRegressive Moving Average) with an exogenous signal (i.e. a control variable ut () is present). The largest integer of the triple ( na, nb, nc ) is called the model order. To see how Eq. (0) relates to Eq. (8), note that it can be written as Bq ( ) Cq ( ) yt () = ut () et () Aq ( ) + Aq ( ). (3) Thus, for the model structure Eq. (0), Gq ( ; θ ) Hq ( ; θ ) = = Λ ( θ) = 2. Bq ( ) Aq ( ) Cq ( ) A( q ) λ The set D is given by (4) D= { θ The polynomial C( z) has all zeros outside the unit circel}. (5) A more standard formulation of the requirement polynomial θ D is that the reciprocal nc nc nc ( ) = + + + nc = ( ) (6) C z z c z c z C z has all zeros inside the unit circle. There are several important special cases of Eq. (0): An autoregressive (AR) model is obtained when nb = nc = 0. (Then a pure time series is modeled, i.e. no input signal is assumed to be present.) For this case

A( q ) y( t) = e( t) θ = ( a... a ) T na (7). A moving average (MA) model is obtained when na = nb = 0. Then yt () = Cq ( ) et () θ = ( c... c ) T nc. (8) An autoregressive moving average (ARMA) model is obtained when nb = 0. Then A( q ) y() t = C( q )() e t θ = ( a... a c... c ) T (9) na nc. When A( q ) is constrained to contain a factor q the model is called autoregressive integrated moving average (ARIMA). Such models are useful for describing drifting disturbances. The above three special cases do all apply to a pure time series (there is no input signal). Some other special cases are: A finite impulse response (FIR) model is obtained when na = nc = 0. It can also be called a truncated weighting function model. Then yt () = Bq ( ) ut () + et () θ = ( b... b ) nb. (20) Another special case is when nc = 0. The model structure then becomes A( q ) y() t = B( q ) u() t + e() t θ = ( a... a b... b ) T na This is sometimes called an ARX (controlled autoregressive) model. This structure can also be viewed as a linear regression, nb. (2) T yt () = ϕ () tθ + et (), (22) where ϕ ( t) = ( y( t )... y( t na) u( t )... u( t nb)) T. (23)

There are some other ways than the ARMAX model to introduce polynomials for linear single-input single-output models. If A( q ) and Cq ( ) in Eq. (0) are constrained to coincide, the so-called output error model Bq ( ) yt () = ut () + et () (24) Fq ( ) is obtained, where all parameters are used to model the filter no description in the model of the disturbance. Gq ( ; θ ), while there is The linear SISO models, such as ARX and ARMA models, can be extended to the multivariable case, but some complications occur in that there is no unique way to introduce free parameters. In case all matrix coefficients are left free, the model will lose uniqueness and it will then not be possible to identify the parameters, no matter what amount of data and what experimental condition is used. - - - Bibliography TO ACCESS ALL THE 23 PAGES OF THIS CHAPTER, Click here Some reference for more extensive background material include Anderson, B. D. O. and Moore, J. B. (979): Linear Optimal Filtering,357pp. Prentice Hall, Englewood Cliffs, J. [Includes background and derivations for spectral factorization, innovations form, and optimal prediction]. Gevers, M. and Wertz, V. (987): Techniques for selection of identifiable parametrizations for multivariable linear systems. In C T Leondes (ed.), Control and Dynamic Systems, vol 26; pp 35-86, System Identification and Adaptive Control Advances in Theory and Application, Academic Press, ew York. [Parameterization issues of linear multivariable systems]. Ljung, L. (999): System Identification. Theory for the User, 2 nd edition. 609pp.Prentice Hall. Upper Saddle River, J. [Description and analysis of prediction error and other identification methods]. Söderström, T. (2002): Discrete-time Stochastic Systems, 2 nd ed., 375pp.Springer-Verlag, London, UK. [Includes background and derivations for spectral factorization, innovations form, and optimal prediction]. Söderström, T. and Stoica, P. (989): System Identification,62pp.Prentice Hall International, Hemel Hempstead, UK[Description and analysis of prediction error and other identification methods]. Wahlberg. B. and. Ljung, L. (986): Design variables for bias distribution in transfer function estimation. IEEE Transactions on Automatic Control, vol AC-3, pp 34-44. [Description of the effects of undermodeling]

Biographical Sketch Torsten Söderström was born in Malmö, Sweden, in 945. He received the MSc degree (`civilingenjör') in engineering physics in 969 and the PhD degree in automatic control in 973, both from Lund Institute of Technology, Lund, Sweden. He is a Fellow of IEEE. In the period 967-974 he held various teaching positions at the Lund Institute of Technology. Since 974, he has been at Department of Systems and Control, Uppsala University, Uppsala, Sweden, where he is a professor of automatic control. Dr Söderström is the author or coauthor of more than 200 technical papers. His main research interests are in the fields of system identification, signal processing, and adaptive systems. He is the (co)author of four books: `Theory and Practice of Recursive Identification', MIT Press, 983 (with L. Ljung), `The Instrumental Variable Methods for System Identification', Springer Verlag, 983 (with P. Stoica), `System Identification', Prentice-Hall, 989 (with P Stoica) and `Discrete-time Stochastic Systems', Prentice-Hall, 994, second revised edition, Springer-Verlag, 2002. In 98 he was, with coauthors, given an Automatica Prize Paper Award. Within IFAC (International Federation of Automatic Control) he has served in several capacities including Automatica editor for the area of System Parameter Estimation since 992, Council member 996-2002, member of Executive Board and Chair of Awards Committee 999-2002, IPC chairman of the SYSID'94 symposium, and guest associate editor or editor for three special issues of Automatica.