Closed Loop Identification (L26)

Similar documents
Lecture 7 Open-loop & closedloop experiments. The Bias Formula goes closed-loop and the returns

EL1820 Modeling of Dynamical Systems

Identification, Model Validation and Control. Lennart Ljung, Linköping

EECE Adaptive Control

FRTN 15 Predictive Control

IDENTIFICATION FOR CONTROL

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.

System Identification: From Data to Model

On Input Design for System Identification

I. D. Landau, A. Karimi: A Course on Adaptive Control Adaptive Control. Part 9: Adaptive Control with Multiple Models and Switching

Matlab software tools for model identification and data analysis 11/12/2015 Prof. Marcello Farina

6.435, System Identification

EECE Adaptive Control

Advanced Process Control Tutorial Problem Set 2 Development of Control Relevant Models through System Identification

On Identification of Cascade Systems 1

Ten years of progress in Identification for Control. Outline

IDENTIFICATION OF A TWO-INPUT SYSTEM: VARIANCE ANALYSIS

Introduction to System Identification and Adaptive Control

Optimal Polynomial Control for Discrete-Time Systems

Direct adaptive suppression of multiple unknown vibrations using an inertial actuator

System Identification for MPC

EL1820 Modeling of Dynamical Systems

Nonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto

Cautious Data Driven Fault Detection and Isolation applied to the Wind Turbine Benchmark

Problem Set 5 Solutions 1

REGLERTEKNIK AUTOMATIC CONTROL LINKÖPING

Identification in closed-loop, MISO identification, practical issues of identification

OPTIMAL EXPERIMENT DESIGN IN CLOSED LOOP. KTH, Signals, Sensors and Systems, S Stockholm, Sweden.

Control Systems Lab - SC4070 System Identification and Linearization

6.241 Dynamic Systems and Control

Chapter 6: Nonparametric Time- and Frequency-Domain Methods. Problems presented by Uwe

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Prediction Error Methods - Torsten Söderström

Introduction. Performance and Robustness (Chapter 1) Advanced Control Systems Spring / 31

Wind Turbine Control

Subspace-based Identification

Non-parametric identification

Dynamic measurement: application of system identification in metrology

Chapter 3. LQ, LQG and Control System Design. Dutch Institute of Systems and Control

EL2520 Control Theory and Practice

14 th IFAC Symposium on System Identification, Newcastle, Australia, 2006

Notes for System Identification: Impulse Response Functions via Wavelet

Lecture 6: Deterministic Self-Tuning Regulators

COMBINED ADAPTIVE CONTROLLER FOR UAV GUIDANCE

Introduction to system identification

LTI Systems (Continuous & Discrete) - Basics

GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL of ELECTRICAL & COMPUTER ENGINEERING FINAL EXAM. COURSE: ECE 3084A (Prof. Michaels)

Industrial Model Predictive Control

L2 gains and system approximation quality 1

Lecture 4 Stabilization

Dynamic Factor Models, Factor Augmented VARs, and SVARs in Macroeconomics. -- Part 2: SVARs and SDFMs --

A FEEDBACK STRUCTURE WITH HIGHER ORDER DERIVATIVES IN REGULATOR. Ryszard Gessing

arxiv: v4 [cs.sy] 24 Oct 2017

Dynamic Regression Models (Lect 15)

Time domain identification, frequency domain identification. Equivalencies! Differences?

System Identification & Parameter Estimation

Problem 1: Ship Path-Following Control System (35%)

Online monitoring of MPC disturbance models using closed-loop data

GLS. Miguel Sarzosa. Econ626: Empirical Microeconomics, Department of Economics University of Maryland

Further Results on Model Structure Validation for Closed Loop System Identification

Close Loop System Identification of Dynamic Parameters of an Airplane. Michal Chvojka

Improving Convergence of Iterative Feedback Tuning using Optimal External Perturbations

Control Systems I. Lecture 6: Poles and Zeros. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

B y t = γ 0 + Γ 1 y t + ε t B(L) y t = γ 0 + ε t ε t iid (0, D) D is diagonal

New Direct Closed-Loop Identification Method for Unstable Systems and Its Application to Magnetic Suspension System

Feedback Control of Turbulent Wall Flows

Linear State Feedback Controller Design

System Identification

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

Lecture 12. Upcoming labs: Final Exam on 12/21/2015 (Monday)10:30-12:30

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

EE5356 Digital Image Processing

Change Detection with prescribed false alarm and detection probabilities. Mogens Blanke

Improving performance and stability of MRI methods in closed-loop

Near Optimal LQR Performance for Uncertain First Order Systems

Closed-Loop Identification of Fractionalorder Models using FOMCON Toolbox for MATLAB

ESTIMATION ALGORITHMS

Identifiability in dynamic network identification

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics

Discrete-time Controllers

CONTROL DESIGN FOR SET POINT TRACKING

Raktim Bhattacharya. . AERO 422: Active Controls for Aerospace Vehicles. Dynamic Response

Control System Design

Linear Approximations of Nonlinear FIR Systems for Separable Input Processes

Uncertainty and Robustness for SISO Systems

LTI Approximations of Slightly Nonlinear Systems: Some Intriguing Examples

Temporal-Difference Q-learning in Active Fault Diagnosis

Nonlinear Observers. Jaime A. Moreno. Eléctrica y Computación Instituto de Ingeniería Universidad Nacional Autónoma de México

Spectral factorization and H 2 -model following

Time-Varying Parameters

Structure of the Hessian

Identification of ARX, OE, FIR models with the least squares method

f-domain expression for the limit model Combine: 5.12 Approximate Modelling What can be said about H(q, θ) G(q, θ ) H(q, θ ) with

12. Prediction Error Methods (PEM)

Iterative Learning Control (ILC)

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)

Pole-Placement Design A Polynomial Approach

Dr Ian R. Manchester

Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science : MULTIVARIABLE CONTROL SYSTEMS by A.

H 2 Optimal State Feedback Control Synthesis. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Transcription:

Stochastic Adaptive Control (02421) www.imm.dtu.dk/courses/02421 Niels Kjølstad Poulsen Build. 303B, room 016 Section for Dynamical Systems Dept. of Applied Mathematics and Computer Science The Technical University of Denmark Email: nkpo@dtu.dk phone: +45 4525 3356 mobile: +45 2890 3797 2018-05-08 22:10 Closed Loop Identification (L26) 1 / 28

System identification - Open loop ID y Exciter u System e The input (u) do not depend on y (and e). 2 / 28

System identification - Closed Loop ID w y Controller u System e Why? unstable plant in a production, economic and safety reasons inherent feedback mechanics (e.g. economics) adaptive control point of operation same as the intended use ( ) input spectrum same as in intended use ( ) 3 / 28

Closed loop identification Avoid closed loop identification if possible (at all cost) 1987 Use closed loop identification whenever possible 4 / 28

Pitch fall - Example I Plant: y t +ay t 1 = bu t 1 +e t e t N iid ( 0,σ 2 ) Control (Notice: we are only considering the regulation problem ): u t = fy t In closed loop: y t +(a+bf)y t 1 = e t Spectral analysis φ yu(ω) = G(e jω )φ u(ω) Ĝ(e jω ) = ˆφ yu(ω) ˆφ u(ω) 5 / 28

Example I Let us analyse the convergence point. Ψ u(z) = H u(z)h u(z 1 )σ 2 Here H xis the transfer function from noise to x Ψ yu(z) = H y(z)h u(z 1 )σ 2 Ĝ(e jω ) = φyu(ω) φ u(ω) = Hy(z) H u(z) z = e jω 6 / 28

Example I In closed loop we have: y t +(a+bf)y t 1 = e t and u t = fy t or H y(z) = 1 f H u(z) = 1+(a+bf)z 1 1+(a+bf)z 1 Then: Ĝ(z) = Hy(z) H = 1 1+(a+bf)z 1 = 1 u(z) 1+(a+bf)z 1 f f We might have to take care when doing ID in closed loop. Here the problem is (lack of) ensuring causality. 7 / 28

Pitch fall - Example II Prediction error method The plant: y t +ay t 1 = bu t 1 +e t e t N iid ( 0,σ 2 ) have the prediction error ε t = y t +ây t 1 ˆbu t 1 but in closed loop (u t = fy t) : ε t = y t +(â+ˆbf)y t 1 ŷ t t 1 = ây t 1 +ˆbu t 1 ŷ t t 1 = (â+ˆbf)y t 1 That means that any â = a 0 +γf ˆb = b0 γ (where γ is arbitrary scalar) gives equally good predictions. The controller is too simple (ie. not complex enough). 8 / 28

Closed loop ID If the plant A(q 1 )y t = q k B(q 1 )u t +C(q 1 )e t e t N iid ( 0,σ 2 ) is controlled by u t = S(q 1 ) R(q 1 ) yt Notice only focus on the regulation problem. The closed loop is [ AR+q k BS ] y t = RCe t The number of equations outnumbers the number of parameters (in A and B) if ] Max [n r n b, n s +k n a 1+n p where n p is the number of common factors in C and AR+q k BS. The controller has to be adequately complex 9 / 28

Closed loop ID Proof: Number of parameters: n a +n b +1 Order of denominator, AR +q k BD, (equals number of equations): Max(n a +n r,k +n b +n s) n p Match: ] Max [n r n b,n s +k n a = 1+n p Number of common factors (between denominator and numerator), n p. 10 / 28

The Minimal Variance Controller Ay t = q k Bu t +Ce t { } J = E yt+k 2 C = AG+q k S y t+k = 1 C [BGu t +Sy t ] +Ge t+k y t = q kb A ut + C A et = q kb A ( AG+q k S ) y t = CGe t S BG yt + C A et n r = n b +k 1 n p = n c n s = n a 1 ] Max [n r n b,n s +k n a 1+n p k n c +2 11 / 28

Example II (again) Plant: y t +ay t 1 = bu t 1 +e t Control: u t = fy t Here: ] Max [n r n b,n s +k n a = 1+n p = Max [ 0 0,0+1 1 ] = 0 < 1+0 = 1 12 / 28

Informative data Sufficient informative: A set of data, {z t}, is sufficient informative with respect to a model set M if, for two models, G 1 and G 2, in the set { Ē (G 1 (q) G 2 (q))z t 2} = 0 implies that G 1 (e jω ) G 2 (e jω ) for almost all ω. Persistenly exciting 1 r k = lim N N N i=1 } E {u tu t k R n = r 0 r 1 r m 1 r 1 r 0... r n 1 r n 2 r 0 u t pe(n) if R n > 0 13 / 28

The bad news - in summary The closed loop experiment may be non-information even if the input in itself is persistently exciting. Reason: the controller might be too simple. Spectral analysis applied in the straightforward fashion will give erroneous results. The estimate of G will converge to G = G 0φ w Fφ v φ w + F 2 φ v y = Gu+v Correlation methods will give biased estimate of the impulse response. OE gives unbiased estimate of G in open loop experiments, even if the additive noise (v) is not white. This is not true in close loop. The subspace methods will typically not give consistent estimate when applied to close loop data. 14 / 28

The good news PE methods will give consistent estimate of the system if The data is informative. The model set contains the true system (S M). The closed loop experiment is informative if the reference w t (or another probe signal) is persistently exciting. 15 / 28

Patches Persistingly exciting reference or probe signal w r y Controller u System e Time-varying ( adaptive ) or nonlinear controller Shift between m different LTI controllers where u t = F i (q)y t m 1+ nu n y i = 1,... m 16 / 28

Closed loop identification methods Approaches to closed loop identification Direct approach Indirect approach Joint Input- output Approach We assume in the following that the reference (or a probe signal) is persistingly exciting. 17 / 28

Closed loop identification w F r u G v y 18 / 28

Direct Approach The system is identified in exactly the same way as in open loop identification. It works regardless of the complexity of the controller and requires no knowledge about the character of the feedback. No special algorithms or software are required. (A word from our sponsers). Consistency and optimal accuracy are obtained if model structure contains the true system. Unstable system can be handled without problems (as long as the closed loop and the predictor are stable). Drawback: we need good noise models. (Not a problem if true system (G,H) is contained in model structure). If noise model is incorrect (fixed incorrectly or not contain the true noise model) bias of G will be introduced. 19 / 28

Indirect approach Assume the plant is given by ( y t = Gu t +v t = Gu t +He t e t N iid 0,σ 2 ) and the control is (using partially the notation in LL): u t = w t Fy t Then the closed loop is given by y t = GSw t +Sv t = u t = Sw t FSv t = where the sensitivity functions is: S = 1 1+FG G 1+FG wt + H 1+FG et 1 FH wt 1+FG 1+FG et 20 / 28

Indirect approach In the indirect method the closed transfer functions are estimated and from these the plant parameters (or transfer functions) are determined. y t = G cl w t +H cl e t where G cl = Consequently: Ĝ = G 1+FG Ĝ cl 1 ĜclF H cl = H 1+FG Ĝ cl Ĥ = Ĥcl(1+FĜ) = Ĥcl(1+F 1 ĜclF ) + Any (open loop) method such as spectral analysis, instrumental variable, subspace and prediction error methods can be applied. - Any error in F will transported directly to the estimate of the model. (Notice saturation, manual operator a.o.) 21 / 28

Indirect approach Prediction error methods The parametrization might be directly in the system parameters ie. and G cl = G(θ) 1+FG(θ) ) ε t = H 1 cl (y t G cl u t H cl = H(θ) 1+FG(θ) We then have the mapping ] θ [G cl,h cl ε t J = 1 2 ε2 t Other parametrizations exists. Methods based on the Youla-Kucera parametrization has been proposed by Hansen, Franklin and Kosut (1989) and Schrama (1991). 22 / 28

Joint Input-Output Approach Let the system be y t = Gu t +v t If the controller is v t = He t u t = w t Fy t +z t where z t is an unknown signal (partially unknown controller), then the closed loop is characterized by y t = GSw t + Sv t +GSz t = G cl w t +v 1 u t = Sw t FSv t + Sz t = G uww t +v 2 where S = 1 1+FG 23 / 28

Approach 1 Let the model [ ] [ yt G = S u t 1 or be used. ] [ w t +S [ ] [ ] yt et = Gw u t +H t z t H G FH 1 ][ et z t ] S = 1 1+FG Then a ML or a PEM method is basicly a minimization of where J = t ε i R 1 ε i i=1 ε i = H 1[[ y i u i ] Gw i ] where G(θ), H(θ) and F(θ). { [ ] e } R = Var i z i This parametrization might be independent, i.e. G(θ), H(γ) and F(η). 24 / 28

Approach 1 Consider a special case, [ ] σ 2 R = e 0 0 σz 2 It can be shown (after some manipulations, LL p. 438) that this in essence is equivalent to minimizing: J = + 1 t σe 2 (H 1 (θ)(y i G(θ)u i )) 2 i=1 1 t σz 2 (u i w i +F(θ)y i ) 2 i=1 If the parametrization of (G, H) and F are independent, this is a direct method for estimation G, H and F. 25 / 28

Approach 2 Here the correlation is disregarded and the two equations are treated as separate and as [ ] [ ] [ ] yt Gcl v1 = w u t G t + uw v 2 J = 1 σ 2 1 This approach has many variants, which have in common that Ĝ = Ĝcl Ĝ uw (G cl = GS G uw = S) t (y i G cl w i ) 2 + 1 t σ 2 (u i G uww i ) 2 i=1 2 i=1 If nothing is done to prevent it, Ĝ would be of unnecessarily high order. (Lack of perfect cancellation) One way to prevents this is to use an independent parametrization of G and S which results in [ ] ϑ G cl = G(ϑ)S(η) G uw = S(η) θ = η 26 / 28

Approach 2 - A two stage approach If additionally the two noise sources are (assumed to be) independent with a variance ratio β then the cost to be minimized is t t J(ϑ,η) = β (y i G(ϑ)S(η)w t) 2 + (u i S(η)w i ) 2 i=0 i=0 If β 0 then η is determined to fit the second part of the cost. A two step procedure could then consist of (Van den Hof+Schrama, Forssel+Ljung): 1 Estimate η for β 0. 2 Use û t = S(ˆη)w t and estimate ϑ by fitting : y t = G(ϑ)û t +v 1 One example on a parameterization of S is m S(η) = s k q k k= m which is a non causal filter. 27 / 28

Furthermore Special courses - Master projects Control of dynamical/stochastic systems, dynamics optimization Modelling and System identification Kalman filtering (state estimation and monitoring) and fault diagnosis Navigation (mobile robots)[sonardyne, Rovsing, DTU-elektro (or,mb)] Wind energy (wind turbine and farms) [hm, Vestas, Siemens, Risø] Artificial pancreas [jbj, hm, Novo] Fault diagnosis [DTU-elektro (hhn)] Adv. system identification (02904) Time series analysis (ord. and adv.) [hm] Robust and fault tolerant [mb, hhn] MPC course [jbj] Static and dynamic optimization (42111/02711) 28 / 28