On Input Design for System Identification

Size: px
Start display at page:

Download "On Input Design for System Identification"

Transcription

1 On Input Design for System Identification Input Design Using Markov Chains CHIARA BRIGHENTI Masters Degree Project Stockholm, Sweden March 2009 XR-EE-RT 2009:002

2 Abstract When system identification methods are used to construct mathematical models of real systems, it is important to collect data that reveal useful information about the systems dynamics. Experimental data are always corrupted by noise and this causes uncertainty in the model estimate. Therefore, design of input signals that guarantee a certain model accuracy is an important issue in system identification. This thesis studies input design problems for system identification where time domain constraints have to be considered. A finite Markov chain is used to model the input of the system. This allows to directly include input amplitude constraints into the input model, by properly choosing the state space of the Markov chain. The state space is defined so that the model generates a binary signal. The probability distribution of the Markov chain is shaped in order to minimize an objective function defined in the input design problem. Two identification issues are considered in this thesis: parameter estimation and NMP zeros estimation of linear systems. Stochastic approximation is needed to minimize the objective function in the parameter estimation problem, while an adaptive algorithm is used to consistently estimate NMP zeros. One of the main advantages of this approach is that the input signal can be easily generated by extracting samples from the designed optimal distribution. No spectral factorization techniques or realization algorithms are required to generate the input signal. Numerical examples show how these models can improve system identification with respect to other input realization techniques. 1

3 Acknowledgements Working on my thesis at the Automatic Control department at KTH has been a great experience. I would like to thank my supervisor Professor Bo Wahlberg and my advisor Dr. Cristian R. Rojas for having given me the possibility to work on very interesting research topics. Thank you for your guidance. A special thank goes to Cristian R. Rojas, for the time he spent on me and for the many ideas shared with me. Thanks to all the people I had the pleasure to know during this work period: Mitra, Andre B., Fotis, Pedro, Mohammad, Andre, Pierluigi, Matteo, Davide and Alessandro. I really enjoyed your company. I would like to thank Professor Giorgio Picci who made this experience possible. Finally, thanks to my family and my friends for their continuous help and support. 2

4 Acronyms AR Autoregressive FDSA Finite Difference Stochastic Approximation FIR Finite Impulse Response LMI Linear Matrix Inequality LTI Linear Time Invariant NMP Non Minimum Phase PEM Prediction Error Method PRBS Pseudo Random Binary Signal RLS Recursive Least Squares SISO Single Input Single Output SPSA Simultaneous Perturbation Stochastic Approximation WN White Noise I

5 Contents Abstract I 1 Introduction Thesis outline and contributions System Identification System and model description Identification method Estimate uncertainty Parameter uncertainty Frequency response uncertainty Conclusions Input Design for System Identification Optimal input design problem Measures of estimate accuracy Quality constraint based on the parameter covariance Quality constraint based on the model variance Quality constraint based on the confidence region Input spectra parametrization Finite dimensional spectrum parametrization Partial correlation parametrization Covariance matrix parametrization Signal constraints parametrization Limitations of input design in the frequency domain Conclusions Markov Chain Input Model Introduction Markov chains model Markov chain state space II

6 4.2.2 Markov chains spectra More general Markov chains Conclusions Estimation Using Markov Chains Problem formulation Solution approach Cost function evaluation Algorithm description Numerical example Conclusions Zero Estimation Problem formulation FIR ARX General linear SISO systems Adaptive algorithm for time domain input design A motivation Algorithm description Algorithm modification for Markov chain signals generation Numerical example Conclusions Summary and future work Summary Future work References 55 III

7 List of Figures 4.1 Graph representation of the two states Markov chain S Graph representation of the four states Markov chain S Graph representation of the three states Markov chain S Mass-spring-damper system Cost functions f 2 and f 3 in the interval of acceptable values of the variables u max and y max, respectively Cost functions f 1, f 4,f 5 in the interval of acceptable values Estimate of the cost function on a discrete set of points for the two states Markov chain in the case Estimation of the best transition probability for the two states Markov chain in the case Estimation of the best transition probability p for the four states Markov chain in the case Estimation of the best transition probability r for the four states Markov chain in the case Estimate of the cost function on a discrete set of points for the two states Markov chain in the case Estimate of the cost function on a discrete set of points for the four states Markov chain in the case Bode diagrams of the optimal spectra of the 2 states Markov chains in the cases 1, 2 and 3 of Table 5.3, and of the real discrete system Bode diagrams of the optimal spectra of the 4 states Markov chains in the cases 1, 2 and 3 of Table 5.3, and of the real discrete system Estimates of the frequency response of the system using the optimal two states Markov chains for the cases 1, 2 and 3 in Table IV

8 5.13 Estimates of the frequency response of the system using the optimal four states Markov chains for the cases 1, 2 and 3 in Table Representation of the adaptive algorithm iteration at step k. u k denotes the vector of all collected input values from the beginning of the experiment, used for the output prediction ŷ (k) A zero estimate trajectory produced by the adaptive algorithms described in Sections and Normalized variance of the estimation error for the adaptive algorithms described in Sections and V

9 List of Tables 4.1 Poles and zeros of the canonical spectral factor of the spectrum of s m of Example Maximum threshold values in the three analyzed cases Results of 100 Monte-Carlo simulations of the algorithm with the 2 states Markov chain Optimal values of the transition probabilities in the cases 1, 2 and 3, obtained after algorithm iterations Total cost function values obtained with the optimal Markov inputs, a PRBS and white noise in case Trace of the covariance matrix obtained with the optimal Markov inputs, a PRBS, white noise, a binary input having the optimal correlation function and the optimal spectrum in case Total cost function value obtained with the optimal Markov inputs, a PRBS and white noise in case Estimated values of the parameters of the continuous real system and relative percentage errors, obtained with the optimal two states Markov chains Estimated values of the parameters of the continuous real system and relative percentage errors, obtained with the optimal four states Markov chains VI

10 Chapter 1 Introduction Mathematical models for systems are necessary in order to predict their behavior and as parts of their control systems. This work focuses on models constructed and validated from experimental input/output data, by means of identification methods. Information obtained through experiments on the real system depend on the input excitation which is often limited by amplitude or power constraints. For this reason, experiment design is necessary in order to obtain system estimates within a given accuracy, saving time and cost of the experiment [1]. Robustness of input design for system identification is also one of the most important issues, specially when the model of the system is used for projecting its control system. In [2]- [7] some studies on this problem are presented. The effects of undermodeling on input design are pointed out in [8] and [9]. Depending on the cost function considered in this setting, input design can be typically solved as a constrained optimization problem. In the Prediction Error Method (PEM) framework it is common to use, as a measure of the estimate accuracy, a function of the asymptotic covariance matrix of the parameter estimate. This matrix depends on the input spectrum that can then be shaped in order to obtain a small covariance matrix and improve the estimate accuracy (see [10], [11]). Usually, a constraint on the input power is also included; in this way, time domain amplitude constraints are approximately translated in the frequency domain [12]. A first disadvantage of these methods is that they are strongly influenced by the initial knowledge of the system. Secondly, solving the problem in the frequency domain does not provide any further information on how to generate the input signal in the time domain: the input can be represented as filtered white noise, but many probability distributions can be used to 1

11 CHAPTER 1. INTRODUCTION 2 generate white noise. Furthermore, in practical applications time domain constraints on signals have to be considered and the power constraint that is usually set in the frequency domain does not assure that these constraints are respected. For this reason, in [13] a method is proposed to generate a binary input with a prescribed correlation function; once an optimal spectrum or correlation function is found solving the input design problem in the frequency domain, it is possible to generate a binary signal which approximates the optimal input. Also in [14] a method is proposed that provides a locally optimal binary input in the time domain. This thesis studies the input design problem in the probability domain. Compared to design methods in the frequency domain, a solution in the probability domain makes it easier to generate input trajectories to apply to the real system, by extracting samples from a given distribution. Inputs are modeled by finite stationary Markov chains which generate binary signals. Binary signals are often used in system identification and one of the reasons is that they achieve the largest power in the set of all signals having the same maximum amplitude and it is well known that this improves parameter estimation for linear models. The idea of modeling the input by a finite Markov chain derives from the possibility of including input amplitude constraints directly into the input model, by suitably choosing the state space of the Markov chain. Furthermore, unlike the design in the frequency domain, this approach keeps more degrees of freedom in the choice of the optimal spectrum, which in general is non unique [12]. Two identification problems are considered here: parameter estimation and non-minimum phase zeros estimation for LTI systems. For the first problem, the optimal distribution is found by minimizing the cost function defined in the input design problem with respect to the one-step transition probabilities of the Markov chain. In this analysis, a stochastic approximation algorithm is used since a closed-form solution to the optimization problem is not available and the cost is a stochastic function of these transition probabilities and is contaminated with noise (see [15], [16] for details). For the second problem, it will be shown that the Markov chain input model has exactly the optimal spectrum for the zero estimation problem of a FIR or ARX model [6]. In general, the spectrum of a two states Markov chain can be made equal to the spectrum of the AR process which guarantees a consistent estimate of the NMP zero of a linear SISO system [17]. Therefore, an optimal or consistent input can be generated in the time domain by Markov chain distributions. The adaptive algorithm introduced

12 CHAPTER 1. INTRODUCTION 3 in [18] for input design in the time domain when undermodeling is considered, is modified here in order to generate Markov chain signals having the same spectrum as the general inputs designed in the original version of the algorithm. The advantage is that a binary signal is then used to identify the non-minimum phase zero, by keeping the same input variance and spectrum. The outline of the thesis is presented in the next section. 1.1 Thesis outline and contributions The subject of this thesis is input design for system identification. In particular, the objective of this study is to analyze a new approach to input design: work in the probability domain model the input signal as a finite Markov chain. The first chapters summarize some known results of system identification in the PEM framework and input design in the frequency domain, in order to compare methods and results obtained with the classical input design method in the frequency domain and with the method proposed here. In Chapter 2 PEM and its asymptotic properties are reviewed. In Chapter 3 the most commonly adopted input design methods are described. These formulate input design problems as convex optimization problems. The solution is given as an optimal input spectrum. In Chapter 4 general Markov chains that model binary input signals are defined. Some spectral properties are also described. In Chapter 5 are presented the input design problem for parameter estimation and the solution approach based on the Markov chain input model. In Chapter 6 input design for identification of NMP zeros of a LTI system is considered. The chapter discusses classical solutions in the frequency domain and adaptive solutions in the time domain in undermodeling conditions where the input is modeled as an AR or a Markov chain process. Chapter 7 concludes the thesis.

13 Chapter 2 System Identification This chapter introduces system identification in the PEM framework for parameter estimation of LTI models. Once a model structure is defined, this method finds the model s parameters that minimize the prediction error. Even when the model structure is able to capture the true system dynamics, the estimate error will not be zero, since data used in the identification procedure are finite and corrupted by noise. The purpose of input design is to minimize the estimate error by minimizing the variance of the parameter estimates, assuming the estimation method is consistent. Section 2.1 defines systems and model structures considered in this work, while Section 2.2 discusses PEM and its asymptotic properties. 2.1 System and model description This thesis considers discrete-time LTI SISO systems, lying in the set M of parametric models y (t) = G (q, θ) u (t) + H (q, θ) e (t), (2.1) G (q, θ) = q n k b 1 + b 2 q b nb q n b 1 + a 1 q a na q na H (q, θ) = 1 + c 1q c nc q n c 1 + d 1 q d nd q n d, θ = [b 1,..., b nb, a 1,... a na, c 1,... c nc, d 1,... d nd ] T R b 1 where u (t) is the input, y (t) is the output and e (t) is zero mean white noise with finite variance. The symbol q 1 represents the delay operator (q 1 u (t) = u (t 1)). Assume H (q, θ) is stable, monic and minimum phase, i.e. poles and zeros lie inside the unit circle. 4

14 CHAPTER 2. SYSTEM IDENTIFICATION 5 The real system S is given as y (t) = G 0 (q) u (t) + H 0 (q) e 0 (t), (2.2) where e 0 (t) has finite variance λ 0. Assume there exists θ 0, parameter vector such that G (q, θ 0 ) = G 0 (q) and H (q, θ 0 ) = H 0 (q), i.e. assume there is no undermodeling: S M. (2.3) This condition is hardly satisfied in practice, since real systems often are of high order or non linear. Nevertheless, as it will be explained in the next section, this condition is crucial for the consistence and the asymptotic properties of PEM. In Chapter 5, regarding the parameter estimation problem, (2.3) will be supposed to hold, while in Chapter 6, where the zero estimation problem is analyzed, this condition will not necessarily be considered. 2.2 Identification method System identification aims at describing a real system through a mathematical model constructed and validated from experimental input-output data. The identification method considered here is PEM [19]. This method minimizes the function of the prediction error ε F (t, θ) V N ( θ, Z N ) = 1 2N N ε 2 F (t, θ) (2.4) where Z N is a vector containing the collected input-output data, i.e. Z N = [y (1), u (1), y (2), u (2)..., y (N), u (N)]. The prediction error is defined as ε F (t, θ) = y (t, θ) ŷ (t, θ), where the one-step ahead predictor is given by t=1 ŷ (t, θ) = H 1 (q, θ) G (q, θ) u (t) + [ 1 H 1 (q, θ) ] y (t). Suppose all the hypothesis for the consistence of PEM are satisfied; in that case, the parameter estimate ˆθ N converges to the true parameter vector θ 0 as N tends to infinity. Briefly, these conditions are 1. Condition (2.3) holds, i.e. there is no undermodeling 2. The signals y (t) and u (t) are jointly quasi-stationary 3. u (t) is persistently exciting of sufficiently large order

15 CHAPTER 2. SYSTEM IDENTIFICATION Estimate uncertainty Measuring the quality of the model estimate is an important issue in system identification. The measure is chosen depending on the application for which the model is required. One possibility to measure the estimate uncertainty is to use a function of the covariance matrix of the parameter estimate. In other cases, such as in control applications, it could be better to use the variance of the frequency response estimate in the frequency domain. These two cases are now presented in more detail Parameter uncertainty Under the assumptions for the consistency of PEM, it holds that N (ˆθN θ 0 ) N (0, P θ0 ) as N (2.5) P 1 θ 0 = 1 λ 0 E [ ψ (t, θ 0 ) ψ T (t, θ 0 ) ] ψ (t, θ 0 ) = ŷ (t, θ) θ where N denotes the Normal distribution [19]. Therefore, when the model class is sufficiently flexible to describe the real system, the parameter estimate will converge to the true parameter vector as the number of data N used in the estimation goes to infinity, with covariance decaying as 1 /N. From (2.5) it follows that a confidence region in which the parameter estimate will lie with probability α is { ( U θ = θ N θ ˆθ ) T ( N P 1 θ 0 θ ˆθ ) } N χ 2 α (n). (2.6) The covariance matrix defines an ellipsoid asymptotically centered in θ 0. Upon the condition that u and e are independent (that is, data are collected in open-loop), the asymptotic expression in the number of data points N of the inverse of the covariance matrix of the parameter estimate is P 1 θ 0 = N 2πλ 0 R e (θ 0 ) = N 2π π π π π θ0 ( F u e iω ), θ 0 Φu (ω) Fu ( e iω ), θ 0 dω + Re (θ 0 ) (2.7) F e ( e iω, θ 0 ) F e ( e iω, θ 0 ) dω

16 CHAPTER 2. SYSTEM IDENTIFICATION 7 where ( F u e iω ), θ 0 = H 1 ( e iω ) G ( e iω, θ ), θ 0 (2.8) θ θ0 ( F e e iω ), θ 0 = H 1 ( e iω ) H ( e iω, θ ), θ 0 (2.9) θ θ0 and Φ u (ω) is the power spectral density of the input u (t). Here denotes the complex conjugate transpose. Expression (2.7) shows that the asymptotic covariance matrix of the parameter estimate depends on the input spectrum. Therefore, by shaping Φ u (ω) it is possible to obtain estimates within a given accuracy. In the whole thesis it will assumed that there is no feedback in the system, i.e. u and e are independent Frequency response uncertainty In many applications, it could be preferable to measure the quality of the model estimate using the variance of the frequency response estimate, frequency by frequency. In [19] it is shown that under the condition (2.3), the variance of G (e iω, ˆθ ) N can be approximated by ( Var G (e iω, ˆθ )) N m Φ v (ω) N Φ u (ω) (2.10) for large but finite model order m and number of data N, where v is the process defined as v (t) = H 0 (q) e 0 (t). If the model order is not large enough the previous expression is not a good approximation. Instead, by the Gauss approximation formula, it is possible to write ( Var G (e iω, ˆθ )) N 1 G ( e iω, θ ) G ( e iω, θ ). (2.11) N θ θ θ0 θ 0 P θ0 Equation (2.11) expresses the frequency response uncertainty in terms of the parameter uncertainty. Therefore, both equations (2.10) and (2.11) show that a proper choice of the input spectrum can reduce the variance of the frequency response estimate. This is the purpose of input design. 2.4 Conclusions This chapter introduced PEM for system identification and its asymptotic properties that are often used to solve input design problems.

17 CHAPTER 2. SYSTEM IDENTIFICATION 8 Models constructed from experimental data are always affected by uncertainty. In Section 2.3 two possible measures of the model uncertainty were discussed: parameter uncertainty and frequency response uncertainty. The choice between the two depends on the application. Typically, in control applications a measure in the frequency domain is preferable. This work considers parameter uncertainty.

18 Chapter 3 Input Design for System Identification This chapter presents general input design problems for system identification. Typically, input design aims at optimizing some performance function under constraints on the estimate accuracy and on the input signal. The solution approach will be presented in detail, describing how input design problems can be formulated as convex optimization problems. Ideas and drawbacks of the general input design framework are reviewed in Section 3.1. The most widely used measures of estimate accuracy are presented in Section 3.2. Here is also shown how quality constraints can be written as convex constraints. Sections 3.3 to 3.5 describe some techniques used for spectra and signal constraints parametrization, needed for handling finitely parametrized problems. 3.1 Optimal input design problem In a general formulation, input design problems are constrained optimization problems, where the constraints are typically on the input signal spectrum or power and the estimate accuracy. In this framework the objective function to be optimized can be any performance criterion, which usually depends on the practical application. For example, input power or experiment time can be minimized. Common input design problem formulations are: 1. Optimize some measure of the estimate accuracy, under constraints on input excitation. 2. Optimize some property of the input signal, given constraints on the 9

19 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 10 estimate accuracy. As it will be discussed in the next section, typical measures of the estimate accuracy are functions of the uncertainty in the model estimate, like (2.7), (2.10) and (2.11). As was shown in Section 2.3, these functions depend on the input spectrum, which therefore can be used to optimize the objective function. A formal expression of the first problem formulation is that can also be written as subject to g (Φ u ) α, min Φ u f (P θ0 ) (3.1) subject to min Φ u γ (3.2) f (P θ0 ) γ g (Φ u ) α. The next sections will show how this type of constraints can be formulated as convex constraints, upon certain conditions on the functions f and g. The following drawbacks of input design problems like (3.2) have to be enlightened. First of all, notice that the asymptotic expression of the covariance matrix depends on the true parameter θ 0 that is not known. Secondly, the constraints may be non-convex and infinite dimensional. In that case, a parametrization of the input spectrum is necessary in order to handle finitely parametrized optimization problems. Furthermore, once the optimal input spectrum has been found, an input signal having that optimal spectrum has to be generated. This can be done by filtering white noise with an input spectral factor 1. Nevertheless, no information on the probability distribution of the white noise is given in this solution approach. 3.2 Measures of estimate accuracy In the usual input design framework, three types of quality measures are typically considered. These are described in the following sections. 1 By spectral factor is meant an analytic function L (z) such that Φ u (z) = L (z) L ( z 1).

20 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION Quality constraint based on the parameter covariance In Section 2.3 it has been shown that the asymptotic covariance matrix of the parameter estimate is an affine function of the input spectrum, through (2.7). If the purpose of the identification procedure is parameter estimation, then typical scalar measures of estimate accuracy are the trace, the determinant or the maximum eigenvalue of the covariance matrix [12]. Then the following quality constraints can be introduced: TrP θ0 γ (3.3) detp θ0 δ (3.4) λ max (P θ0 ) ɛ. (3.5) It is possible to prove that these constraints can be manipulated to be convex in P 1 θ 0 ; proofs can be found in [20] and [21] for (3.4) and (3.5), respectively. For example, (3.5) is equivalent to [ ] Iɛ I 0 (3.6) I P 1 θ 0 which is an LMI in P 1 θ 0. The constraint (3.3) is a special case of the more general weighted trace constraint that will be considered in the next subsection. Notice that all these quality constraints depend on the true parameter vector θ 0. Many solutions have been presented in the literature to handle this problem, as will be discussed afterwards Quality constraint based on the model variance Consider the quality constraint based on the variance of the frequency response 1 π ( F (ω) Var G (e iω, 2π ˆθ )) N dω γ, (3.7) π where F (ω) is a weighting function. By substituting the variance expression (2.11), it results that this quality constraint can be written as where W = 1 π 1 2π π N TrW P θ0 γ, (3.8) G ( e iω, θ ) F (ω) G ( e iω, θ ) dω. (3.9) θ θ θ0 θ 0 See [12] for details. The following Lemma generalizes the previous result. A proof can be found in [6].

21 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 12 Lemma The problem TrW (ω) P θ0 γ, ω W (ω) = V (ω) V (ω) 0, ω P θ0 0 can be formulated as an LMI of the form [ Z V (ω) γ TrZ ] 0 V (ω) 0, ω. P 1 θ 0 (3.10) Notice that this formulation is convex in the matrices P 1 θ 0 and Z (see [12], [6]). Notice also that this type of constraint includes the constraint on the trace of the covariance matrix (3.3) as a special case, when W is the identity matrix Quality constraint based on the confidence region In control applications, it is often preferable to have frequency by frequency constraints on the estimate error. Consider the measure [6, 12] ( e iω, θ ) = T ( e iω) G ( 0 e iω ) G ( e iω, θ ) G (e iω. (3.11), θ) The input has to be designed so that ( e iω) γ, ω, θ U θ. (3.12) This constraint can also be formulated as a convex constraint in P 1 θ 0, as proven in [22]. 3.3 Input spectra parametrization As discussed in the last section, the typical measures of estimate accuracy are functions of the covariance matrix P θ0. Expression (2.7), derived in the asymptotic analysis of PEM, shows that the input spectrum Φ u can be used to optimize the estimate performance. The problem of finding an optimal input spectrum has an infinite number of parameters, since Φ u (ω) is a continuous function of the frequency ω. Nevertheless, by a proper spectrum parametrization, it is possible to formulate the problem as a convex

22 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 13 and finite dimensional optimization problem, since the parametrization of the input spectrum leads to a parametrization of the inverse of the covariance matrix [6, 12]. A spectrum can always be written in the general form Φ u (ω) = k= c k B k ( e iω ), (3.13) where { ( B k e iω )} are proper stable rational basis functions that span2 k= L 2. It is always possible to choose basis functions having the hermitian properties B k = Bk and B ( k e iω ) = Bk ( e iω ) [6]. The coefficients c k satisfy the symmetry property c k = c k and must be such that Φ u (ω) 0, ω, (3.14) otherwise Φ u would not be a spectrum. For example, the FIR representation of a spectrum is obtained by choosing B k (ω) = e iωk ; consequently, it results c k = r k, where r k is the correlation function of the process u. By substituting (3.13) into (2.7), the inverse of the covariance matrix becomes an affine function of the coefficients c k of the form P 1 θ 0 = k= c k Q k + Q. (3.15) This parametrization of the input spectrum leads to a denumerable but infinite number of parameters in the optimization problem. Two possible spectra parametrizations are described in the following subsections, which make the problem finitely parametrized Finite dimensional spectrum parametrization The finite dimensional spectrum parametrization has the form Φ u (ω) = Ψ ( e iω) + Ψ ( e iω) Ψ ( e iω) = M 1 k=0 c k B k ( e iω ). (3.16) This parametrization forces the coefficients c M, c M+1,... to be zero. Therefore, the condition Φ u (ω) 0 must be assured through the coefficients { c k } M 1 k=0. The following result, deriving from an application of the Positive Real Lemma [23], can be used to assure the constraint (3.14). 2 L 2 denotes the set { f f (x) 2 dx < }

23 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 14 Lemma Let {A, B, C, D} be a controllable state-space realization of Ψ ( e iω). Then there exists a matrix Q = Q T such that ( ) ( ) Q A T QA A T QB 0 C T B T QA B T + QB C D + D T 0 (3.17) if and only if Φ u (ω) M 1 k=0 c [ ( k Bk e iω ) + Bk ( e iω )] 0, ω. The state-space realization of the positive real part of the input spectrum can be easily constructed. For example (Example 3.5 in [6]), an FIR spectrum has positive real part given by Ψ ( e iω) = 1 M 1 2 r 0 + r k e iωk. (3.18) A controllable state-space realization of Ψ ( e iω) is ( ) O 1 M 2 0 ( ) T A =, B = I M 2 O M 2 1 ( ) C = r 1 r 2... r M 1, D = 1 2 r 0. (3.19) Therefore, in this example the constraint (3.14) can be written as an LMI in Q and r 1,..., r M 1. k= Partial correlation parametrization The partial correlation parametrization uses the finite expansion M 1 k= (M 1) c k B k ( e iω ) (3.20) in order to design only the first M coefficients c k. In this case, it is necessary to assure that there exists a sequence c M, c M+1,... such that the complete sequence { c k } k=0 defines a spectrum. That is, the condition (3.14) must hold. This means that (3.20) does not necessary define a spectrum itself, but the designed coefficients are extendable to a sequence that parametrizes a spectrum. As explained in [6], if an FIR spectrum is considered, a necessary and sufficient condition for (3.14) to hold is r 0 r 1... r M 1 r 1 r 0... r M (3.21) r M 1 r M 2... r 0

24 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 15 (see [24] and [25]). This condition also applies to more general basis functions, like B k ( e iω ) = L ( e iω) e iωk, where L ( e iω) > 0 [6, 12]. The constraint (3.21) is an LMI in the first M correlation coefficients and therefore is convex in these variables. Notice that (3.21) is less restrictive than the condition imposed in Lemma for the finite dimensional parametrization. Furthermore, as it will be discussed in the next section, the finite spectrum parametrization allows to handle spectral constraints on input and output signals, that the partial correlation parametrization cannot handle, since the parametrization (3.20) is not a spectrum. An advantage of the latter parametrization, though, is that it uses the minimum number of free parameters. 3.4 Covariance matrix parametrization By using one of the input spectrum parametrizations in the expression (2.7), the inverse of the asymptotic covariance matrix can be written as P 1 θ 0 = M 1 k= (M 1) c k Q k + Q, (3.22) where Q k = N π 2πλ 0 π F ( u e iω ) (, θ 0 Bk e iω ) Fu ( e iω ), θ 0 dω and Q = Re (θ 0 ). Then, P 1 θ 0 is expressed as a linear and finitely parametrized function of the coefficients c 0,..., c M 1 (since the symmetry condition c k = c k holds). Therefore, any quality constraint that is convex in P 1 θ 0 is also convex in c 0,..., c M 1. Some common quality constraints have been introduced in Section 3.2, which were all convex functions of P 1 θ Signal constraints parametrization Constraints on the input spectrum are also considered in input design. Typically, they are frequency by frequency or power constraints. A detailed discussion of the power constraints parametrization is presented in [6]. Briefly, consider power constraints of the type 1 2π 1 2π π π π π Wu ( e iω ) 2 Φ u (ω) dω α u (3.23) Wy ( e iω ) 2 Φ y (ω) dω α y. (3.24)

25 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 16 By using a finite spectrum parametrization these constraints can be written as convex finite-dimensional functions of { c k } M 1 k=0. For example, a constraint on the input power for an FIR spectrum becomes r 0 α u. For frequency by frequency constraints of the form β u (ω) Φ u (ω) γ u (ω) (3.25) β y (ω) Φ y (ω) γ y (ω), (3.26) Lemma can be applied to write them as convex constraints in { c k } M 1 k=0, upon the condition that the constraining functions are rational [6]. 3.6 Limitations of input design in the frequency domain The previous sections introduced input design problems where constraints on the signal spectra as well as on a measure of the estimate accuracy are considered. It has been shown that they can be formulated as finitely parametrized convex optimization problems, upon the condition that the measure of the estimate accuracy is a convex function of P 1 θ 0 and the input spectrum is parametrized as proposed in Section 3.3. Therefore, by solving a constrained optimization problem, the optimal variables c 0,..., c M 1 are found. The FIR spectrum representation is commonly used, so that the optimization procedure returns the first M terms of the correlation function. If a partial correlation parametrization is used, the optimal spectrum can be found by solving the Yule-Walker equations as described in [26]. From the optimal spectrum is then necessary to generate a signal in the time domain to apply to the real system. This is a realization problem that caracterizes solutions in the frequency domain. The input can be generated as filtered white noise, by spectral factorization of the optimal spectrum. Nevertheless, many probability distributions can be used to generate white noise. Also, it has to be noticed that in general the optimal spectrum is non unique and the input design approach so far considered only finds one of the optimal spectra. In fact, a finite dimensional spectrum parametrization forces the input correlation coefficients r M, r M+1,... to be zero; on the other hand, the partial correlation parametrization needs to complete the correlation sequence by solving Yule-Walker equations which give only one particular correlation sequence. Furthermore, in practical applications time domain constraints on the signals have to be considered and the power constraint that is usually set in

26 CHAPTER 3. INPUT DESIGN FOR SYSTEM IDENTIFICATION 17 the frequency domain does not assure that these constraints are respected. For these reasons, this thesis proposes to analyze the performance of an input design method in the probability domain, as it will be presented in the next chapters. 3.7 Conclusions Classical input design in the frequency domain has been presented. The advantage of this approach is that input design problems can be formulated as convex optimization problems. Some limitations of the method concern time domain constraints on signals and realization techniques.

27 Chapter 4 Markov Chain Input Model The drawbacks of input design in the frequency domain, presented in Section 3.6, suggest to study the possibility of a different approach. This chapter will introduce the idea of input design in the probability domain. In particular, reasons and advantages of modeling the input signal as a Markov chain will be presented in Section 4.1. In Section 4.2 the Markov chain input model will be described in detail. 4.1 Introduction What in general is required for system identification in practical applications is an input signal in the time domain that guarantees a sufficiently accurate estimate of the system while respecting some amplitude constraints. As discussed above, input design in the frequency domain does not handle time domain constraints on signals. Another disadvantage is that it does not define how to generate the input signal to apply to the real system from the optimal spectrum. Furthermore, input design in the frequency domain does not use the degrees of freedom in the choice of the optimal spectrum, which is generally non unique. That approach, in fact, only finds an optimal solution that fits with the optimal correlation coefficients r 0,..., r M 1. All the other possible solutions are not considered. The idea of input design in the probability domain arises from this observation: a solution in the probability domain makes it easier to generate input trajectories to apply to the real system, by extracting samples from the optimal distribution. In this way, no spectral factorization or realization algorithm are required. Markov chains having finite state space could then be used in order to directly include the time domain amplitude constraints into the input 18

28 CHAPTER 4. MARKOV CHAIN INPUT MODEL 19 model, by suitably choosing the state space. The idea is to use Markov chain distributions to generate binary signals. Binary signals are often used in system identification and one of the reasons is that they achieve the largest power in the set of all signals having the same maximum amplitude and this improves parameter estimation for linear models. Also, if the Markov chain has spectrum of sufficiently high order (and this depends on the state space dimension), when designing the optimal probability distribution there are more degrees of freedom in the choice of the optimal spectrum. A finite stationary Markov chain is used as input model for system identification. The probability distribution will be shaped in order to optimize the objective function defined in the input design problem. This function is then minimized with respect to the transition probabilities of the Markov chain that completely define its distribution [27]. 4.2 Markov chains model This section describes the state space structure of the general Markov chain input model and some of its spectral features Markov chain state space Consider a finite stationary Markov chain having states of the form (u t n, u t n+1,..., u t ) (4.1) where u i represents the value of the input at the time instant i; it can be equal to either u max or u max, where u max is the maximum tolerable input amplitude, imposed by the real system. This model allows the present value of the input to depend on the last n past values, rather than only on the previous one. Note that at the time instant t, the state can transit only to either the state (u t n+1, u t n+2,..., u t, u max ) or (u t n+1, u t n+2,..., u t, u max ) with probabilities p (ut n,...,u t) and 1 p (ut n,...,u t), respectively. Not all the transitions between states are possible; therefore the transition matrix will present several zeros corresponding to those forbidden state transitions. The last component of the Markov chain state will generate the binary signal to apply to the real system. Example Consider a Markov chain having state space S 2 = {1, 1}. The graph representation is shown in Figure 4.1 and the corresponding tran-

29 CHAPTER 4. MARKOV CHAIN INPUT MODEL 20 sition matrix is Π 2 = ( p 1 p 1 q q ). (4.2) Figure 4.1: Graph representation of the two states Markov chain S 2. This simple model generates a binary signal where each sample at time t depends only on the previous value at time t 1. Example A more general model is the four states Markov chain, with state space S = {(1, 1), (1, 1), ( 1, 1), ( 1, 1)}. The transition matrix is Π = p 1 p s 1 s 0 0 q 1 q r 1 r 0 0 and the corresponding graph is shown in Figure 4.2. (4.3) Figure 4.2: Graph representation of the four states Markov chain S 4. Note that when p = r and s = q the four states Markov chain model is equivalent to the two states Markov chain of Example

30 CHAPTER 4. MARKOV CHAIN INPUT MODEL 21 These examples show one of the advantages of the proposed Markov chain input model: each model includes all the models of lower dimension as special cases, for proper choices of the transition probabilities Markov chains spectra This section presents a method to calculate the expression of the Markov chains spectra. Some examples will illustrate the type of signals these models generate. By means of Markov chains and state space realization theories (see [27] and [28]), it is possible to derive a general expression for the spectrum of a finite stationary Markov chain s m having state space S = {S 1, S 2,..., S J }. For the general Markov chains considered in the previous section, each state has the form (4.1) and the number of states is J = 2 n+1. Let Π denote the transition matrix whose elements are( the conditional ) probabilities Π (i, j) = P {s m+1 = S j s m = S i } and p = p 1... p J the solution of the linear system p = pπ, containing the stationary probabilities p i = P {s n = S i }. Consider the states S i as column vectors. Defining ( ) A s = S 1... S J and D s = p p J it is possible to write the correlation coefficients of the output signal in the matricial form r k = A s D s Π k A T s, k = 0, 1, 2... (4.4) For k < 0 the correlation can be obtained by the symmetry condition r k = r k, since the process s m is real. To calculate the spectrum of s m as the Fourier transform of the correlation function, note that r k can be viewed as the impulse response, for k = 1, 2,..., of the linear system x k+1 = Πx k + ΠA T s u k y k = A s D s x k. (4.5) Therefore, the transfer function W (z) = A s D s (zi Π) 1 ΠA T s of the system (4.5) is the Z-transform of the causal part of the correlation function, that is {r k } k=1.

31 CHAPTER 4. MARKOV CHAIN INPUT MODEL 22 Consequently, W ( z 1) is the Z-transform of the anticausal part of the correlation function, {r k } 1 k=. The spectrum of the Markov chain signal s m can then be expressed as Φ s (z) = W (z) + r 0 + W ( z 1), (4.6) The correlation r k is in general a matricial function, i.e. for each k, r k is an n n matrix. The correlation function of the input signal is given by the sequence obtained for the element in position (n, n): {r k (n, n)} k=0. Then, the input signal spectrum is given by the element in position (n, n) of the matrix Φ s (z). Consider now the two Markov chains in Examples and of the previous section. Example By calculating the transfer function W (z) and the autocorrelation r 0, the following expression for the spectrum of the Markov chain of Figure 4.1 is obtained: Φ s (z) = (1 + α) (1 γ) (1 + α) (1 γ) (z α) (z 1, (4.7) α) where α = p + q 1 [ 1, 1] and γ = (p+q 1)(p+q 2) (p q)2 (p+q 2). Notice that (4.7) is the spectrum of a first order AR process. This also means that it is possible to generate any first order AR process through a two states Markov chain. The mean value of s m is E [s m ] = p q 2 p q. By forcing p = q the mean value would be zero and since γ = α the spectrum would depend on only one parameter. If p = q the variance of the Markov chain is 1. Example Consider the Example of the previous section where the mean value of the Markov chain is set to zero. Then s = (1 q)r (1 p) and the stationary probabilities are p = r(1 q) 2(1 p)(1 q)+2r(1 q) (1 p)(1 q) 2(1 p)(1 q)+2r(1 q) r(1 q) 2(1 p)(1 q)+2r(1 q) (1 p)(1 q) 2(1 p)(1 q)+2r(1 q) ( ) By definition, A s =. However, only the second component of the Markov chain is of interest, since it will represent the T

32 CHAPTER 4. MARKOV CHAIN INPUT MODEL 23 input signal. Therefore, the ( correlation r k of the ) input process can be calculated using the vector Ās = instead of A s. The analytic calculation of the spectrum turns out to be too involved, so it has only been evaluated numerically. Some values of poles and zeros of the canonical spectral factor 1 are reported in Table 4.1. These data show that it is not possible to model s m by an AR process, because in general there are also non null zeros in the spectrum. The four states Markov chain has a higher order spectrum than the two states Markov chain of Example 4.2.3; the number of poles and zeros depends on the values of the probabilities p, r and q and can be up to eight. For some values of the probabilities p, r and q, there are zero-pole cancelations that reduce the spectrum order; in particular, when p = q = r the spectrum has the same simple structure obtained for the previous case (see Table 4.1), as it has already been shown in the previous section. A particular choice of the transition matrix for this Markov chain is p 1 p r 1 r Π 4 = (4.8) 0 0 p 1 p r 1 r 0 0 This choice of the transition matrix makes the Markov chain symmetric in the sense that the transition probabilities are invariant with respect to exchanges in the sign of the states components. Even if the input is designed in the probability domain, the spectrum is shaped by the choice of n and of the transition probabilities of the Markov chain. Notice that Φ u is only subject to (4.6) and no other structural constraints are imposed, except for the constraint that the transition probabilities have to lie in [0, 1]. For the spectrum parametrizations described in Section 3.3 this does not happen. For example, if the FIR representation of the spectrum is used, the finite dimensional parametrization forces the positive real part of the spectrum to be an FIR system. The Markov chain signals have spectrum where poles and zeros are related each other, since the number of free parameters in the problem is J (the transition probabilities), as the number of poles of W (z). However, the positive real part of the spectrum is not forced to be FIR. From these observations it is possible to conclude that looking at input design using Markov chains in the frequency domain, this 1 By canonical spectral factor is meant the analytic and minimum phase function L (z) such that Φ s (z) = L (z) L ( z 1).

33 CHAPTER 4. MARKOV CHAIN INPUT MODEL 24 p q r poles zeros i i i i i i i i i i i i Table 4.1: Poles and zeros of the canonical spectral factor of the spectrum of s m of Example

34 CHAPTER 4. MARKOV CHAIN INPUT MODEL 25 approach preserves more degrees of freedom for the choice of the optimal spectrum than the input design approach described in Chapter 3, since less constraints on the structure are imposed. In order to optimize an objective function for input design purposes by shaping the input process distribution, it is necessary to define the Markov chain state space and its transition probabilities. That is, once the input model structure is defined, the objective function is optimized with respect to the transition probabilities. For the examples considered in this section, the purpose of the input design problem would be to optimize the objective function J (u, θ 0 ) with respect to the transition probabilities, p in the first case, p and r in the second. 4.3 More general Markov chains Input signals generated by the Markov chains described in the previous sections are binary signals. It is also possible to extend this type of Markov models to more general input signals. For example, Markov chains with state space S = {0, 1, 1, 2, 2, 3, 3,...} would generate signals having more than two amplitude levels. As a simple example, consider a three states Markov chain generating a ternary input signal. The Markov chain has state space S 3 = {0, 1, 1} and transition graph in Figure 4.3. Figure 4.3: Graph representation of the three states Markov chain S 3. To easily calculate an expression for the spectrum, the mean value of the process is set to zero, so that p = q.

35 CHAPTER 4. MARKOV CHAIN INPUT MODEL 26 By solving p = pπ, the vector p = found. The spectrum results ( 1 r 3 2r p 1 p 3 2r p 1 r 3 2r p ) is Φ s (z) = 3 (r 1) (p 1) (1 + 3p) 2 (3 2r p) ( z 3p ) ( z 1 3p 1 2 ) This spectrum has the same structure as the one found for the two states Markov chain in Example In this case, Φ s depends on both r and p: p determines the pole and r the gain of the spectrum. The expression found for Φ s (z) shows that this input model does not provide a higher order spectrum than the two states Markov chain. In this case, it is the preferable to use a simpler model. This work focuses on input models that generate binary signals. The input models described in this section will not be further considered. 4.4 Conclusions This chapter defined finite stationary Markov chains generating binary signals. These processes have been introduced to model input signals for system identification purposes. The main advantages of using Markov chains as input models are that amplitude constraints are directly included into the input model and the input signal can be easily generated from the optimal distribution, avoiding the realization problem. Spectral properties of these processes have been also analyzed through some examples in Section

36 Chapter 5 Estimation Using Markov Chains This chapter proposes a method for parameter estimation of an LTI SISO system by using the Markov chain input model presented in Chapter 4. Section 5.1 defines the input design problem that will be studied here. The solution approach is presented in Section 5.2. The design method is described in detail in Sections 5.3 and 5.4. A numerical example, analyzed in Section 5.5, concludes the chapter. 5.1 Problem formulation Consider the system (2.2) and the parametric model (2.1) defined in Section 2.1. In this chapter, objective of the identification procedure is the estimation of the parameter vector θ. The input design problem considered in this study is to minimize a measure of the estimate error, f (P θ0 ), where f is a convex function of the covariance matrix of the parameter estimate. Often in practice, it is also necessary to take into account some constraints on the real signals. In that case, a general cost function can be considered J (u, θ 0 ) = f (P θ0 (u)) + g, (5.1) where g is a term which represents the cost of the experiment. As explained in Chapter 3, typical functions f are the trace, the determinant or the largest eigenvalue of the covariance matrix P θ0 [1]. This problem formulation is slightly different from the one presented in Chapter 3. No constraints are set in the optimization problem explicitly, but they are included into the cost function through the term g. The reason 27

37 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 28 for this is that the stochastic approximation framework is used to minimize the objective function, since no analytic convex formulation of the problem is available. This can be seen as a classical multiobjective optimization approach (an overview can be found in [29] and [30]). 5.2 Solution approach Since the analytic expressions for the covariance matrix P θ0 as a function of the transition probabilities of the Markov chain modeling the input are quite involved, simulation techniques are required to evaluate the cost function. The estimate of P θ0 is a stochastic function of the one-step transition probabilities and is contaminated with noise; therefore, it can only be evaluated through randomly generated input and noise signals. In this framework, stochastic approximation algorithms are needed to minimize the cost function (5.1) with respect to the transition probabilities of the Markov chain (see [15], [16] for details). An expression for P θ0 that suits in the stochastic approximation approach is found in Section 5.3; P θ0 is estimated as a function of input and noise data. The stochastic algorithm used to minimize the cost function (5.1) is described in Section Cost function evaluation From the model expression (2.1) it is possible to write e (t) = H (q, θ) 1 (y (t) G (q, θ) u (t)) and by linearizing the functions G (q, θ) and H (q, θ) at θ = θ 0 G (q, θ) G 0 (q) + G (q, θ) H (q, θ) H 0 (q) + H (q, θ), where G (q, θ) = (θ θ 0 ) T G(q,θ) θ, H (q, θ) = (θ θ 0 ) T H(q,θ) θ0 θ, the θ0 following expression is derived: e (t) = (H 0 (q) + H (q, θ)) 1 (H 0 (q) e 0 (t) G (q, θ) u (t)). (5.2) By substituting the Taylor expansion (H 0 (q) + H (q, θ)) 1 1 H 0 (q) H (q, θ) H 0 (q) 2

38 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 29 and the expressions of G (q, θ) and H (q, θ), it results ( e 0 (t) (θ θ 0 ) T 1 H (q, θ) H 0 (q) θ e 0 (t) + θ0 ) 1 G (q, θ) H 0 (q) θ u (t) + e (t). (5.3) θ0 The problem of estimating the parameter θ for the model (2.1) is asymptotically equivalent to solving the least squares problem for (5.3) where e 0 and u are known, when the number of data points N used for estimation goes to infinity. Therefore, the asymptotic expression (2.7) can be approximated as where S = ( w 1... w b ) R N b and w i R N 1 is the sequence ob- u (t) + 1 e 0 (t). θ0 θ0 tained from w it = 1 H 0 (q) G(q,θ) θ P 1 θ 0 = 1 λ 0 ( S T S ) H 0 (q) H(q,θ) θ Therefore, at each iteration of the algorithm, the cost function is evaluated using randomly generated input and noise signals. 5.4 Algorithm description When evaluating the cost function by simulation, it is necessary to consider that the cost function estimate is a stochastic variable that depends on the transition probabilities of the Markov chains and on the noise process e (t). Therefore, the cost function values generated through simulation have to be considered as samples of that stochastic variable. The true value of the cost function for a given transition probability would be the mean of that stochastic variable. For these reasons, stochastic approximation is necessary in order to minimize the cost function with respect to the transition probabilities of the Markov chain describing the input. One of the most common stochastic approximation methods that do not require the knowledge of the cost function gradient is the finite difference stochastic approximation (FDSA) [15]. It uses the recursion ˆp k+1 = ˆp k a k J k, (5.4) where J k is an estimate of the gradient of J at the k-th step and a k is a sequence such that lim k a k = 0. The FDSA estimates the gradient of the

39 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 30 cost function as J ki = J (ˆp k + c k e i ) J (ˆp k c k e i ) 2c k, where e i denotes the unit vector in the i-th direction, J ki is the i-th component of the gradient vector and c k is a sequence of coefficients converging to zero as k. Depending on the number d of parameters with respect to which minimize the cost function, a simultaneous perturbation stochastic approximation (SPSA) may be more efficient than the FDSA [31] ; when d increases the number of cost function evaluations in a FDSA procedure may be too large and the algorithm be very slow. In that case the SPSA algorithm described in [31] gives better performance, since it requires only two evaluations of the cost function regardless of d. In fact, SPSA estimates the gradient by J ki = J (ˆp k + c k k ) J (ˆp k c k k ) 2c k ki where k is a d-dimensional random perturbation vector, whose components are independently generated from a Bernoulli ±1 distribution with probability 0.5 for each outcome [32]. The iteration (5.4) is initialized by a first evaluation of the cost function on a discrete set of points and choosing the minimum in that set. At any point in this set, the cost function is evaluated only once; therefore, the value obtained is a sample extracted from the stochastic variable describing the cost function at that point. Therefore, it could turn out that the initial condition is not close to the true minimum of the cost function, due to noise in the measurements. Nevertheless, in some cases the result of the initialization procedure may be sufficiently accurate, so there could be no need to run many algorithm iterations. This of course will depend on the cost function shape and on the choice of the grid of points. a c (k+1) 1 /3, The sequences a k and c k can be chosen as a k = A+k+1 and c k = which are asymptotically optimal for the FDSA algorithm (see [15]). A method for choosing A, a and c may be to estimate the gradient of the cost function at the initial condition, so that the product a 0 J 0 has magnitude approximately equal to the expected changes among the elements of ˆp k in the early iterations [32]. The coefficient c (as suggested in [32]) ought to be greater than the variance of the noise in the cost function measurements in order to have a good estimate of the gradient. This variance may be estimated at the initial condition of the algorithm. An analytic proof of the algorithm convergence can be found in [15].

40 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 31 Figure 5.1: Mass-spring-damper system. 5.5 Numerical example Consider a mass-spring-damper system (Figure 5.1), where the input u is the force applied to the mass and the output y is the mass position. It is described by the transfer function G 0 (s) = 1 m s 2 + c m s + k m with m = 100 Kg, k = 10 N /m and c = Ns /m, resulting the natural frequency ω n = rad /s and the damping ξ = 0.1. The power here is defined as pw (t) = u (t) ẏ (t). White noise with variance λ 0 = is added at the output and an output-error model is used [19]. Data are sampled with T s = 1 s and the number of data points generated is N = As a measure of the estimate accuracy, the trace of the covariance matrix P θ0 is used. In order to consider also some practical constraints on the amplitude of the input and output signals and the maximum and mean input power, a general cost function will be used: J (u, θ 0 ) = f 1 (T rp θ0 (u)) + f 2 (u max ) (5.5) + f 3 (y max ) + f 4 (pw max ) + f 5 (pw mean ) where u max and y max are the absolute maximum values of the input and output signals, pw max and pw mean are the maximum and mean input power. Thresholds for T rp θ0, u max, y max, pw max and pw mean have been set, which define the maximum values allowed for each of these variables. Figure 5.2 and 5.3 show the cost functions f 2, f 3 and f 1, f 4, f 5, respectively: when the variables T rp θ0, u max, y max, pw max and pw mean reach their maximum

41 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 32 Figure 5.2: Cost functions f 2 and f 3 in the interval of acceptable values of the variables u max and y max, respectively. Figure 5.3: Cost functions f 1, f 4,f 5 in the interval of acceptable values.

42 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 33 acceptable value (100%), the cost is one. Outside the interval of acceptable values, the cost functions continue growing linearly. As input models, the two simple examples of Markov chains introduced in Section 4.2 are considered here. They are described by the transition matrices ( ) p 1 p Π 2 = 1 p p Π 4 = p 1 p r 1 r 0 0 p 1 p r 1 r 0 0. The set of points used for the algorithm initialization as explained in Section 5.4 is {0.1, 0.2, }. In the case analyzed here, since the cost function depends on not more than two parameters, FDSA is used. The algorithm coefficients have been chosen by the method suggested in the previous section. Three cases that have been studied: 1. The cost associated to T rp θ0 and the costs associated to the physical constraints have comparable values. 2. No power and amplitude constraints are considered. 3. Very strict power constraints are considered. These cases are summarized in 5.1. Case T rp θ0 u max y max pw max pw mean N 1 m 0.3 Nm s 0.03 Nm s Inf N Inf m Inf Nm s N 1 m 0.03 Nm s Inf Nm s Nm s Table 5.1: Maximum threshold values in the three analyzed cases. As a term of comparison for the performance of the Markov input model, a pseudo-random binary signal and white noise with unit variance (the same as the variance of the Markov chains) have been applied as inputs to the system. The results of the simulation runs for the three cases listed above are shown in Tables 5.4, 5.5 and 5.6. The cost function values are estimated by evaluating the average of 100 simulation runs using the optimal input found by the algorithm, the PRBS and white noise inputs. Table 5.5, related to case 2, shows the optimal value of the trace of the covariance matrix calculated by solving the LMI formulation of the input design problem in

43 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 34 the frequency domain, as explained in [12]. Furthermore, by the method described in [13], a binary signal having the optimal correlation function is generated. The minimum obtained with this input signal is also shown in Table 5.5. The second case, which is the most standard in input design problems, is first analyzed in detail. Figure 5.4 presents the cost function, estimated on a fine grid of points, as the average of 100 simulations. Table 5.2 exhibits the Figure 5.4: Estimate of the cost function on a discrete set of points for the two states Markov chain in the case 2. results of two Monte-Carlo simulations (each consisting of 100 runs), which show that the variance of the algorithm output decreases approximately as 1 N Iter, where N Iter is the number of algorithm iterations; this guarantees the empirical algorithm convergence. With iterations the algorithm N Iter Mean value Eˆp Variance Varˆp Table 5.2: Results of 100 Monte-Carlo simulations of the algorithm with the 2 states Markov chain. produces the results in Figure 5.5. The optimality of the probability ˆp found by the algorithm has been verified by using the expression of the two states

44 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 35 Markov chain spectrum in the asymptotic expression (2.7) and minimizing T rp θ0 with respect to α; it turns out that the optimal value ˆp = is very close to the one found by the stochastic algorithm after iterations, that is ˆp = (Table 5.3). This confirms that the stochastic algorithm converges to the true optimal value. In practice, it is not necessary to run the algorithm for iterations, since already at the initial condition the cost function is very close to the minimum and the variance of the estimate after iterations is of the order of It has been done here, anyway, to show that the final value obtained is the true optimal one. Figure 5.5: Estimation of the best transition probability for the two states Markov chain in the case 2. Case S 2 [ ] [ S 4 1 ˆp = ˆp [ ˆr = ] [ ˆp = ˆp [ ˆr = ] [ ˆp = ˆp ˆr = ] ] ] Table 5.3: Optimal values of the transition probabilities in the cases 1, 2 and 3, obtained after algorithm iterations. Notice from the results in Table 5.5 that the Markov chains give lower values of the trace of P θ0 (u) than all the other inputs, except the true optimal

45 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 36 spectrum. The frequencies of the optimal input spectrum for the case 2 S 2 S 4 PRBS WN J (u, θ 0 ) Table 5.4: Total cost function values obtained with the optimal Markov inputs, a PRBS and white noise in case 1. S 2 S 4 PRBS WN BI Optimum T rp θ0 1.43e e e e e e-8 Table 5.5: Trace of the covariance matrix obtained with the optimal Markov inputs, a PRBS, white noise, a binary input having the optimal correlation function and the optimal spectrum in case 2. S 2 S 4 PRBS WN J (u, ϑ 0 ) Table 5.6: Total cost function value obtained with the optimal Markov inputs, a PRBS and white noise in case 3. have been estimated by means of the Multiple Signal Classification (MUSIC) methods, described in [26]. It results that the optimal input consists of two sinusoids of frequencies rad /s and rad /s, respectively, where the main contribution is given by the sinusoid of high frequency, which has approximately 5.6 times the power of the first component. Note that these frequencies are very close to the natural frequency of the system and to the poles of the Markov chains spectra (Figures 5.10 and 5.11). Figures 5.6 and 5.7 show the trajectories of the probabilities estimates obtained for the four states Markov chain in case 2. The empirical speed of convergence is lower than for the two states Markov chain. Nevertheless, the cost function value does not change significantly if the algorithm is stopped after 2000 iterations. Case 1 analyzes the more practical situation in which amplitude and power constraints on signals have to be considered. In this case, the cost functions obtained for the two and the four states Markov chain are presented in Figures 5.8 and 5.9. Notice that despite the presence of noise, the cost function is convex; therefore, the problem has a solution.

46 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 37 Figure 5.6: Estimation of the best transition probability p for the four states Markov chain in the case 2. Figure 5.7: Estimation of the best transition probability r for the four states Markov chain in the case 2.

47 CHAPTER 5. ESTIMATION USING MARKOV CHAINS 38 Figure 5.8: Estimate of the cost function on a discrete set of points for the two states Markov chain in the case 1. Figure 5.9: Estimate of the cost function on a discrete set of points for the four states Markov chain in the case 1. Power constraints move the minimum of the cost function to smaller probability values; that is, the transition probabilities cannot be too large, otherwise the input excitation would not respect power constraints. In case 1, the Markov inputs and the PRBS signal give almost the same cost value

THIS paper studies the input design problem in system identification.

THIS paper studies the input design problem in system identification. 1534 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 50, NO. 10, OCTOBER 2005 Input Design Via LMIs Admitting Frequency-Wise Model Specifications in Confidence Regions Henrik Jansson Håkan Hjalmarsson, Member,

More information

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Prediction Error Methods - Torsten Söderström

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Prediction Error Methods - Torsten Söderström PREDICTIO ERROR METHODS Torsten Söderström Department of Systems and Control, Information Technology, Uppsala University, Uppsala, Sweden Keywords: prediction error method, optimal prediction, identifiability,

More information

On Identification of Cascade Systems 1

On Identification of Cascade Systems 1 On Identification of Cascade Systems 1 Bo Wahlberg Håkan Hjalmarsson Jonas Mårtensson Automatic Control and ACCESS, School of Electrical Engineering, KTH, SE-100 44 Stockholm, Sweden. (bo.wahlberg@ee.kth.se

More information

EL1820 Modeling of Dynamical Systems

EL1820 Modeling of Dynamical Systems EL1820 Modeling of Dynamical Systems Lecture 10 - System identification as a model building tool Experiment design Examination and prefiltering of data Model structure selection Model validation Lecture

More information

Optimal input design for nonlinear dynamical systems: a graph-theory approach

Optimal input design for nonlinear dynamical systems: a graph-theory approach Optimal input design for nonlinear dynamical systems: a graph-theory approach Patricio E. Valenzuela Department of Automatic Control and ACCESS Linnaeus Centre KTH Royal Institute of Technology, Stockholm,

More information

Identification of ARX, OE, FIR models with the least squares method

Identification of ARX, OE, FIR models with the least squares method Identification of ARX, OE, FIR models with the least squares method CHEM-E7145 Advanced Process Control Methods Lecture 2 Contents Identification of ARX model with the least squares minimizing the equation

More information

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece.

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece. Machine Learning A Bayesian and Optimization Perspective Academic Press, 2015 Sergios Theodoridis 1 1 Dept. of Informatics and Telecommunications, National and Kapodistrian University of Athens, Athens,

More information

EL1820 Modeling of Dynamical Systems

EL1820 Modeling of Dynamical Systems EL1820 Modeling of Dynamical Systems Lecture 9 - Parameter estimation in linear models Model structures Parameter estimation via prediction error minimization Properties of the estimate: bias and variance

More information

Further Results on Model Structure Validation for Closed Loop System Identification

Further Results on Model Structure Validation for Closed Loop System Identification Advances in Wireless Communications and etworks 7; 3(5: 57-66 http://www.sciencepublishinggroup.com/j/awcn doi:.648/j.awcn.735. Further esults on Model Structure Validation for Closed Loop System Identification

More information

I. D. Landau, A. Karimi: A Course on Adaptive Control Adaptive Control. Part 9: Adaptive Control with Multiple Models and Switching

I. D. Landau, A. Karimi: A Course on Adaptive Control Adaptive Control. Part 9: Adaptive Control with Multiple Models and Switching I. D. Landau, A. Karimi: A Course on Adaptive Control - 5 1 Adaptive Control Part 9: Adaptive Control with Multiple Models and Switching I. D. Landau, A. Karimi: A Course on Adaptive Control - 5 2 Outline

More information

y k = ( ) x k + v k. w q wk i 0 0 wk

y k = ( ) x k + v k. w q wk i 0 0 wk Four telling examples of Kalman Filters Example : Signal plus noise Measurement of a bandpass signal, center frequency.2 rad/sec buried in highpass noise. Dig out the quadrature part of the signal while

More information

EECE Adaptive Control

EECE Adaptive Control EECE 574 - Adaptive Control Recursive Identification in Closed-Loop and Adaptive Control Guy Dumont Department of Electrical and Computer Engineering University of British Columbia January 2010 Guy Dumont

More information

On Input Design in System Identification for Control MÄRTA BARENTHIN

On Input Design in System Identification for Control MÄRTA BARENTHIN On Input Design in System Identification for Control MÄRTA BARENTHIN Licentiate Thesis Stockholm, Sweden 2006 TRITA-EE 2006:023 ISSN 1653-5146 ISBN 91-7178-400-4 KTH School of Electrical Engineering SE-100

More information

Experiment design for batch-to-batch model-based learning control

Experiment design for batch-to-batch model-based learning control Experiment design for batch-to-batch model-based learning control Marco Forgione, Xavier Bombois and Paul M.J. Van den Hof Abstract An Experiment Design framewor for dynamical systems which execute multiple

More information

MODEL PREDICTIVE CONTROL and optimization

MODEL PREDICTIVE CONTROL and optimization MODEL PREDICTIVE CONTROL and optimization Lecture notes Model Predictive Control PhD., Associate professor David Di Ruscio System and Control Engineering Department of Technology Telemark University College

More information

Advanced Digital Signal Processing -Introduction

Advanced Digital Signal Processing -Introduction Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary

More information

Data mining for system identi cation

Data mining for system identi cation LERTEKNIK REG AU T O MA RO T IC C O N T L applications to process André Carvalho Bittencourt Automatic Control, Linköping Sweden Outline 1. Problem formulation. Theoretical guiding principles Modeling

More information

Analysis of Discrete-Time Systems

Analysis of Discrete-Time Systems TU Berlin Discrete-Time Control Systems 1 Analysis of Discrete-Time Systems Overview Stability Sensitivity and Robustness Controllability, Reachability, Observability, and Detectabiliy TU Berlin Discrete-Time

More information

Advanced Process Control Tutorial Problem Set 2 Development of Control Relevant Models through System Identification

Advanced Process Control Tutorial Problem Set 2 Development of Control Relevant Models through System Identification Advanced Process Control Tutorial Problem Set 2 Development of Control Relevant Models through System Identification 1. Consider the time series x(k) = β 1 + β 2 k + w(k) where β 1 and β 2 are known constants

More information

On Application Oriented Experiment Design for Closed-loop System Identification AFROOZ EBADAT

On Application Oriented Experiment Design for Closed-loop System Identification AFROOZ EBADAT On Application Oriented Experiment Design for Closed-loop System Identification AFROOZ EBADAT Licentiate Thesis Stockholm, Sweden 2015 TRITA-EE 2015:002 ISSN 1653-5146 ISBN 978-91-7595-410-3 KTH Royal

More information

Matlab software tools for model identification and data analysis 11/12/2015 Prof. Marcello Farina

Matlab software tools for model identification and data analysis 11/12/2015 Prof. Marcello Farina Matlab software tools for model identification and data analysis 11/12/2015 Prof. Marcello Farina Model Identification and Data Analysis (Academic year 2015-2016) Prof. Sergio Bittanti Outline Data generation

More information

MOOSE: Model Based Optimal Input Design Toolbox for Matlab

MOOSE: Model Based Optimal Input Design Toolbox for Matlab MOOSE: Model Based Optimal Input Design Toolbox for Matlab Technical Report Mariette Annergren mariette.annergren@ee.kth.se Christian Larsson christian.larsson@ee.kth.se September 23, 2011 Contents 1 Introduction

More information

Properties of Zero-Free Spectral Matrices Brian D. O. Anderson, Life Fellow, IEEE, and Manfred Deistler, Fellow, IEEE

Properties of Zero-Free Spectral Matrices Brian D. O. Anderson, Life Fellow, IEEE, and Manfred Deistler, Fellow, IEEE IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL 54, NO 10, OCTOBER 2009 2365 Properties of Zero-Free Spectral Matrices Brian D O Anderson, Life Fellow, IEEE, and Manfred Deistler, Fellow, IEEE Abstract In

More information

Automatic Control Systems theory overview (discrete time systems)

Automatic Control Systems theory overview (discrete time systems) Automatic Control Systems theory overview (discrete time systems) Prof. Luca Bascetta (luca.bascetta@polimi.it) Politecnico di Milano Dipartimento di Elettronica, Informazione e Bioingegneria Motivations

More information

Analysis of Discrete-Time Systems

Analysis of Discrete-Time Systems TU Berlin Discrete-Time Control Systems TU Berlin Discrete-Time Control Systems 2 Stability Definitions We define stability first with respect to changes in the initial conditions Analysis of Discrete-Time

More information

Statistical and Adaptive Signal Processing

Statistical and Adaptive Signal Processing r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory

More information

OPTIMAL EXPERIMENT DESIGN IN CLOSED LOOP. KTH, Signals, Sensors and Systems, S Stockholm, Sweden.

OPTIMAL EXPERIMENT DESIGN IN CLOSED LOOP. KTH, Signals, Sensors and Systems, S Stockholm, Sweden. OPTIMAL EXPERIMENT DESIGN IN CLOSED LOOP Henrik Jansson Håkan Hjalmarsson KTH, Signals, Sensors and Systems, S-00 44 Stockholm, Sweden. henrik.jansson@s3.kth.se Abstract: In this contribution we extend

More information

Let the plant and controller be described as:-

Let the plant and controller be described as:- Summary of Fundamental Limitations in Feedback Design (LTI SISO Systems) From Chapter 6 of A FIRST GRADUATE COURSE IN FEEDBACK CONTROL By J. S. Freudenberg (Winter 2008) Prepared by: Hammad Munawar (Institute

More information

Andrea Zanchettin Automatic Control AUTOMATIC CONTROL. Andrea M. Zanchettin, PhD Spring Semester, Linear systems (frequency domain)

Andrea Zanchettin Automatic Control AUTOMATIC CONTROL. Andrea M. Zanchettin, PhD Spring Semester, Linear systems (frequency domain) 1 AUTOMATIC CONTROL Andrea M. Zanchettin, PhD Spring Semester, 2018 Linear systems (frequency domain) 2 Motivations Consider an LTI system Thanks to the Lagrange s formula we can compute the motion of

More information

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. ISS0031 Modeling and Identification Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. Aleksei Tepljakov, Ph.D. October 21, 2015 Discrete-time Transfer Functions

More information

Expressions for the covariance matrix of covariance data

Expressions for the covariance matrix of covariance data Expressions for the covariance matrix of covariance data Torsten Söderström Division of Systems and Control, Department of Information Technology, Uppsala University, P O Box 337, SE-7505 Uppsala, Sweden

More information

Long-Run Covariability

Long-Run Covariability Long-Run Covariability Ulrich K. Müller and Mark W. Watson Princeton University October 2016 Motivation Study the long-run covariability/relationship between economic variables great ratios, long-run Phillips

More information

f-domain expression for the limit model Combine: 5.12 Approximate Modelling What can be said about H(q, θ) G(q, θ ) H(q, θ ) with

f-domain expression for the limit model Combine: 5.12 Approximate Modelling What can be said about H(q, θ) G(q, θ ) H(q, θ ) with 5.2 Approximate Modelling What can be said about if S / M, and even G / G? G(q, ) H(q, ) f-domain expression for the limit model Combine: with ε(t, ) =H(q, ) [y(t) G(q, )u(t)] y(t) =G (q)u(t) v(t) We know

More information

Exploring Granger Causality for Time series via Wald Test on Estimated Models with Guaranteed Stability

Exploring Granger Causality for Time series via Wald Test on Estimated Models with Guaranteed Stability Exploring Granger Causality for Time series via Wald Test on Estimated Models with Guaranteed Stability Nuntanut Raksasri Jitkomut Songsiri Department of Electrical Engineering, Faculty of Engineering,

More information

LMI Based Model Order Reduction Considering the Minimum Phase Characteristic of the System

LMI Based Model Order Reduction Considering the Minimum Phase Characteristic of the System LMI Based Model Order Reduction Considering the Minimum Phase Characteristic of the System Gholamreza Khademi, Haniyeh Mohammadi, and Maryam Dehghani School of Electrical and Computer Engineering Shiraz

More information

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω ECO 513 Spring 2015 TAKEHOME FINAL EXAM (1) Suppose the univariate stochastic process y is ARMA(2,2) of the following form: y t = 1.6974y t 1.9604y t 2 + ε t 1.6628ε t 1 +.9216ε t 2, (1) where ε is i.i.d.

More information

Properties of Open-Loop Controllers

Properties of Open-Loop Controllers Properties of Open-Loop Controllers Sven Laur University of Tarty 1 Basics of Open-Loop Controller Design Two most common tasks in controller design is regulation and signal tracking. Regulating controllers

More information

Control Systems I. Lecture 6: Poles and Zeros. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Control Systems I. Lecture 6: Poles and Zeros. Readings: Emilio Frazzoli. Institute for Dynamic Systems and Control D-MAVT ETH Zürich Control Systems I Lecture 6: Poles and Zeros Readings: Emilio Frazzoli Institute for Dynamic Systems and Control D-MAVT ETH Zürich October 27, 2017 E. Frazzoli (ETH) Lecture 6: Control Systems I 27/10/2017

More information

EEG- Signal Processing

EEG- Signal Processing Fatemeh Hadaeghi EEG- Signal Processing Lecture Notes for BSP, Chapter 5 Master Program Data Engineering 1 5 Introduction The complex patterns of neural activity, both in presence and absence of external

More information

Theory and Problems of Signals and Systems

Theory and Problems of Signals and Systems SCHAUM'S OUTLINES OF Theory and Problems of Signals and Systems HWEI P. HSU is Professor of Electrical Engineering at Fairleigh Dickinson University. He received his B.S. from National Taiwan University

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted

More information

Time Series: Theory and Methods

Time Series: Theory and Methods Peter J. Brockwell Richard A. Davis Time Series: Theory and Methods Second Edition With 124 Illustrations Springer Contents Preface to the Second Edition Preface to the First Edition vn ix CHAPTER 1 Stationary

More information

EECE Adaptive Control

EECE Adaptive Control EECE 574 - Adaptive Control Basics of System Identification Guy Dumont Department of Electrical and Computer Engineering University of British Columbia January 2010 Guy Dumont (UBC) EECE574 - Basics of

More information

A Mathematica Toolbox for Signals, Models and Identification

A Mathematica Toolbox for Signals, Models and Identification The International Federation of Automatic Control A Mathematica Toolbox for Signals, Models and Identification Håkan Hjalmarsson Jonas Sjöberg ACCESS Linnaeus Center, Electrical Engineering, KTH Royal

More information

FIR Filters for Stationary State Space Signal Models

FIR Filters for Stationary State Space Signal Models Proceedings of the 17th World Congress The International Federation of Automatic Control FIR Filters for Stationary State Space Signal Models Jung Hun Park Wook Hyun Kwon School of Electrical Engineering

More information

Stability of Parameter Adaptation Algorithms. Big picture

Stability of Parameter Adaptation Algorithms. Big picture ME5895, UConn, Fall 215 Prof. Xu Chen Big picture For ˆθ (k + 1) = ˆθ (k) + [correction term] we haven t talked about whether ˆθ(k) will converge to the true value θ if k. We haven t even talked about

More information

Control Systems I. Lecture 7: Feedback and the Root Locus method. Readings: Jacopo Tani. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Control Systems I. Lecture 7: Feedback and the Root Locus method. Readings: Jacopo Tani. Institute for Dynamic Systems and Control D-MAVT ETH Zürich Control Systems I Lecture 7: Feedback and the Root Locus method Readings: Jacopo Tani Institute for Dynamic Systems and Control D-MAVT ETH Zürich November 2, 2018 J. Tani, E. Frazzoli (ETH) Lecture 7:

More information

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036

More information

Frequency methods for the analysis of feedback systems. Lecture 6. Loop analysis of feedback systems. Nyquist approach to study stability

Frequency methods for the analysis of feedback systems. Lecture 6. Loop analysis of feedback systems. Nyquist approach to study stability Lecture 6. Loop analysis of feedback systems 1. Motivation 2. Graphical representation of frequency response: Bode and Nyquist curves 3. Nyquist stability theorem 4. Stability margins Frequency methods

More information

12. Prediction Error Methods (PEM)

12. Prediction Error Methods (PEM) 12. Prediction Error Methods (PEM) EE531 (Semester II, 2010) description optimal prediction Kalman filter statistical results computational aspects 12-1 Description idea: determine the model parameter

More information

ESC794: Special Topics: Model Predictive Control

ESC794: Special Topics: Model Predictive Control ESC794: Special Topics: Model Predictive Control Discrete-Time Systems Hanz Richter, Professor Mechanical Engineering Department Cleveland State University Discrete-Time vs. Sampled-Data Systems A continuous-time

More information

Elements of Multivariate Time Series Analysis

Elements of Multivariate Time Series Analysis Gregory C. Reinsel Elements of Multivariate Time Series Analysis Second Edition With 14 Figures Springer Contents Preface to the Second Edition Preface to the First Edition vii ix 1. Vector Time Series

More information

Closed and Open Loop Subspace System Identification of the Kalman Filter

Closed and Open Loop Subspace System Identification of the Kalman Filter Modeling, Identification and Control, Vol 30, No 2, 2009, pp 71 86, ISSN 1890 1328 Closed and Open Loop Subspace System Identification of the Kalman Filter David Di Ruscio Telemark University College,

More information

Least costly probing signal design for power system mode estimation

Least costly probing signal design for power system mode estimation 1 Least costly probing signal design for power system mode estimation Vedran S. Perić, Xavier Bombois, Luigi Vanfretti KTH Royal Institute of Technology, Stockholm, Sweden NASPI Meeting, March 23, 2015.

More information

Adaptive Dual Control

Adaptive Dual Control Adaptive Dual Control Björn Wittenmark Department of Automatic Control, Lund Institute of Technology Box 118, S-221 00 Lund, Sweden email: bjorn@control.lth.se Keywords: Dual control, stochastic control,

More information

On Moving Average Parameter Estimation

On Moving Average Parameter Estimation On Moving Average Parameter Estimation Niclas Sandgren and Petre Stoica Contact information: niclas.sandgren@it.uu.se, tel: +46 8 473392 Abstract Estimation of the autoregressive moving average (ARMA)

More information

Chapter 7 Interconnected Systems and Feedback: Well-Posedness, Stability, and Performance 7. Introduction Feedback control is a powerful approach to o

Chapter 7 Interconnected Systems and Feedback: Well-Posedness, Stability, and Performance 7. Introduction Feedback control is a powerful approach to o Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology c Chapter 7 Interconnected

More information

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi Signal Modeling Techniques in Speech Recognition Hassan A. Kingravi Outline Introduction Spectral Shaping Spectral Analysis Parameter Transforms Statistical Modeling Discussion Conclusions 1: Introduction

More information

Time Series Examples Sheet

Time Series Examples Sheet Lent Term 2001 Richard Weber Time Series Examples Sheet This is the examples sheet for the M. Phil. course in Time Series. A copy can be found at: http://www.statslab.cam.ac.uk/~rrw1/timeseries/ Throughout,

More information

Problem Set 4 Solution 1

Problem Set 4 Solution 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.245: MULTIVARIABLE CONTROL SYSTEMS by A. Megretski Problem Set 4 Solution Problem 4. For the SISO feedback

More information

A test for improved forecasting performance at higher lead times

A test for improved forecasting performance at higher lead times A test for improved forecasting performance at higher lead times John Haywood and Granville Tunnicliffe Wilson September 3 Abstract Tiao and Xu (1993) proposed a test of whether a time series model, estimated

More information

ADAPTIVE FILTER THEORY

ADAPTIVE FILTER THEORY ADAPTIVE FILTER THEORY Fourth Edition Simon Haykin Communications Research Laboratory McMaster University Hamilton, Ontario, Canada Front ice Hall PRENTICE HALL Upper Saddle River, New Jersey 07458 Preface

More information

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver Stochastic Signals Overview Definitions Second order statistics Stationarity and ergodicity Random signal variability Power spectral density Linear systems with stationary inputs Random signal memory Correlation

More information

Design Methods for Control Systems

Design Methods for Control Systems Design Methods for Control Systems Maarten Steinbuch TU/e Gjerrit Meinsma UT Dutch Institute of Systems and Control Winter term 2002-2003 Schedule November 25 MSt December 2 MSt Homework # 1 December 9

More information

IDENTIFICATION FOR CONTROL

IDENTIFICATION FOR CONTROL IDENTIFICATION FOR CONTROL Raymond A. de Callafon, University of California San Diego, USA Paul M.J. Van den Hof, Delft University of Technology, the Netherlands Keywords: Controller, Closed loop model,

More information

Let H(z) = P(z)/Q(z) be the system function of a rational form. Let us represent both P(z) and Q(z) as polynomials of z (not z -1 )

Let H(z) = P(z)/Q(z) be the system function of a rational form. Let us represent both P(z) and Q(z) as polynomials of z (not z -1 ) Review: Poles and Zeros of Fractional Form Let H() = P()/Q() be the system function of a rational form. Let us represent both P() and Q() as polynomials of (not - ) Then Poles: the roots of Q()=0 Zeros:

More information

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied Linear Signal Models Overview Introduction Linear nonparametric vs. parametric models Equivalent representations Spectral flatness measure PZ vs. ARMA models Wold decomposition Introduction Many researchers

More information

14 - Gaussian Stochastic Processes

14 - Gaussian Stochastic Processes 14-1 Gaussian Stochastic Processes S. Lall, Stanford 211.2.24.1 14 - Gaussian Stochastic Processes Linear systems driven by IID noise Evolution of mean and covariance Example: mass-spring system Steady-state

More information

Robust Loop Shaping Controller Design for Spectral Models by Quadratic Programming

Robust Loop Shaping Controller Design for Spectral Models by Quadratic Programming Robust Loop Shaping Controller Design for Spectral Models by Quadratic Programming Gorka Galdos, Alireza Karimi and Roland Longchamp Abstract A quadratic programming approach is proposed to tune fixed-order

More information

6.435, System Identification

6.435, System Identification System Identification 6.435 SET 3 Nonparametric Identification Munther A. Dahleh 1 Nonparametric Methods for System ID Time domain methods Impulse response Step response Correlation analysis / time Frequency

More information

Optimal Polynomial Control for Discrete-Time Systems

Optimal Polynomial Control for Discrete-Time Systems 1 Optimal Polynomial Control for Discrete-Time Systems Prof Guy Beale Electrical and Computer Engineering Department George Mason University Fairfax, Virginia Correspondence concerning this paper should

More information

LTI Systems, Additive Noise, and Order Estimation

LTI Systems, Additive Noise, and Order Estimation LTI Systems, Additive oise, and Order Estimation Soosan Beheshti, Munther A. Dahleh Laboratory for Information and Decision Systems Department of Electrical Engineering and Computer Science Massachusetts

More information

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015 EC402: Serial Correlation Danny Quah Economics Department, LSE Lent 2015 OUTLINE 1. Stationarity 1.1 Covariance stationarity 1.2 Explicit Models. Special cases: ARMA processes 2. Some complex numbers.

More information

Spectral Analysis. Jesús Fernández-Villaverde University of Pennsylvania

Spectral Analysis. Jesús Fernández-Villaverde University of Pennsylvania Spectral Analysis Jesús Fernández-Villaverde University of Pennsylvania 1 Why Spectral Analysis? We want to develop a theory to obtain the business cycle properties of the data. Burns and Mitchell (1946).

More information

Vector Auto-Regressive Models

Vector Auto-Regressive Models Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

A New Subspace Identification Method for Open and Closed Loop Data

A New Subspace Identification Method for Open and Closed Loop Data A New Subspace Identification Method for Open and Closed Loop Data Magnus Jansson July 2005 IR S3 SB 0524 IFAC World Congress 2005 ROYAL INSTITUTE OF TECHNOLOGY Department of Signals, Sensors & Systems

More information

Multiobjective Optimization Applied to Robust H 2 /H State-feedback Control Synthesis

Multiobjective Optimization Applied to Robust H 2 /H State-feedback Control Synthesis Multiobjective Optimization Applied to Robust H 2 /H State-feedback Control Synthesis Eduardo N. Gonçalves, Reinaldo M. Palhares, and Ricardo H. C. Takahashi Abstract This paper presents an algorithm for

More information

Switching Regime Estimation

Switching Regime Estimation Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms

More information

3. ARMA Modeling. Now: Important class of stationary processes

3. ARMA Modeling. Now: Important class of stationary processes 3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average

More information

M E M O R A N D U M. Faculty Senate approved November 1, 2018

M E M O R A N D U M. Faculty Senate approved November 1, 2018 M E M O R A N D U M Faculty Senate approved November 1, 2018 TO: FROM: Deans and Chairs Becky Bitter, Sr. Assistant Registrar DATE: October 23, 2018 SUBJECT: Minor Change Bulletin No. 5 The courses listed

More information

VAR Models and Applications

VAR Models and Applications VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Frequency Domain and Filtering

Frequency Domain and Filtering 3 Frequency Domain and Filtering This is page i Printer: Opaque this 3. Introduction Comovement and volatility are key concepts in macroeconomics. Therefore, it is important to have statistics that describe

More information

Lecture 7 Open-loop & closedloop experiments. The Bias Formula goes closed-loop and the returns

Lecture 7 Open-loop & closedloop experiments. The Bias Formula goes closed-loop and the returns Lecture 7 Open-loop & closedloop experiments The Bias Formula goes closed-loop and the returns Closed-loop vs open-loop Open-loop system disturbance Feedback-free system input output Control input independent

More information

Analysis and Synthesis of Single-Input Single-Output Control Systems

Analysis and Synthesis of Single-Input Single-Output Control Systems Lino Guzzella Analysis and Synthesis of Single-Input Single-Output Control Systems l+kja» \Uja>)W2(ja»\ um Contents 1 Definitions and Problem Formulations 1 1.1 Introduction 1 1.2 Definitions 1 1.2.1 Systems

More information

RECURSIVE SUBSPACE IDENTIFICATION IN THE LEAST SQUARES FRAMEWORK

RECURSIVE SUBSPACE IDENTIFICATION IN THE LEAST SQUARES FRAMEWORK RECURSIVE SUBSPACE IDENTIFICATION IN THE LEAST SQUARES FRAMEWORK TRNKA PAVEL AND HAVLENA VLADIMÍR Dept of Control Engineering, Czech Technical University, Technická 2, 166 27 Praha, Czech Republic mail:

More information

Modeling and Identification of Dynamic Systems (vimmd312, 2018)

Modeling and Identification of Dynamic Systems (vimmd312, 2018) Modeling and Identification of Dynamic Systems (vimmd312, 2018) Textbook background of the curriculum taken. In parenthesis: material not reviewed due to time shortage, but which is suggested to be read

More information

Learn2Control Laboratory

Learn2Control Laboratory Learn2Control Laboratory Version 3.2 Summer Term 2014 1 This Script is for use in the scope of the Process Control lab. It is in no way claimed to be in any scientific way complete or unique. Errors should

More information

6.435, System Identification

6.435, System Identification SET 6 System Identification 6.435 Parametrized model structures One-step predictor Identifiability Munther A. Dahleh 1 Models of LTI Systems A complete model u = input y = output e = noise (with PDF).

More information

Part 2 Introduction to Microlocal Analysis

Part 2 Introduction to Microlocal Analysis Part 2 Introduction to Microlocal Analysis Birsen Yazıcı & Venky Krishnan Rensselaer Polytechnic Institute Electrical, Computer and Systems Engineering August 2 nd, 2010 Outline PART II Pseudodifferential

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides

More information

Lecture 9 Infinite Impulse Response Filters

Lecture 9 Infinite Impulse Response Filters Lecture 9 Infinite Impulse Response Filters Outline 9 Infinite Impulse Response Filters 9 First-Order Low-Pass Filter 93 IIR Filter Design 5 93 CT Butterworth filter design 5 93 Bilinear transform 7 9

More information

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems

On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. xx, No. x, Xxxxxxx 00x, pp. xxx xxx ISSN 0364-765X EISSN 156-5471 0x xx0x 0xxx informs DOI 10.187/moor.xxxx.xxxx c 00x INFORMS On the Power of Robust Solutions in

More information

Efficient Monte Carlo computation of Fisher information matrix using prior information

Efficient Monte Carlo computation of Fisher information matrix using prior information Efficient Monte Carlo computation of Fisher information matrix using prior information Sonjoy Das, UB James C. Spall, APL/JHU Roger Ghanem, USC SIAM Conference on Data Mining Anaheim, California, USA April

More information

MOOSE2: Model Based Optimal Input Design Toolbox for MATLAB R (version 2)

MOOSE2: Model Based Optimal Input Design Toolbox for MATLAB R (version 2) MOOSE2: Model Based Optimal Input Design Toolbox for MATLAB R (version 2) User s Guide Mariette Annergren mariette.annergren@ee.kth.se Christian Larsson christian.y.larsson@scania.com April 15, 2016 Contents

More information

Part 2 Introduction to Microlocal Analysis

Part 2 Introduction to Microlocal Analysis Part 2 Introduction to Microlocal Analysis Birsen Yazıcı& Venky Krishnan Rensselaer Polytechnic Institute Electrical, Computer and Systems Engineering March 15 th, 2010 Outline PART II Pseudodifferential(ψDOs)

More information

GATE EE Topic wise Questions SIGNALS & SYSTEMS

GATE EE Topic wise Questions SIGNALS & SYSTEMS www.gatehelp.com GATE EE Topic wise Questions YEAR 010 ONE MARK Question. 1 For the system /( s + 1), the approximate time taken for a step response to reach 98% of the final value is (A) 1 s (B) s (C)

More information

An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic

An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic Chapter 6 ESTIMATION OF THE LONG-RUN COVARIANCE MATRIX An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic standard errors for the OLS and linear IV estimators presented

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector

More information

A NUMERICAL METHOD TO SOLVE A QUADRATIC CONSTRAINED MAXIMIZATION

A NUMERICAL METHOD TO SOLVE A QUADRATIC CONSTRAINED MAXIMIZATION A NUMERICAL METHOD TO SOLVE A QUADRATIC CONSTRAINED MAXIMIZATION ALIREZA ESNA ASHARI, RAMINE NIKOUKHAH, AND STEPHEN L. CAMPBELL Abstract. The problem of maximizing a quadratic function subject to an ellipsoidal

More information

7. MULTIVARATE STATIONARY PROCESSES

7. MULTIVARATE STATIONARY PROCESSES 7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability

More information