Design and Implementation of Predictive Controller with KHM Clustering Algorithm

Size: px
Start display at page:

Download "Design and Implementation of Predictive Controller with KHM Clustering Algorithm"

Transcription

1 Journal of Computational Information Systems 9: ) Available at Design and Implementation of Predictive Controller with KHM Clustering Algorithm Jing ZENG 1,, Jun WANG 2 1 College of Information Engineering, Shenyang University of Chemical Technology, Shenyang , China 2 College of Computer Science and Technology, Shenyang University of Chemical Technology, Shenyang , China Abstract A KHM multi-model predictive control method is presented based on data-driven for a sort of unknownstructure nonlinear system based on vast history data only. A database search strategy based on K- Harmonic Means is given, hard cluster is softening. The similar cluster is selected, neighborhood and model parameters are determined on line according to the change of system operation point. This method can shorten searching time and improve precision of the data which used to get optimal local model. A multiple generalized predictive control strategy based on KHM is developed. The simulation test has stated the validity of this approach. Keywords: Clustering Algorithm; Fitting; Local Polynomial Regression; Predictive Control 1 Introduction Nonlinear system modeling and identification is very complicated, when there are a lot of observation data, determining model structure and related optimal problem will become more complex, it is difficult to get accurate global model for actual industrial process [1]. Many scholars modeling online using the present sampling data based on data driven [2-4], and it s generally believed that local polynomial fitting algorithm can achieve good control effect. Clustering analysis is one of the basic technologies of data mining, statistical analysis, machine learning and many other fields of application. In the study of clustering problem, k-means KM) clustering algorithm became one of the universal clustering algorithms because of its rapid convergence performance [5]. However, KM algorithm is very sensitive to the system initial value, different initial value can lead to different clustering results. K-Harmonic Means KHM) clustering algorithm can make up for the defect [6-9], this algorithm has the characteristic that be not sensitive to the initial value and advantages of good clustering performance. Generalized predictive control algorithm is a kind of advanced control algorithm which can overcome the system lag Corresponding author. address: zengjing0066@sohu.com Jing ZENG) / Copyright 2013 Binary Information Press DOI: /jcis6421 July 15, 2013

2 5620 J. Zeng et al. /Journal of Computational Information Systems 9: ) and deal with open loop unstable non-minimum phase system [10,11]. This paper presents a K- Harmonic Means KHM) local polynomial generalized predictive control algorithm, using KHM clustering algorithm to cluster analysis on historical database, finding out the recent class to the current working point, getting the specific working point neighborhood of input vector in this class, computing the corresponding weights and model parameters, and finally combining with generalized predictive control algorithm, designing the predictive controller. This paper gives the simulation case analysis, and the results show that the dynamic response of closed-loop system can be improved significantly. 2 Modeling Process Assume that the input and output dataset can expressed as {Y i, X i )} N i=1, it can characterize all kinds of possible basic condition of the nonlinear process, the relationship of input and output of the process can be expressed as Y i = fx i ) + ε i, i = 1,...N 1) f.) is the nonlinear mapping between input and output data, ε i is noise variable. By the Taylor formula, it is known that at the input vector x, the corresponding local function of the system f ) can be expressed by p order polynomial, fx, ξ) = ξ 0 + ξ 1 X i x) ξ r X i x) p 2) After get the similar data sample, local polynomial fitting is essentially a weighted optimal problem, its purpose is to minimize the mismatching degree between the model and data, ˆξ = arg min ξ i Ψ k x) ly i p ξ j X i x) j )w i x) 3) j=0 w i x) denote weight, Ψ k x) denote the specific working point neighborhood of input vector, it contains k sampling point. 3 K-harmonic Mean Database Searching Modeling process first need to determine the input-output history database, which should cover all kinds of working conditions, for industrial field data, removing noise filtering, smoothing and eliminating related redundant information should also be considered. Suppose the dataset which contain n group data can expressed as {x 1, x 2,..., x n }, clustering analysis the dataset, suppose k {2,..., n 1} is clustering number, V k is the kth class, k n is segmentation matrix, satisfy the following formula { uij = 1 if x j V i u ij = 0 if x j / V i

3 J. Zeng et al. /Journal of Computational Information Systems 9: ) and k i=1 u ij = 1, n j=1 u ij > 0 i = 1,..., k; j = 1,..., n). v i Ris defined as every kind of clustering centers, V = {v 1,..., v k } is clustering center set. The distance between data point x j and clustering center v i is defined as distance measure d ij, and define differ matrix by d d 1n D =..... d k1 d kn D can be calculated by X and V using appropriate norm, using the Euclidean norm to describe as follows d ij = x j v i )x j v i ) T Using the harmonic mean distance between data point and clustering center instead of the minimum distance between data point and clustering center k k i=1 1 d 2 ij min { d 2 ij i = 1,..., k } Objective function is expressed as J KHM V, X) = n j=1 k k i=1 1 d 2 ij 4) Make J / V =0 and then get the clustering center update methods as formula 5) shows. According to initial value continuously iteration, make formula 4) continuously reduce until stability. v i = n 1 ) 2 x j j=1 k d 4 1 ij d l=1 2 l,j n 1 j=1 k d 4 1 ij d l=1 2 l,j ) 2 5) Relative to the KM algorithm, KHM using the harmonic mean distance between data point and clustering center instead of the minimum distance between data point and clustering center. The conditional probability from clustering center to data point and dynamic weight of every iteration process are introduced. Algorithm procedure is as follows, Step 1 k samples from the historical input-output database {Y i, X i )} N i=1 are selected randomly as the initial clustering center v 1, v 2,..., v k. Step 2 According to formula 4), {Y i, X i )} N i=1 are divided into these K clusters according to the minimum harmonic mean principle between the data and clustering center. Step 3 The new clustering centers v 1 *, v 2 *,... v k * are computed by formula 5).

4 5622 J. Zeng et al. /Journal of Computational Information Systems 9: ) Step 4 If v l *= v l l = 1, 2,...,k ), jump into step 5, or turn to step 2. Step 5 Find out the recent class to the current working point, compute the specific working point neighborhood of input vector x in the specific class and compute weight w i x), compute model parameters according to formula 3). 4 Controller Design Consider the following NARX non-linear auto-regressive with exogenous inputs) system, yk) = fφk)) + ek), K = 1, M 6) f ) is an unknown nonlinear curve, e ) is an error random variables for example a random variable with mean zero and variance σk 2 ), φt) is regression vector, φt) = [yt 1)...yt n a )ut n k )...ut n b n k )] T n a, n b and n k are on behalf of the previous output, input and model hysteresis uantity respectively. Supposing formula 3) using uadratic form for l.), lε) = ε 2, and model structure is local linear, f φ k), ζ) = ζ 0 + ζ T 1 φ k) φ t)) 7) If ˆβ 0 and ˆβ 1 represent the parameter estimator, then the estimator of yt) can be expressed as: ŷt) = fφt), ˆζ) = ˆζ 0 8) By formula 7 can learn, a linear input-output form is available in the local neighborhood of φt), A 1 )yt) = B 1 )ut 1) + α 9) A 1 ) = 1 + a a n n B 1 ) = b 0 + b b nb n b A 1 ) and B 1 ) are polynomials about the recession operator 1, which can be obtained by ˆζ 1. α = ˆζ 0 ζ T 1 φt) 10) α is a deviation item. Multi-step prediction strategy is introduced at here, and in order to find out the optimal output predictive value after j step, Diophantine formula is introduced as follows 1 = E j 1 ) A 1) + j F j 1 ) 11) E j 1 ) = e j,0 + e j, e j,j 1 j+1 F j 1 ) = f j,0 + f j, f j,na n a By formula 9) 11) can learn y t + j) = ) Ḡj 1 u t + j + 1) + F ) j 1 y t) + E ) j 1 ξ t + j)

5 J. Zeng et al. /Journal of Computational Information Systems 9: ) Ḡ j 1 ) = E j 1 ) B 1) = B 1 ) 1 j F 1 )) A 1 ) = ḡ j,0 + ḡ j, The optimal predictive value at t + j can be expressed as follows ŷtj) = Ḡj 1 ) ut + j 1) + F j 1 )yt) 12) Another Diophantine formula is introduced which can decompose Ḡj 1 ) u t + j + 1) into two parts: known input item and unknown input item E j 1 ) B 1) = G j 1 ) + j H j 1 ) 13) G j 1 ) = g 0 + g g j 1 j+1 H j 1 ) = h j,0 + h j, h j,nb 1 n b+1 Then formula 12) can be expressed as vector form Ŷ = G U + f 14) Ŷ = [ŷ t + 1), ŷ t + 2), ŷt + 3)...ŷ t + N 2 )] T U = [ u t), u t + 1), ut + 2)... u t + N u )] T f t + 1) F 1 f t + 2) F 1 f = = H u t 1) + y t).. f t + N 2 ) F N2 G 1 1 ) g 1 [G 2 1 ) g 2 1 g 1 ] H = 2 [G 3 1 ) g 3 2 g 2 1 g 1 ]. [ ] N 2 1 G N2 1 ) g N2 N1+1 g 1 Then the output of the controlled object can be predicted by the known input, output information and the future input value. By formula 14), the performance index function can be expressed as follows { } J = E G U + f ω) T Q y G U + f ω) + U T Q u U 15) [ ω = ω t + 1) ω t + 2) ϖ t + N 2 ) ] T

6 5624 J. Zeng et al. /Journal of Computational Information Systems 9: ) Therefore, the future input which minimize the control performance index in the unconstrained conditions can be expressed U = G T Q y G + Q u I ) 1 G T Q y w f) 16) Then the real-time optimal control incremental is taken for the first element u t) = d T w f) 17) d T is the first line of the matrix G T Q y G + Q u I ) 1 G T Q y. d T = 1 0 0) G T Q y G + Q u I ) 1 G T Q y The first element of U in formula 16) is u t), then the real-time input u t) can be expressed by u t) = u t 1) + d T w f) 18) In the industrial controlled process, when there are input constraints u min ut + k) u max 19) u min ut + k) u max 20) and because the controlled variables can be written as vector form U = T U + Ū 21) T Ū ut 1) ut 1). ut 1) Then the constraints can be expressed as [ ] C U T U T U T U T c 22) C I I T T u max c u min u max Ū u min + Ū It is a uadratic programming problem as follows: min U T G T Q y G + Q u I) U 2ω f)q y G U U [ ] s.t. C U T U T U T U T c

7 J. Zeng et al. /Journal of Computational Information Systems 9: ) Simulation Study Consider the following strong nonlinear system y t)1 + y t) ) = ut) This system gain decreases along with the increase of the output signal amplitude. Because the open loop system is unstable, we generate 2500 input-output data sets in this paper by using closed loop proportional controller. The sampling time is selected as T s =0.1. The controller is designed by the data sets and the method in this paper, the simulation results are compared with the conventional generalized predictive control method. Suppose the system output is step signal, at t=1, the output is jumped form 0 to 1. In the simulation process, the predictive time domain is set as N=8, the controlled time domain is set as N u =6, the weight of output variable is Qy=1, the increment weight of the controlled signal is Q U = The output results using conventional generalized predictive control algorithm is shown in figure 1, the results by the method in this paper is shown in figure 2, the real line represent system output, dotted line represent reference in figure a), the controlled input signal is shown in figure b). Form the diagram we can see that compared with routine predictive control algorithm, the system output overshoot by the controller of our method is decrease obviously, and system response time is shortened, the input signal freuency jump is avoided. The superiority of the algorithm in this paper is seen clearly. a) Fig. 1: Step response curve of conventional predictive controller b) a) Fig. 2: Step response curve of our algorithm b)

8 5626 J. Zeng et al. /Journal of Computational Information Systems 9: ) Conclusion A K-harmonic mean clustering algorithm which is not sensitive to initial value selection is introduced in this paper, the harmonic mean distance between data point and clustering center instead of the minimum distance between data point and clustering center. The conditional probability from clustering center to data point and dynamic weight of every iteration process are also introduced. The recent class to current working point is selected by tracking real time condition changes, the specific working point neighborhood and model parameters are computed in the class, the appropriate control increment is calculated too. The simulation results show the effectiveness of the method. Acknowledgement This work is supported by the Scientific Research Project of Liaoning Province, China L , L ). References [1] N.N. Nandola and S. Bhartiya. Hybrid system identification using a structural approach and its model based control: An experimental validation. Nonlinear Analysis: Hybrid systems, 32): , [2] Yunan S, Maozai T. Adaptive local linear uantile regression. Acta Mathematicae Applicata Sinica, 273): , [3] Seuenz H, Schreiber A, Isermann R. Identification of nonlinear static processes with local polynomial regression and subset selection. IFAC Symposium on System Identification, , Saint- Malo, France, July [4] Hang Yue; Jones E G, Revesz P. Local polynomial regression models for average traffic speed estimation and forecasting in linear constraint databases. 17th International symposium on temporal representation and reasoning, [5] Kumar A, Sinha R and Bhattacherjee V. Modeling using K-means clustering algorithm st International Conference on Rencent Advances in Information Technology, , [6] B. Zhang, M. Hsu, and U. Dayal. K-harmonic means a data clustering algorithm. Technical Report HPL , HP Laboratories Palo Alto, [7] A. Inler and Z. Gungor. Applying k-harmonic means clustering to the part-machine classification problem. Expert Systems with Applications, 362): , [8] Jing Zeng, Hui Jin. Local Polynomial Regression with K-harmonic Means and Subset Selection. Journal of Computational Information Systems, 2012, 822): [9] K. Thangavel and N. Karthikeyani Visalakshi. Ensemble based distributed k-harmonic means clustering. International Journal of Recent Trends in Engineering, 21): , November [10] Wei Wang. Theory and application of generalized predictive control. Science press, [11] N.N. Nandola and D.E Rivera. A novel model predictive control formulation for hybrid systems with application to adaptive behavioral interventions. ACC, , Baltimore, Maryland, USA, June, 2010.

A Hybrid Time-delay Prediction Method for Networked Control System

A Hybrid Time-delay Prediction Method for Networked Control System International Journal of Automation and Computing 11(1), February 2014, 19-24 DOI: 10.1007/s11633-014-0761-1 A Hybrid Time-delay Prediction Method for Networked Control System Zhong-Da Tian Xian-Wen Gao

More information

! # % & () +,.&/ 01),. &, / &

! # % & () +,.&/ 01),. &, / & ! # % & () +,.&/ ),. &, / & 2 A NEW METHOD FOR THE DESIGN OF ENERGY TRANSFER FILTERS Xiaofeng Wu, Z Q Lang and S. A. Billings Department of Automatic Control and Systems Engineering The University of Sheffield

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

Hybrid particle swarm algorithm for solving nonlinear constraint. optimization problem [5].

Hybrid particle swarm algorithm for solving nonlinear constraint. optimization problem [5]. Hybrid particle swarm algorithm for solving nonlinear constraint optimization problems BINGQIN QIAO, XIAOMING CHANG Computers and Software College Taiyuan University of Technology Department of Economic

More information

Bearing fault diagnosis based on EMD-KPCA and ELM

Bearing fault diagnosis based on EMD-KPCA and ELM Bearing fault diagnosis based on EMD-KPCA and ELM Zihan Chen, Hang Yuan 2 School of Reliability and Systems Engineering, Beihang University, Beijing 9, China Science and Technology on Reliability & Environmental

More information

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction ISSN 1749-3889 (print), 1749-3897 (online) International Journal of Nonlinear Science Vol.11(2011) No.2,pp.153-158 Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method Yigui Ou, Jun Zhang

More information

Feature gene selection method based on logistic and correlation information entropy

Feature gene selection method based on logistic and correlation information entropy Bio-Medical Materials and Engineering 26 (2015) S1953 S1959 DOI 10.3233/BME-151498 IOS Press S1953 Feature gene selection method based on logistic and correlation information entropy Jiucheng Xu a,b,,

More information

Chap 1. Overview of Statistical Learning (HTF, , 2.9) Yongdai Kim Seoul National University

Chap 1. Overview of Statistical Learning (HTF, , 2.9) Yongdai Kim Seoul National University Chap 1. Overview of Statistical Learning (HTF, 2.1-2.6, 2.9) Yongdai Kim Seoul National University 0. Learning vs Statistical learning Learning procedure Construct a claim by observing data or using logics

More information

Curriculum Vitae Wenxiao Zhao

Curriculum Vitae Wenxiao Zhao 1 Personal Information Curriculum Vitae Wenxiao Zhao Wenxiao Zhao, Male PhD, Associate Professor with Key Laboratory of Systems and Control, Institute of Systems Science, Academy of Mathematics and Systems

More information

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. ISS0031 Modeling and Identification Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. Aleksei Tepljakov, Ph.D. October 21, 2015 Discrete-time Transfer Functions

More information

Soft Sensor Modelling based on Just-in-Time Learning and Bagging-PLS for Fermentation Processes

Soft Sensor Modelling based on Just-in-Time Learning and Bagging-PLS for Fermentation Processes 1435 A publication of CHEMICAL ENGINEERING TRANSACTIONS VOL. 70, 2018 Guest Editors: Timothy G. Walmsley, Petar S. Varbanov, Rongin Su, Jiří J. Klemeš Copyright 2018, AIDIC Servizi S.r.l. ISBN 978-88-95608-67-9;

More information

EL1820 Modeling of Dynamical Systems

EL1820 Modeling of Dynamical Systems EL1820 Modeling of Dynamical Systems Lecture 10 - System identification as a model building tool Experiment design Examination and prefiltering of data Model structure selection Model validation Lecture

More information

Iterative Controller Tuning Using Bode s Integrals

Iterative Controller Tuning Using Bode s Integrals Iterative Controller Tuning Using Bode s Integrals A. Karimi, D. Garcia and R. Longchamp Laboratoire d automatique, École Polytechnique Fédérale de Lausanne (EPFL), 05 Lausanne, Switzerland. email: alireza.karimi@epfl.ch

More information

Nonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto

Nonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto Dr-Ing Sudchai Boonto Department of Control System and Instrumentation Engineering King Mongkut s Unniversity of Technology Thonburi Thailand Nonlinear System Identification Given a data set Z N = {y(k),

More information

Optimization. Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison

Optimization. Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison Optimization Benjamin Recht University of California, Berkeley Stephen Wright University of Wisconsin-Madison optimization () cost constraints might be too much to cover in 3 hours optimization (for big

More information

Research Article A Novel Differential Evolution Invasive Weed Optimization Algorithm for Solving Nonlinear Equations Systems

Research Article A Novel Differential Evolution Invasive Weed Optimization Algorithm for Solving Nonlinear Equations Systems Journal of Applied Mathematics Volume 2013, Article ID 757391, 18 pages http://dx.doi.org/10.1155/2013/757391 Research Article A Novel Differential Evolution Invasive Weed Optimization for Solving Nonlinear

More information

Statistical Methods for Data Mining

Statistical Methods for Data Mining Statistical Methods for Data Mining Kuangnan Fang Xiamen University Email: xmufkn@xmu.edu.cn Support Vector Machines Here we approach the two-class classification problem in a direct way: We try and find

More information

4.0 Update Algorithms For Linear Closed-Loop Systems

4.0 Update Algorithms For Linear Closed-Loop Systems 4. Update Algorithms For Linear Closed-Loop Systems A controller design methodology has been developed that combines an adaptive finite impulse response (FIR) filter with feedback. FIR filters are used

More information

Kaggle.

Kaggle. Administrivia Mini-project 2 due April 7, in class implement multi-class reductions, naive bayes, kernel perceptron, multi-class logistic regression and two layer neural networks training set: Project

More information

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Song Li 1, Peng Wang 1 and Lalit Goel 1 1 School of Electrical and Electronic Engineering Nanyang Technological University

More information

Stochastic Analogues to Deterministic Optimizers

Stochastic Analogues to Deterministic Optimizers Stochastic Analogues to Deterministic Optimizers ISMP 2018 Bordeaux, France Vivak Patel Presented by: Mihai Anitescu July 6, 2018 1 Apology I apologize for not being here to give this talk myself. I injured

More information

Learning with multiple models. Boosting.

Learning with multiple models. Boosting. CS 2750 Machine Learning Lecture 21 Learning with multiple models. Boosting. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Learning with multiple models: Approach 2 Approach 2: use multiple models

More information

Stochastic Subgradient Method

Stochastic Subgradient Method Stochastic Subgradient Method Lingjie Weng, Yutian Chen Bren School of Information and Computer Science UC Irvine Subgradient Recall basic inequality for convex differentiable f : f y f x + f x T (y x)

More information

Research Article Identifying a Global Optimizer with Filled Function for Nonlinear Integer Programming

Research Article Identifying a Global Optimizer with Filled Function for Nonlinear Integer Programming Discrete Dynamics in Nature and Society Volume 20, Article ID 7697, pages doi:0.55/20/7697 Research Article Identifying a Global Optimizer with Filled Function for Nonlinear Integer Programming Wei-Xiang

More information

Adaptive Primal Dual Optimization for Image Processing and Learning

Adaptive Primal Dual Optimization for Image Processing and Learning Adaptive Primal Dual Optimization for Image Processing and Learning Tom Goldstein Rice University tag7@rice.edu Ernie Esser University of British Columbia eesser@eos.ubc.ca Richard Baraniuk Rice University

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

C.-H. Lamarque. University of Lyon/ENTPE/LGCB & LTDS UMR CNRS 5513

C.-H. Lamarque. University of Lyon/ENTPE/LGCB & LTDS UMR CNRS 5513 Nonlinear Dynamics of Smooth and Non-Smooth Systems with Application to Passive Controls 3rd Sperlonga Summer School on Mechanics and Engineering Sciences on Dynamics, Stability and Control of Flexible

More information

Bifurcation Analysis of the Aeroelastic Galloping Problem via Input-Output Parametric Modelling

Bifurcation Analysis of the Aeroelastic Galloping Problem via Input-Output Parametric Modelling Bifurcation Analysis of the Aeroelastic Galloping Problem via Input-Output Parametric Modelling Nicholas F. Giannelis 1 and Gareth A. Vio School of Aeronautical, Mechanical and Mechatronic Engineering

More information

EL1820 Modeling of Dynamical Systems

EL1820 Modeling of Dynamical Systems EL1820 Modeling of Dynamical Systems Lecture 9 - Parameter estimation in linear models Model structures Parameter estimation via prediction error minimization Properties of the estimate: bias and variance

More information

Method for Recognizing Mechanical Status of Container Crane Motor Based on SOM Neural Network

Method for Recognizing Mechanical Status of Container Crane Motor Based on SOM Neural Network IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS Method for Recognizing Mechanical Status of Container Crane Motor Based on SOM Neural Network To cite this article: X Q Yao et

More information

6.036 midterm review. Wednesday, March 18, 15

6.036 midterm review. Wednesday, March 18, 15 6.036 midterm review 1 Topics covered supervised learning labels available unsupervised learning no labels available semi-supervised learning some labels available - what algorithms have you learned that

More information

Chapter 4. Replication Variance Estimation. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28

Chapter 4. Replication Variance Estimation. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28 Chapter 4 Replication Variance Estimation J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28 Jackknife Variance Estimation Create a new sample by deleting one observation n 1 n n ( x (k) x) 2 = x (k) = n

More information

An artificial chemical reaction optimization algorithm for. multiple-choice; knapsack problem.

An artificial chemical reaction optimization algorithm for. multiple-choice; knapsack problem. An artificial chemical reaction optimization algorithm for multiple-choice knapsack problem Tung Khac Truong 1,2, Kenli Li 1, Yuming Xu 1, Aijia Ouyang 1, and Xiaoyong Tang 1 1 College of Information Science

More information

ROBUST PASSIVE OBSERVER-BASED CONTROL FOR A CLASS OF SINGULAR SYSTEMS

ROBUST PASSIVE OBSERVER-BASED CONTROL FOR A CLASS OF SINGULAR SYSTEMS INTERNATIONAL JOURNAL OF INFORMATON AND SYSTEMS SCIENCES Volume 5 Number 3-4 Pages 480 487 c 2009 Institute for Scientific Computing and Information ROBUST PASSIVE OBSERVER-BASED CONTROL FOR A CLASS OF

More information

Dynamic-Inner Partial Least Squares for Dynamic Data Modeling

Dynamic-Inner Partial Least Squares for Dynamic Data Modeling Preprints of the 9th International Symposium on Advanced Control of Chemical Processes The International Federation of Automatic Control MoM.5 Dynamic-Inner Partial Least Squares for Dynamic Data Modeling

More information

Kalman Filter and Parameter Identification. Florian Herzog

Kalman Filter and Parameter Identification. Florian Herzog Kalman Filter and Parameter Identification Florian Herzog 2013 Continuous-time Kalman Filter In this chapter, we shall use stochastic processes with independent increments w 1 (.) and w 2 (.) at the input

More information

STK Statistical Learning: Advanced Regression and Classification

STK Statistical Learning: Advanced Regression and Classification STK4030 - Statistical Learning: Advanced Regression and Classification Riccardo De Bin debin@math.uio.no STK4030: lecture 1 1/ 42 Outline of the lecture Introduction Overview of supervised learning Variable

More information

Discriminative Models

Discriminative Models No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models

More information

Forecasting Data Streams: Next Generation Flow Field Forecasting

Forecasting Data Streams: Next Generation Flow Field Forecasting Forecasting Data Streams: Next Generation Flow Field Forecasting Kyle Caudle South Dakota School of Mines & Technology (SDSMT) kyle.caudle@sdsmt.edu Joint work with Michael Frey (Bucknell University) and

More information

Dual Estimation and the Unscented Transformation

Dual Estimation and the Unscented Transformation Dual Estimation and the Unscented Transformation Eric A. Wan ericwan@ece.ogi.edu Rudolph van der Merwe rudmerwe@ece.ogi.edu Alex T. Nelson atnelson@ece.ogi.edu Oregon Graduate Institute of Science & Technology

More information

Journal of Chemical and Pharmaceutical Research, 2014, 6(3): Research Article

Journal of Chemical and Pharmaceutical Research, 2014, 6(3): Research Article Available online www.jocpr.com Journal of Chemical and Pharmaceutical Research, 2014, 6(3):813-817 Research Article ISSN : 0975-7384 COEN(USA) : JCPRC5 Research On the Burning Zone Temperature Control

More information

Support Vector Machines for Classification: A Statistical Portrait

Support Vector Machines for Classification: A Statistical Portrait Support Vector Machines for Classification: A Statistical Portrait Yoonkyung Lee Department of Statistics The Ohio State University May 27, 2011 The Spring Conference of Korean Statistical Society KAIST,

More information

LEAST-SQUARES FINITE ELEMENT MODELS

LEAST-SQUARES FINITE ELEMENT MODELS LEAST-SQUARES FINITE ELEMENT MODELS General idea of the least-squares formulation applied to an abstract boundary-value problem Works of our group Application to Poisson s equation Application to flows

More information

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18 CSE 417T: Introduction to Machine Learning Lecture 11: Review Henry Chai 10/02/18 Unknown Target Function!: # % Training data Formal Setup & = ( ), + ),, ( -, + - Learning Algorithm 2 Hypothesis Set H

More information

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature suggests the design variables should be normalized to a range of [-1,1] or [0,1].

More information

A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong

A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,

More information

APPLICATION OF A KERNEL METHOD IN MODELING FRICTION DYNAMICS

APPLICATION OF A KERNEL METHOD IN MODELING FRICTION DYNAMICS APPLICATION OF A KERNEL METHOD IN MODELING FRICTION DYNAMICS Yufeng Wan, Chian X. Wong, Tony J. Dodd, Robert F. Harrison Department of Automatic Control and Systems Engineering, The University of Sheffield,

More information

A Wavelet Neural Network Forecasting Model Based On ARIMA

A Wavelet Neural Network Forecasting Model Based On ARIMA A Wavelet Neural Network Forecasting Model Based On ARIMA Wang Bin*, Hao Wen-ning, Chen Gang, He Deng-chao, Feng Bo PLA University of Science &Technology Nanjing 210007, China e-mail:lgdwangbin@163.com

More information

Statistics 910, #15 1. Kalman Filter

Statistics 910, #15 1. Kalman Filter Statistics 910, #15 1 Overview 1. Summary of Kalman filter 2. Derivations 3. ARMA likelihoods 4. Recursions for the variance Kalman Filter Summary of Kalman filter Simplifications To make the derivations

More information

Closed-loop Identification of Hammerstein Systems Using Iterative Instrumental Variables

Closed-loop Identification of Hammerstein Systems Using Iterative Instrumental Variables Proceedings of the 18th World Congress The International Federation of Automatic Control Closed-loop Identification of Hammerstein Systems Using Iterative Instrumental Variables Younghee Han and Raymond

More information

Introduction to Statistical modeling: handout for Math 489/583

Introduction to Statistical modeling: handout for Math 489/583 Introduction to Statistical modeling: handout for Math 489/583 Statistical modeling occurs when we are trying to model some data using statistical tools. From the start, we recognize that no model is perfect

More information

Iterative Reweighted Least Squares

Iterative Reweighted Least Squares Iterative Reweighted Least Squares Sargur. University at Buffalo, State University of ew York USA Topics in Linear Classification using Probabilistic Discriminative Models Generative vs Discriminative

More information

Machine Learning and Data Mining. Linear regression. Kalev Kask

Machine Learning and Data Mining. Linear regression. Kalev Kask Machine Learning and Data Mining Linear regression Kalev Kask Supervised learning Notation Features x Targets y Predictions ŷ Parameters q Learning algorithm Program ( Learner ) Change q Improve performance

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Support Vector Machine via Nonlinear Rescaling Method

Support Vector Machine via Nonlinear Rescaling Method Manuscript Click here to download Manuscript: svm-nrm_3.tex Support Vector Machine via Nonlinear Rescaling Method Roman Polyak Department of SEOR and Department of Mathematical Sciences George Mason University

More information

Engine fault feature extraction based on order tracking and VMD in transient conditions

Engine fault feature extraction based on order tracking and VMD in transient conditions Engine fault feature extraction based on order tracking and VMD in transient conditions Gang Ren 1, Jide Jia 2, Jian Mei 3, Xiangyu Jia 4, Jiajia Han 5 1, 4, 5 Fifth Cadet Brigade, Army Transportation

More information

Refined Instrumental Variable Methods for Identifying Hammerstein Models Operating in Closed Loop

Refined Instrumental Variable Methods for Identifying Hammerstein Models Operating in Closed Loop Refined Instrumental Variable Methods for Identifying Hammerstein Models Operating in Closed Loop V. Laurain, M. Gilson, H. Garnier Abstract This article presents an instrumental variable method dedicated

More information

Lazy learning for control design

Lazy learning for control design Lazy learning for control design Gianluca Bontempi, Mauro Birattari, Hugues Bersini Iridia - CP 94/6 Université Libre de Bruxelles 5 Bruxelles - Belgium email: {gbonte, mbiro, bersini}@ulb.ac.be Abstract.

More information

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition Data Mining Classification: Basic Concepts and Techniques Lecture Notes for Chapter 3 by Tan, Steinbach, Karpatne, Kumar 1 Classification: Definition Given a collection of records (training set ) Each

More information

Lecture 9. Time series prediction

Lecture 9. Time series prediction Lecture 9 Time series prediction Prediction is about function fitting To predict we need to model There are a bewildering number of models for data we look at some of the major approaches in this lecture

More information

Large-scale Collaborative Ranking in Near-Linear Time

Large-scale Collaborative Ranking in Near-Linear Time Large-scale Collaborative Ranking in Near-Linear Time Liwei Wu Depts of Statistics and Computer Science UC Davis KDD 17, Halifax, Canada August 13-17, 2017 Joint work with Cho-Jui Hsieh and James Sharpnack

More information

Online Learning. Jordan Boyd-Graber. University of Colorado Boulder LECTURE 21. Slides adapted from Mohri

Online Learning. Jordan Boyd-Graber. University of Colorado Boulder LECTURE 21. Slides adapted from Mohri Online Learning Jordan Boyd-Graber University of Colorado Boulder LECTURE 21 Slides adapted from Mohri Jordan Boyd-Graber Boulder Online Learning 1 of 31 Motivation PAC learning: distribution fixed over

More information

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms Neural networks Comments Assignment 3 code released implement classification algorithms use kernels for census dataset Thought questions 3 due this week Mini-project: hopefully you have started 2 Example:

More information

A Fast Approximation Algorithm for Set-Membership System Identification

A Fast Approximation Algorithm for Set-Membership System Identification A Fast Approximation Algorithm for Set-Membership System Identification J.A. Castaño F. Ruiz J. Régnier Pontificia Universidad Javeriana, Bogotá, Colombia. e-mail: {j-castano,ruizf}@javeriana.edu.co Université

More information

A NEW METHOD FOR VIBRATION MODE ANALYSIS

A NEW METHOD FOR VIBRATION MODE ANALYSIS Proceedings of IDETC/CIE 25 25 ASME 25 International Design Engineering Technical Conferences & Computers and Information in Engineering Conference Long Beach, California, USA, September 24-28, 25 DETC25-85138

More information

Approximation Bound for Fuzzy-Neural Networks with Bell Membership Function

Approximation Bound for Fuzzy-Neural Networks with Bell Membership Function Approximation Bound for Fuzzy-Neural Networks with Bell Membership Function Weimin Ma, and Guoqing Chen School of Economics and Management, Tsinghua University, Beijing, 00084, P.R. China {mawm, chengq}@em.tsinghua.edu.cn

More information

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2015 http://www.astro.cornell.edu/~cordes/a6523 Lecture 23:! Nonlinear least squares!! Notes Modeling2015.pdf on course

More information

Nonlinear PD Controllers with Gravity Compensation for Robot Manipulators

Nonlinear PD Controllers with Gravity Compensation for Robot Manipulators BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 4, No Sofia 04 Print ISSN: 3-970; Online ISSN: 34-408 DOI: 0.478/cait-04-00 Nonlinear PD Controllers with Gravity Compensation

More information

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 12, NO. 5, SEPTEMBER 2001 1215 A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing Da-Zheng Feng, Zheng Bao, Xian-Da Zhang

More information

Data Exploration and Unsupervised Learning with Clustering

Data Exploration and Unsupervised Learning with Clustering Data Exploration and Unsupervised Learning with Clustering Paul F Rodriguez,PhD San Diego Supercomputer Center Predictive Analytic Center of Excellence Clustering Idea Given a set of data can we find a

More information

Gaussian Process for Internal Model Control

Gaussian Process for Internal Model Control Gaussian Process for Internal Model Control Gregor Gregorčič and Gordon Lightbody Department of Electrical Engineering University College Cork IRELAND E mail: gregorg@rennesuccie Abstract To improve transparency

More information

Data-driven methods in application to flood defence systems monitoring and analysis Pyayt, A.

Data-driven methods in application to flood defence systems monitoring and analysis Pyayt, A. UvA-DARE (Digital Academic Repository) Data-driven methods in application to flood defence systems monitoring and analysis Pyayt, A. Link to publication Citation for published version (APA): Pyayt, A.

More information

Fine-grained Photovoltaic Output Prediction using a Bayesian Ensemble

Fine-grained Photovoltaic Output Prediction using a Bayesian Ensemble Fine-grained Photovoltaic Output Prediction using a Bayesian Ensemble 1,2, Manish Marwah 3,, Martin Arlitt 3, and Naren Ramakrishnan 1,2 1 Department of Computer Science, Virginia Tech, Blacksburg, VA

More information

Model-free Predictive Control

Model-free Predictive Control Model-free Predictive Control Anders Stenman Department of Electrical Engineering Linköping University, S-581 83 Linköping, Sweden WWW: http://wwwcontrolisyliuse Email: stenman@isyliuse February 25, 1999

More information

Supplemental Material for Discrete Graph Hashing

Supplemental Material for Discrete Graph Hashing Supplemental Material for Discrete Graph Hashing Wei Liu Cun Mu Sanjiv Kumar Shih-Fu Chang IM T. J. Watson Research Center Columbia University Google Research weiliu@us.ibm.com cm52@columbia.edu sfchang@ee.columbia.edu

More information

A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL FUZZY CLUSTERING *

A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL FUZZY CLUSTERING * No.2, Vol.1, Winter 2012 2012 Published by JSES. A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL * Faruk ALPASLAN a, Ozge CAGCAG b Abstract Fuzzy time series forecasting methods

More information

Neural Network Training

Neural Network Training Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification

More information

The analog data assimilation: method, applications and implementation

The analog data assimilation: method, applications and implementation The analog data assimilation: method, applications and implementation Pierre Tandeo pierre.tandeo@imt-atlantique.fr IMT-Atlantique, Brest, France 16 th of February 2018 Before starting Works in collaboration

More information

The Perceptron algorithm

The Perceptron algorithm The Perceptron algorithm Tirgul 3 November 2016 Agnostic PAC Learnability A hypothesis class H is agnostic PAC learnable if there exists a function m H : 0,1 2 N and a learning algorithm with the following

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Ensembles Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique Fédérale de Lausanne

More information

CS229 Final Project. Wentao Zhang Shaochuan Xu

CS229 Final Project. Wentao Zhang Shaochuan Xu CS229 Final Project Shale Gas Production Decline Prediction Using Machine Learning Algorithms Wentao Zhang wentaoz@stanford.edu Shaochuan Xu scxu@stanford.edu In petroleum industry, oil companies sometimes

More information

Weighted Least Squares

Weighted Least Squares Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w

More information

Humanoid Based Intelligence Control Strategy of Plastic Cement Die Press Work-Piece Forming Process for Polymer Plastics

Humanoid Based Intelligence Control Strategy of Plastic Cement Die Press Work-Piece Forming Process for Polymer Plastics Journal of Materials Science and Chemical Engineering, 206, 4, 9-6 Published Online June 206 in SciRes. http://www.scirp.org/journal/msce http://dx.doi.org/0.4236/msce.206.46002 Humanoid Based Intelligence

More information

Discriminative Models

Discriminative Models No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models

More information

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2 Nearest Neighbor Machine Learning CSE546 Kevin Jamieson University of Washington October 26, 2017 2017 Kevin Jamieson 2 Some data, Bayes Classifier Training data: True label: +1 True label: -1 Optimal

More information

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:

More information

Location Prediction of Moving Target

Location Prediction of Moving Target Location of Moving Target Department of Radiation Oncology Stanford University AAPM 2009 Outline of Topics 1 Outline of Topics 1 2 Outline of Topics 1 2 3 Estimation of Stochastic Process Stochastic Regression:

More information

SIMPLIFIED MARGINAL LINEARIZATION METHOD IN AUTONOMOUS LIENARD SYSTEMS

SIMPLIFIED MARGINAL LINEARIZATION METHOD IN AUTONOMOUS LIENARD SYSTEMS italian journal of pure and applied mathematics n. 30 03 (67 78) 67 SIMPLIFIED MARGINAL LINEARIZATION METHOD IN AUTONOMOUS LIENARD SYSTEMS Weijing Zhao Faculty of Electronic Information and Electrical

More information

Machine Learning And Applications: Supervised Learning-SVM

Machine Learning And Applications: Supervised Learning-SVM Machine Learning And Applications: Supervised Learning-SVM Raphaël Bournhonesque École Normale Supérieure de Lyon, Lyon, France raphael.bournhonesque@ens-lyon.fr 1 Supervised vs unsupervised learning Machine

More information

Ad Placement Strategies

Ad Placement Strategies Case Study 1: Estimating Click Probabilities Tackling an Unknown Number of Features with Sketching Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox 2014 Emily Fox January

More information

Multi-scale Geometric Summaries for Similarity-based Upstream S

Multi-scale Geometric Summaries for Similarity-based Upstream S Multi-scale Geometric Summaries for Similarity-based Upstream Sensor Fusion Duke University, ECE / Math 3/6/2019 Overall Goals / Design Choices Leverage multiple, heterogeneous modalities in identification

More information

Scuola di Calcolo Scientifico con MATLAB (SCSM) 2017 Palermo 31 Luglio - 4 Agosto 2017

Scuola di Calcolo Scientifico con MATLAB (SCSM) 2017 Palermo 31 Luglio - 4 Agosto 2017 Scuola di Calcolo Scientifico con MATLAB (SCSM) 2017 Palermo 31 Luglio - 4 Agosto 2017 www.u4learn.it Ing. Giuseppe La Tona Sommario Machine Learning definition Machine Learning Problems Artificial Neural

More information

L-statistics based Modification of Reconstruction Algorithms for Compressive Sensing in the Presence of Impulse Noise

L-statistics based Modification of Reconstruction Algorithms for Compressive Sensing in the Presence of Impulse Noise L-statistics based Modification of Reconstruction Algorithms for Compressive Sensing in the Presence of Impulse Noise Srdjan Stanković, Irena Orović and Moeness Amin 1 Abstract- A modification of standard

More information

Adaptive Dual Control

Adaptive Dual Control Adaptive Dual Control Björn Wittenmark Department of Automatic Control, Lund Institute of Technology Box 118, S-221 00 Lund, Sweden email: bjorn@control.lth.se Keywords: Dual control, stochastic control,

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Notion of Distance. Metric Distance Binary Vector Distances Tangent Distance

Notion of Distance. Metric Distance Binary Vector Distances Tangent Distance Notion of Distance Metric Distance Binary Vector Distances Tangent Distance Distance Measures Many pattern recognition/data mining techniques are based on similarity measures between objects e.g., nearest-neighbor

More information

Overfitting, Bias / Variance Analysis

Overfitting, Bias / Variance Analysis Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic

More information

Logarithmic quantisation of wavelet coefficients for improved texture classification performance

Logarithmic quantisation of wavelet coefficients for improved texture classification performance Logarithmic quantisation of wavelet coefficients for improved texture classification performance Author Busch, Andrew, W. Boles, Wageeh, Sridharan, Sridha Published 2004 Conference Title 2004 IEEE International

More information

The Phase Detection Algorithm of Weak Signals Based on Coupled Chaotic-oscillators Sun Wen Jun Rui Guo Sheng, Zhang Yang

The Phase Detection Algorithm of Weak Signals Based on Coupled Chaotic-oscillators Sun Wen Jun Rui Guo Sheng, Zhang Yang 4th International Conference on Mechatronics, Materials, Chemistry and Computer Engineering (ICMMCCE 205) The Phase Detection Algorithm of Weak Signals Based on Coupled Chaotic-oscillators Sun Wen Jun

More information

EECE Adaptive Control

EECE Adaptive Control EECE 574 - Adaptive Control Basics of System Identification Guy Dumont Department of Electrical and Computer Engineering University of British Columbia January 2010 Guy Dumont (UBC) EECE574 - Basics of

More information