AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET

Size: px
Start display at page:

Download "AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET"

Transcription

1 Identification of Linear and Nonlinear Dynamical Systems Theme : Nonlinear Models Grey-box models Division of Automatic Control Linköping University Sweden General Aspects Let Z t denote all available (input-output) data up to time t A mathematical model for the system is a function from these data to the space where the output at time t, y(t) lives, in general ŷ(t t ) = g(z t, t) The function can be thought of as a predictor of the next output A parametric model structure is a parameterized family of such models: g(z t, θ) Choice of regressors and nonlinear function Functions for a scalar regressor Expansion into multiple regressors Examples of named structures All aspects on curve fitting applies pretty much also to this case The difficulty is the enormous richness in possibilities of parameterizations There are two main cases : General models of great flexibility Grey-box models: Models that incorporate some knowledge of the character of the actual system Black-box Models: General Comments NL Black Box: Choice of g 6 The general mapping g(z t, θ) is normally too flexible Let us split it into one mapping from Z t to a regression vector ϕ(t) of fixed dimension d and a mapping g from R d to R (assuming the output to be scalar): Leaves two problems g(z t, θ) = g(ϕ(t), θ) ϕ(t) = ϕ(z t ) Choose the mapping g(ϕ, θ) Choose the regression vector ϕ(t) ( or ϕ(t, θ) = ϕ(z t, θ)) First, consider ϕ to be scalar Basic form g(ϕ, θ) = N α k (β k (ϕ γ k )) k= (x) = cos(x): Fourier transform (x) = U(x): Unit pulse, gives piecewise constant functions g Soft version: (x) = e x / (x) = H(x): Step at x =, gives also piecewise constant functions Soft version: (x) = +e x α coordinates, β scale or dilation, γ location Several Regressors 7 Examples of Named Structures 8 Consider now ϕ to be a d-dimensional vector, but let still (x) be a function of one variable How to interpret (β(ϕ γ))? Radial β(ϕ γ) = ϕ γ β = (ϕ γ) T β(ϕ γ) γ a d-dimensional vector, β a d d-matrix (positive definite) or scaled version of the identity matrix with β a scalar Describes an ellipsoid in R d Ridge β(ϕ γ) = β T ϕ γ β a d-dimensional vector, γ a scalar Describes a hyperplane in R d Tensor is a product of factors corresponding to the components of the vector: (β(ϕ γ)) = d k= (β k(ϕ k γ k )) γ and β are d-dimensional vectors and subscript denotes component ANN: artificial Neural Networks One hidden layer sigmoidal: (x) = +e x, ridge extension Radial Basis Networks: (x) = e x /, radial extension Wavelets: is the mother wavelet and β j = j, γ k = j k (double indexing) as fixed choices (Neuro)-Fuzzy models: are the membership functions, tensor expansion

2 Simulation and Prediction 9 Choice of Regressors Suppose ϕ(t) = [y(t ), u(t )] T The (one-step ahead) predicted output at time for a given model θ is then Four players: Outputs y(t k), Inputs u(t k) ŷ p (t θ) = g([y(t ), u(t )] T, θ) It uses the previous measurement y(t ) A tougher test is to check how the model would behave in simulation, ie when only the input sequence u is used The simulated output is obtained as above, by replacing the measured output by the simulated output from the previous step: ŷ s (t, θ) = g([ŷ s (t, θ), u(t )] T, θ) Simulated model outputs ŷ s (t k, θ) Predicted model outputs ŷ p (t k θ) Regressors for dynamical systems are typically chosen among those: NFIR-models use past inputs NARX-models use past inputs and outputs NOE-models use past inputs and past simulated outputs NARMAX-models use inputs, outputs predicted outputs NBJ-models use all four regressor types Notice a possible stability problem! Recurrent Networks Network Aspects Several Layers For NOE, NARMAX and NBJ, previous outputs from the model have to be fed back into the model computations on-line: ϕ g The model structures are really basis function expansions However, since the basis functions are variants of the same function, a graphical description looks like a network One can also let the regressors be outputs from a previous layer of the network: Input layer Hidden layers Output layer θ θ θ q - q - These are called recurrent networks and require considerable more computational work to fit to data { ϕ g Example: Hydraulic Crane Data Linear Model These are data from a forest harvest machine: OUTPUT # Black: Measured Output Blue: Model Simulated Output Linear model: Fit 7 % INPUT # Sigmoidal ANN model Wavenet Model (Radial BF ANN Model) 6 Sigmoidal NN: Fit 9 % Wavenet model: Fit 7 %

3 7 Physical Modeling 8 Physical Modeling Semi-physical Modeling Block-models Local Linear Models Perform physical modeling (eg in MODELICA) and denote unknown physical parameters by θ Collect the model equations as ẋ(t) = f(x(t), u(t), θ) y(t) = h(x(t), u(t), θ) (or in DAE, Differential Algebraic Equations, form) For each parameter θ this defines a simulated (predicted) output ŷ(t θ) which is the parameterized function ŷ(t θ) = g(z t, θ) in somewhat implicit form To be a correct predictor this really assumes white measurement noise Some more sophistical noise modeling is possible, usually involving ad hoc non-linear observers The approach is conceptually simple, but could be very demanding in practice Example: Missile 9 The Equations function [dx, y] = missile(t, x, p, u); MISSILE A non-linear missile system Output equation y = [x(); x(); x(); -p(8)*u()*(p()*x()+p()*u())/p(); -p(8)*u()*(p()*x()+p()*u())/p() ]; % Angular velocity around x % Angular velocity around y % Angular velocity around z % Acceleration in y-directi % Acceleration in z-directi inputs, outputs, 6 unknown parameters % State equations dx = [/p(9)*(p(7)*p(8)*(p()*x()+*p(6)*p(7)*x()/u()+ % Angular p(7)*u())*u()-(p()-p())*x()*x())+ p()*(u(6)-x()); /p()*(p(7)*p(8)*(p(8)*x()+*p(9)*p(7)*x()/u()+ % Angular velocity around y-axis Equations Initial Fit between Model and Data Tunn: measurement; Tjock: simulation p()*u())*u()-(p(9)-p())*x()*x())+ p()*(u(7)-x()); /p()*(p(7)*p(8)*(p()*x()+p()*x()*x()+ % Angular *p()*p(7)*x()/u()+p()*u()+ p()*u())*u()-(p()-p(9))*x()*x())+ p()*(u(8)-x()); (-p(8)*u()*(p()*x()+p()*u()))/(p()*u())- % Attack x()*x()+x()+p()*(u(9)/u()-x())+p(6)*x()ˆ; (-p(8)*u()*(p()*x()+p()*u()))/(p()*u())- % Slide angle x()+x()*x()+p()*(u()/u()-x()) ]; y y y y y [s] Adjusted Fit between Model and Data Semi-physical Models Tunn: measurement; Tjock: simulation y y y y Apply non-linear transformations to the measured data, so that the transformed data stand a better chance to describe the system in a linear relationship Rules: Only high-school physics and max minutes Simple examples: Another example: y [s]

4 A Solar Heated House Data 6 Storage Temperature (degree C) Fan OFF and ON Solar Panel Pump Pump Storage - hours Solar Intensity 8 6 y(t) : temperature in storage, I(t) : Solar intensity u(t) : Pump velocity Linear Model 7 Think 8 Suppose we had measured the temperature x(t) in the solar panel: x(t + ) x(t) = d I(t) d x(t) d x(t) u(t) y(t + ) y(t) = d x(t) u(t) d y(t) Eliminate x(t): y(t )u(t ) y(t) =( + d )y(t ) + ( d ) u(t ) + (d )( + d ) y(t )u(t ) + d d u(t ) I(t ) u(t ) d u(t )y(t ) + d ( + d )u(t )y(t ) tid (timmar) Reparameterize with θ being the coefficiente above, ignoring links between them Semi-physical Model 9 Neural Network Model tid (timmar) tid (timmar) Block-oriented Models Physical Modeling Semi-physical Modeling Block-oriented models Local Linear Models Building Blocks: Linear Dynamic System G(s) Nonlinear static function f(u)

5 Common Models Other Combinations Wiener Hammerstein Hammerstein- Wiener With the linear blocks parameterized as a linear dynamic system and the static blocks parameterized as a function ( curve ), this gives a parameterization of the output as ŷ(t θ) = g(z t, θ) and the general approach of model fitting can be applied However, in this contexts many algorithmic variants have been suggested The Hydraulic Crane Again Local Linear Models 6 Hammerstein model: Fit 76 % Non-linear systems are often handled by linearization around a working point The idea behind Local Linear Models is to deal with the nonlinearities by selecting or averaging over relevant linearized models Example: Tank with inflow u and outflow y and level h: Equations: ḣ = h + u y = h Linearize around level h with corresponding flows u = y = h : ḣ = h (h h ) + (u u ) y = y + h (h h ) Tank Example, ctd 7 Data and Linear Model 8 Sampled data model around level h : Measured data: Linear Model (d = ) ŷ h (t) = θ T h ϕ(t) ϕ(t) = [ y(t T s ) u(t T s ) ] T θ h = [ γ α h β h ] T T s = sampling time Total model: select or average over these local predictions, computed at a grid of values of h Choices of weights w k : ŷ(t) = w k (h, h k )ŷ hk (t) k= Inflöde u Nivå h Local Linear Models 9 Local Linear Models: General Comments Two models (d=) Five models (d = ) Let the measured working point variable (tank level in example) be denoted by ρ(t) (sometimes called regime variable) If the regime variable is partitioned into d values ρ k, the predicted output will be ŷ(t) = w k (ρ(t), ρ k )ŷ (k) (t) k= If the prediction ŷ (k) (t) corresponding to ρ k is linear in the parameters, ŷ (k) (t) = ϕ T (t)θ (k) the whole model will be a linear regression

6 To Build a Local Linear Model Hybrid Models and LPV Models To build the model, we need to Select the regime variable ρ Decide the partition of the regime variable w k (ρ(t), η) Here η is a parameter that describes the partition Find the local models in each partition If the local models are linear regressions, the total model will be ŷ(t, θ, η) = w k (ρ(t), η)ϕ T (t)θ (k) k= is also an example of a hybrid model (piecewise linear) If the partition is to be estimated too, the problem is considerably more difficult ŷ(t, θ, η) = w k (ρ(t), η)ϕ T (t)θ (k) k= which for fixed η is a linear regression So called Linear Parameter Varying (LPV) are also closely related: ẋ = A(ρ(t))x + B(ρ(t))u y = C(ρ(t))x + D(ρ(t))u Experiment Design Input design Sparsity Local minima The design of inputs for non-linear models is considerable more difficult than for linear models For example, it is clear that a binary input will not work (Think of a Hammerstein model!) In addition to exciting all frequencies, an input for a general nonlinear model must also excite all amplitudes Gaussian noise (white or colored) would be an example of a signal that is generically exciting for nonlinear models Sparsity Sparsity: Some ideas 6 The basic problem: Non-linear surfaces in high dimensions can be very complicated and need support of many observed data points How to find parameterizations of such surfaces that both give a good chance of being close to the true system, and also use a moderate amount of parameters? The data cloud of observations is by necessity sparse in the surface space Using physical insight in grey-box models is one way to allow extrapolation and interpolation in the data space on physical grounds Hoping that most of the non-linear action takes place across hyperplanes or hyperspaces is another idea that will radically reduce the flexibility of the model In statistics the problem to find such subspaces is known as the index problem Finding projections S of dimension r m, r << m such that f(ϕ) = g(sϕ) captures most of nonlinearity consequently is an important problem The Ridge-based neural networks can be seen as one way to find several such hyperplanes that define the structure of the non-linear effects Local Minima 7 Conclusions Theme : Nonlinear Models 8 Adjusting a parameterized model structure to data typically is a non-convex problem and several local minima of the criterion function may exist This is one of the most pressing problem in non-linear identification, and calls for sophisticated initialization procedures In Neural Networks, some normalization is first applied to the data, and then a randomized initialization is made Typically one will have to try several initializations Wavenets use an initialization based on fixed location and dilation parameters, which gives a linear regression A nonlinear model can be seen as nonlinear mapping from past data to the space where the output lives: ŷ(t t ) = g(z t, t) Observations are then y(t) = ŷ(t t ) + e(t) Useful split of mapping: g(z t ) = g(ϕ(z t )) Non-parametric and Parametric methods Black-box and Grey-box parameterizations g(ϕ, θ) Black-box parameterizations usually employ one basic basis-function, that is scaled and located at different points For physical models, algebraic methods may produce linear regressions for initial estimates Block oriented models often employ several steps, fixing linear and/or nonlinear block to create smaller problems Grey-boxes can be based on (serious) physical modeling and on more leisurely semi-physical modeling Non-convexity of the optimization remains one of the more serious problems for most parametric methods

The Problem 2(47) Nonlinear System Identification: A Palette from Off-white to Pit-black. A Common Frame 4(47) This Presentation... 3(47) ˆθ = arg min

The Problem 2(47) Nonlinear System Identification: A Palette from Off-white to Pit-black. A Common Frame 4(47) This Presentation... 3(47) ˆθ = arg min y y y y y Tunn: measurement; Tjock: simulation 6 7 8 9.. 6 7 8 9.. 6 7 8 9 6 7 8 9 6 7 8 9 [s] 6 OUTPUT # 6 - INPUT # - 6.. -. OUTPUT # INPUT # - 6 The Problem (7) : A Palette from Off-white to Pit-black

More information

Identification of Non-linear Dynamical Systems

Identification of Non-linear Dynamical Systems Identification of Non-linear Dynamical Systems Linköping University Sweden Prologue Prologue The PI, the Customer and the Data Set C: I have this data set. I have collected it from a cell metabolism experiment.

More information

System Identification: From Data to Model

System Identification: From Data to Model 1 : From Data to Model With Applications to Aircraft Modeling Division of Automatic Control Linköping University Sweden Dynamic Systems 2 A Dynamic system has an output response y that depends on (all)

More information

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. ISS0031 Modeling and Identification Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. Aleksei Tepljakov, Ph.D. October 21, 2015 Discrete-time Transfer Functions

More information

2 Chapter 1 A nonlinear black box structure for a dynamical system is a model structure that is prepared to describe virtually any nonlinear dynamics.

2 Chapter 1 A nonlinear black box structure for a dynamical system is a model structure that is prepared to describe virtually any nonlinear dynamics. 1 SOME ASPECTS O OLIEAR BLACK-BOX MODELIG I SYSTEM IDETIFICATIO Lennart Ljung Dept of Electrical Engineering, Linkoping University, Sweden, ljung@isy.liu.se 1 ITRODUCTIO The key problem in system identication

More information

Chapter 4 Neural Networks in System Identification

Chapter 4 Neural Networks in System Identification Chapter 4 Neural Networks in System Identification Gábor HORVÁTH Department of Measurement and Information Systems Budapest University of Technology and Economics Magyar tudósok körútja 2, 52 Budapest,

More information

J. Sjöberg et al. (1995):Non-linear Black-Box Modeling in System Identification: a Unified Overview, Automatica, Vol. 31, 12, str

J. Sjöberg et al. (1995):Non-linear Black-Box Modeling in System Identification: a Unified Overview, Automatica, Vol. 31, 12, str Dynamic Systems Identification Part - Nonlinear systems Reference: J. Sjöberg et al. (995):Non-linear Black-Box Modeling in System Identification: a Unified Overview, Automatica, Vol. 3,, str. 69-74. Systems

More information

Approaches to Identification of Nonlinear Systems

Approaches to Identification of Nonlinear Systems Approaches to Identification of Nonlinear Systems Lennart Ljung 1 1. Division of Automatic Control, Linköping University, SE-58183 Linköping, Sweden E-mail: ljung@isy.liu.se Abstract: System Identification

More information

Local Modelling of Nonlinear Dynamic Systems Using Direct Weight Optimization

Local Modelling of Nonlinear Dynamic Systems Using Direct Weight Optimization Local Modelling of Nonlinear Dynamic Systems Using Direct Weight Optimization Jacob Roll, Alexander Nazin, Lennart Ljung Division of Automatic Control Department of Electrical Engineering Linköpings universitet,

More information

AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET. Questions AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET

AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET. Questions AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET The Problem Identification of Linear and onlinear Dynamical Systems Theme : Curve Fitting Division of Automatic Control Linköping University Sweden Data from Gripen Questions How do the control surface

More information

COMS 4771 Introduction to Machine Learning. James McInerney Adapted from slides by Nakul Verma

COMS 4771 Introduction to Machine Learning. James McInerney Adapted from slides by Nakul Verma COMS 4771 Introduction to Machine Learning James McInerney Adapted from slides by Nakul Verma Announcements HW1: Please submit as a group Watch out for zero variance features (Q5) HW2 will be released

More information

COMS 4771 Regression. Nakul Verma

COMS 4771 Regression. Nakul Verma COMS 4771 Regression Nakul Verma Last time Support Vector Machines Maximum Margin formulation Constrained Optimization Lagrange Duality Theory Convex Optimization SVM dual and Interpretation How get the

More information

Artificial Neural Networks Francesco DI MAIO, Ph.D., Politecnico di Milano Department of Energy - Nuclear Division IEEE - Italian Reliability Chapter

Artificial Neural Networks Francesco DI MAIO, Ph.D., Politecnico di Milano Department of Energy - Nuclear Division IEEE - Italian Reliability Chapter Artificial Neural Networks Francesco DI MAIO, Ph.D., Politecnico di Milano Department of Energy - Nuclear Division IEEE - Italian Reliability Chapter (Chair) STF - China Fellow francesco.dimaio@polimi.it

More information

IDENTIFICATION OF A PRODUCTION SYSTEM USING HAMMERSTEIN-WIENER AND NARX MODELS

IDENTIFICATION OF A PRODUCTION SYSTEM USING HAMMERSTEIN-WIENER AND NARX MODELS IDENTIFICATION OF A PRODUCTION SYSTEM USING HAMMERSTEIN-WIENER AND NARX MODELS 1 HICHAM FOURAIJI, 2 BAHLOUL BENSASSI 1 laboratory PMMAT university HASSAN II Facultyof science Casablanca, Morocco. 2 laboratory

More information

Just-in-Time Models with Applications to Dynamical Systems

Just-in-Time Models with Applications to Dynamical Systems Linköping Studies in Science and Technology Thesis No. 601 Just-in-Time Models with Applications to Dynamical Systems Anders Stenman REGLERTEKNIK AUTOMATIC CONTROL LINKÖPING Division of Automatic Control

More information

A Mathematica Toolbox for Signals, Models and Identification

A Mathematica Toolbox for Signals, Models and Identification The International Federation of Automatic Control A Mathematica Toolbox for Signals, Models and Identification Håkan Hjalmarsson Jonas Sjöberg ACCESS Linnaeus Center, Electrical Engineering, KTH Royal

More information

Parameter Estimation in a Moving Horizon Perspective

Parameter Estimation in a Moving Horizon Perspective Parameter Estimation in a Moving Horizon Perspective State and Parameter Estimation in Dynamical Systems Reglerteknik, ISY, Linköpings Universitet State and Parameter Estimation in Dynamical Systems OUTLINE

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

1 What a Neural Network Computes

1 What a Neural Network Computes Neural Networks 1 What a Neural Network Computes To begin with, we will discuss fully connected feed-forward neural networks, also known as multilayer perceptrons. A feedforward neural network consists

More information

Stability and the elastic net

Stability and the elastic net Stability and the elastic net Patrick Breheny March 28 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/32 Introduction Elastic Net Our last several lectures have concentrated on methods for

More information

Experiments with neural network for modeling of nonlinear dynamical systems: Design problems

Experiments with neural network for modeling of nonlinear dynamical systems: Design problems Experiments with neural network for modeling of nonlinear dynamical systems: Design problems Ewa Skubalska-Rafaj lowicz Wroc law University of Technology, Wroc law, Wroc law, Poland Summary Introduction

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data January 17, 2006 Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Other Artificial Neural Networks Samy Bengio IDIAP Research Institute, Martigny, Switzerland,

More information

Linearization problem. The simplest example

Linearization problem. The simplest example Linear Systems Lecture 3 1 problem Consider a non-linear time-invariant system of the form ( ẋ(t f x(t u(t y(t g ( x(t u(t (1 such that x R n u R m y R p and Slide 1 A: f(xu f(xu g(xu and g(xu exist and

More information

Modeling nonlinear systems using multiple piecewise linear equations

Modeling nonlinear systems using multiple piecewise linear equations Nonlinear Analysis: Modelling and Control, 2010, Vol. 15, No. 4, 451 458 Modeling nonlinear systems using multiple piecewise linear equations G.K. Lowe, M.A. Zohdy Department of Electrical and Computer

More information

Linear Models for Regression. Sargur Srihari

Linear Models for Regression. Sargur Srihari Linear Models for Regression Sargur srihari@cedar.buffalo.edu 1 Topics in Linear Regression What is regression? Polynomial Curve Fitting with Scalar input Linear Basis Function Models Maximum Likelihood

More information

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection Instructor: Herke van Hoof (herke.vanhoof@cs.mcgill.ca) Based on slides by:, Jackie Chi Kit Cheung Class web page:

More information

Statistics 910, #15 1. Kalman Filter

Statistics 910, #15 1. Kalman Filter Statistics 910, #15 1 Overview 1. Summary of Kalman filter 2. Derivations 3. ARMA likelihoods 4. Recursions for the variance Kalman Filter Summary of Kalman filter Simplifications To make the derivations

More information

Data Driven Modelling for Complex Systems

Data Driven Modelling for Complex Systems Data Driven Modelling for Complex Systems Dr Hua-Liang (Leon) Wei Senior Lecturer in System Identification and Data Analytics Head of Dynamic Modelling, Data Mining & Decision Making (3DM) Lab Complex

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction

More information

Midterm Review CS 7301: Advanced Machine Learning. Vibhav Gogate The University of Texas at Dallas

Midterm Review CS 7301: Advanced Machine Learning. Vibhav Gogate The University of Texas at Dallas Midterm Review CS 7301: Advanced Machine Learning Vibhav Gogate The University of Texas at Dallas Supervised Learning Issues in supervised learning What makes learning hard Point Estimation: MLE vs Bayesian

More information

CHAPTER 10: Numerical Methods for DAEs

CHAPTER 10: Numerical Methods for DAEs CHAPTER 10: Numerical Methods for DAEs Numerical approaches for the solution of DAEs divide roughly into two classes: 1. direct discretization 2. reformulation (index reduction) plus discretization Direct

More information

NEURAL LANGUAGE MODELS

NEURAL LANGUAGE MODELS COMP90042 LECTURE 14 NEURAL LANGUAGE MODELS LANGUAGE MODELS Assign a probability to a sequence of words Framed as sliding a window over the sentence, predicting each word from finite context to left E.g.,

More information

Contents lecture 6 2(17) Automatic Control III. Summary of lecture 5 (I/III) 3(17) Summary of lecture 5 (II/III) 4(17) H 2, H synthesis pros and cons:

Contents lecture 6 2(17) Automatic Control III. Summary of lecture 5 (I/III) 3(17) Summary of lecture 5 (II/III) 4(17) H 2, H synthesis pros and cons: Contents lecture 6 (7) Automatic Control III Lecture 6 Linearization and phase portraits. Summary of lecture 5 Thomas Schön Division of Systems and Control Department of Information Technology Uppsala

More information

Adaptive Dual Control

Adaptive Dual Control Adaptive Dual Control Björn Wittenmark Department of Automatic Control, Lund Institute of Technology Box 118, S-221 00 Lund, Sweden email: bjorn@control.lth.se Keywords: Dual control, stochastic control,

More information

Control Systems Lab - SC4070 System Identification and Linearization

Control Systems Lab - SC4070 System Identification and Linearization Control Systems Lab - SC4070 System Identification and Linearization Dr. Manuel Mazo Jr. Delft Center for Systems and Control (TU Delft) m.mazo@tudelft.nl Tel.:015-2788131 TU Delft, February 13, 2015 (slides

More information

Linear Models for Regression

Linear Models for Regression Linear Models for Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

Lecture 17: Neural Networks and Deep Learning

Lecture 17: Neural Networks and Deep Learning UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions

More information

Lecture: Section In this section, we will consider more rigorously the question of whether a vector field is path dependent.

Lecture: Section In this section, we will consider more rigorously the question of whether a vector field is path dependent. Lecture: Section 184 Path Dependent Fields and Green s Theorem In this section, we will consider more rigorously the question of whether a vector field is path dependent Graphical Interpretation We know

More information

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011 Compressed Sensing Huichao Xue CS3750 Fall 2011 Table of Contents Introduction From News Reports Abstract Definition How it works A review of L 1 norm The Algorithm Backgrounds for underdetermined linear

More information

Data Driven Discrete Time Modeling of Continuous Time Nonlinear Systems. Problems, Challenges, Success Stories. Johan Schoukens

Data Driven Discrete Time Modeling of Continuous Time Nonlinear Systems. Problems, Challenges, Success Stories. Johan Schoukens 1/51 Data Driven Discrete Time Modeling of Continuous Time Nonlinear Systems Problems, Challenges, Success Stories Johan Schoukens fyuq (,, ) 4/51 System Identification Data Distance Model 5/51 System

More information

Automatic Differentiation and Neural Networks

Automatic Differentiation and Neural Networks Statistical Machine Learning Notes 7 Automatic Differentiation and Neural Networks Instructor: Justin Domke 1 Introduction The name neural network is sometimes used to refer to many things (e.g. Hopfield

More information

Probabilistic & Bayesian deep learning. Andreas Damianou

Probabilistic & Bayesian deep learning. Andreas Damianou Probabilistic & Bayesian deep learning Andreas Damianou Amazon Research Cambridge, UK Talk at University of Sheffield, 19 March 2019 In this talk Not in this talk: CRFs, Boltzmann machines,... In this

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

APPLICATION OF A KERNEL METHOD IN MODELING FRICTION DYNAMICS

APPLICATION OF A KERNEL METHOD IN MODELING FRICTION DYNAMICS APPLICATION OF A KERNEL METHOD IN MODELING FRICTION DYNAMICS Yufeng Wan, Chian X. Wong, Tony J. Dodd, Robert F. Harrison Department of Automatic Control and Systems Engineering, The University of Sheffield,

More information

Towards airborne seismic. Thomas Rapstine & Paul Sava

Towards airborne seismic. Thomas Rapstine & Paul Sava Towards airborne seismic Thomas Rapstine & Paul Sava Research goal measure a strong motion signal from an airborne platform 2 Proposed method Monitor ground motion from an UAV using cameras Unmanned Aerial

More information

Optimization Methods for Machine Learning (OMML)

Optimization Methods for Machine Learning (OMML) Optimization Methods for Machine Learning (OMML) 2nd lecture (2 slots) Prof. L. Palagi 16/10/2014 1 What is (not) Data Mining? By Namwar Rizvi - Ad Hoc Query: ad Hoc queries just examines the current data

More information

Notes on Discriminant Functions and Optimal Classification

Notes on Discriminant Functions and Optimal Classification Notes on Discriminant Functions and Optimal Classification Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Discriminant Functions Consider a classification problem

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Machine Learning Lecture 5

Machine Learning Lecture 5 Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

More information

Computing Neural Network Gradients

Computing Neural Network Gradients Computing Neural Network Gradients Kevin Clark 1 Introduction The purpose of these notes is to demonstrate how to quickly compute neural network gradients in a completely vectorized way. It is complementary

More information

Neural-wavelet Methodology for Load Forecasting

Neural-wavelet Methodology for Load Forecasting Journal of Intelligent and Robotic Systems 31: 149 157, 2001. 2001 Kluwer Academic Publishers. Printed in the Netherlands. 149 Neural-wavelet Methodology for Load Forecasting RONG GAO and LEFTERI H. TSOUKALAS

More information

Modeling and Experimentation: Compound Pendulum

Modeling and Experimentation: Compound Pendulum Modeling and Experimentation: Compound Pendulum Prof. R.G. Longoria Department of Mechanical Engineering The University of Texas at Austin Fall 2014 Overview This lab focuses on developing a mathematical

More information

A Few Games and Geometric Insights

A Few Games and Geometric Insights A Few Games and Geometric Insights Brian Powers Arizona State University brpowers@asu.edu January 20, 2017 1 / 56 Outline 1 Nash/Correlated Equilibria A Motivating Example The Nash Equilibrium Correlated

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Generalized projective synchronization between two chaotic gyros with nonlinear damping

Generalized projective synchronization between two chaotic gyros with nonlinear damping Generalized projective synchronization between two chaotic gyros with nonlinear damping Min Fu-Hong( ) Department of Electrical and Automation Engineering, Nanjing Normal University, Nanjing 210042, China

More information

ADAPTIVE NEURO-FUZZY INFERENCE SYSTEMS

ADAPTIVE NEURO-FUZZY INFERENCE SYSTEMS ADAPTIVE NEURO-FUZZY INFERENCE SYSTEMS RBFN and TS systems Equivalent if the following hold: Both RBFN and TS use same aggregation method for output (weighted sum or weighted average) Number of basis functions

More information

Topic 3: Neural Networks

Topic 3: Neural Networks CS 4850/6850: Introduction to Machine Learning Fall 2018 Topic 3: Neural Networks Instructor: Daniel L. Pimentel-Alarcón c Copyright 2018 3.1 Introduction Neural networks are arguably the main reason why

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology 6.867 Machine Learning, Fall 2006 Problem Set 5 Due Date: Thursday, Nov 30, 12:00 noon You may submit your solutions in class or in the box. 1. Wilhelm and Klaus are

More information

Hammerstein-Wiener Model: A New Approach to the Estimation of Formal Neural Information

Hammerstein-Wiener Model: A New Approach to the Estimation of Formal Neural Information Hammerstein-Wiener Model: A New Approach to the Estimation of Formal Neural Information Reza Abbasi-Asl1, Rahman Khorsandi2, Bijan Vosooghi-Vahdat1 1. Biomedical Signal and Image Processing Laboratory

More information

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas Midterm Review CS 6375: Machine Learning Vibhav Gogate The University of Texas at Dallas Machine Learning Supervised Learning Unsupervised Learning Reinforcement Learning Parametric Y Continuous Non-parametric

More information

Introduction to Biomedical Engineering

Introduction to Biomedical Engineering Introduction to Biomedical Engineering Biosignal processing Kung-Bin Sung 6/11/2007 1 Outline Chapter 10: Biosignal processing Characteristics of biosignals Frequency domain representation and analysis

More information

Neural Networks. Volker Tresp Summer 2015

Neural Networks. Volker Tresp Summer 2015 Neural Networks Volker Tresp Summer 2015 1 Introduction The performance of a classifier or a regression model critically depends on the choice of appropriate basis functions The problem with generic basis

More information

Nonlinear Regression

Nonlinear Regression Nonlinear Regression 28.11.2012 Goals of Today s Lecture Understand the difference between linear and nonlinear regression models. See that not all functions are linearizable. Get an understanding of the

More information

Intelligent Systems and Control Prof. Laxmidhar Behera Indian Institute of Technology, Kanpur

Intelligent Systems and Control Prof. Laxmidhar Behera Indian Institute of Technology, Kanpur Intelligent Systems and Control Prof. Laxmidhar Behera Indian Institute of Technology, Kanpur Module - 2 Lecture - 4 Introduction to Fuzzy Logic Control In this lecture today, we will be discussing fuzzy

More information

Linear Regression 1 / 25. Karl Stratos. June 18, 2018

Linear Regression 1 / 25. Karl Stratos. June 18, 2018 Linear Regression Karl Stratos June 18, 2018 1 / 25 The Regression Problem Problem. Find a desired input-output mapping f : X R where the output is a real value. x = = y = 0.1 How much should I turn my

More information

Lecture 3 Feedforward Networks and Backpropagation

Lecture 3 Feedforward Networks and Backpropagation Lecture 3 Feedforward Networks and Backpropagation CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago April 3, 2017 Things we will look at today Recap of Logistic Regression

More information

Output Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x)

Output Feedback and State Feedback. EL2620 Nonlinear Control. Nonlinear Observers. Nonlinear Controllers. ẋ = f(x,u), y = h(x) Output Feedback and State Feedback EL2620 Nonlinear Control Lecture 10 Exact feedback linearization Input-output linearization Lyapunov-based control design methods ẋ = f(x,u) y = h(x) Output feedback:

More information

Adaptive sparse grids

Adaptive sparse grids ANZIAM J. 44 (E) ppc335 C353, 2003 C335 Adaptive sparse grids M. Hegland (Received 1 June 2001) Abstract Sparse grids, as studied by Zenger and Griebel in the last 10 years have been very successful in

More information

2D Wavelets. Hints on advanced Concepts

2D Wavelets. Hints on advanced Concepts 2D Wavelets Hints on advanced Concepts 1 Advanced concepts Wavelet packets Laplacian pyramid Overcomplete bases Discrete wavelet frames (DWF) Algorithme à trous Discrete dyadic wavelet frames (DDWF) Overview

More information

Input-output data sets for development and benchmarking in nonlinear identification

Input-output data sets for development and benchmarking in nonlinear identification Input-output data sets for development and benchmarking in nonlinear identification Torbjörn Wigren Division of Systems and Control, Department of Information Technology, Uppsala University, PO box 337,

More information

Advances in Extreme Learning Machines

Advances in Extreme Learning Machines Advances in Extreme Learning Machines Mark van Heeswijk April 17, 2015 Outline Context Extreme Learning Machines Part I: Ensemble Models of ELM Part II: Variable Selection and ELM Part III: Trade-offs

More information

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Lecture - 27 Multilayer Feedforward Neural networks with Sigmoidal

More information

Lecture 6. Regression

Lecture 6. Regression Lecture 6. Regression Prof. Alan Yuille Summer 2014 Outline 1. Introduction to Regression 2. Binary Regression 3. Linear Regression; Polynomial Regression 4. Non-linear Regression; Multilayer Perceptron

More information

1 Conjugate gradients

1 Conjugate gradients Notes for 2016-11-18 1 Conjugate gradients We now turn to the method of conjugate gradients (CG), perhaps the best known of the Krylov subspace solvers. The CG iteration can be characterized as the iteration

More information

Nonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto

Nonlinear System Identification Using MLP Dr.-Ing. Sudchai Boonto Dr-Ing Sudchai Boonto Department of Control System and Instrumentation Engineering King Mongkut s Unniversity of Technology Thonburi Thailand Nonlinear System Identification Given a data set Z N = {y(k),

More information

LECTURE NOTE #NEW 6 PROF. ALAN YUILLE

LECTURE NOTE #NEW 6 PROF. ALAN YUILLE LECTURE NOTE #NEW 6 PROF. ALAN YUILLE 1. Introduction to Regression Now consider learning the conditional distribution p(y x). This is often easier than learning the likelihood function p(x y) and the

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

Regression with Numerical Optimization. Logistic

Regression with Numerical Optimization. Logistic CSG220 Machine Learning Fall 2008 Regression with Numerical Optimization. Logistic regression Regression with Numerical Optimization. Logistic regression based on a document by Andrew Ng October 3, 204

More information

Dynamic Data Modeling of SCR De-NOx System Based on NARX Neural Network Wen-jie ZHAO * and Kai ZHANG

Dynamic Data Modeling of SCR De-NOx System Based on NARX Neural Network Wen-jie ZHAO * and Kai ZHANG 2018 International Conference on Modeling, Simulation and Analysis (ICMSA 2018) ISBN: 978-1-60595-544-5 Dynamic Data Modeling of SCR De-NOx System Based on NARX Neural Network Wen-jie ZHAO * and Kai ZHANG

More information

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function

More information

Introduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim

Introduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim Introduction - Motivation Many phenomena (physical, chemical, biological, etc.) are model by differential equations. Recall the definition of the derivative of f(x) f f(x + h) f(x) (x) = lim. h 0 h Its

More information

Using Neural Networks for Identification and Control of Systems

Using Neural Networks for Identification and Control of Systems Using Neural Networks for Identification and Control of Systems Jhonatam Cordeiro Department of Industrial and Systems Engineering North Carolina A&T State University, Greensboro, NC 27411 jcrodrig@aggies.ncat.edu

More information

Consider the following example of a linear system:

Consider the following example of a linear system: LINEAR SYSTEMS Consider the following example of a linear system: Its unique solution is x + 2x 2 + 3x 3 = 5 x + x 3 = 3 3x + x 2 + 3x 3 = 3 x =, x 2 = 0, x 3 = 2 In general we want to solve n equations

More information

Deep Feedforward Networks

Deep Feedforward Networks Deep Feedforward Networks Yongjin Park 1 Goal of Feedforward Networks Deep Feedforward Networks are also called as Feedforward neural networks or Multilayer Perceptrons Their Goal: approximate some function

More information

Statistical Methods for SVM

Statistical Methods for SVM Statistical Methods for SVM Support Vector Machines Here we approach the two-class classification problem in a direct way: We try and find a plane that separates the classes in feature space. If we cannot,

More information

Ensemble Methods and Random Forests

Ensemble Methods and Random Forests Ensemble Methods and Random Forests Vaishnavi S May 2017 1 Introduction We have seen various analysis for classification and regression in the course. One of the common methods to reduce the generalization

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information

CHAPTER 2: QUADRATIC PROGRAMMING

CHAPTER 2: QUADRATIC PROGRAMMING CHAPTER 2: QUADRATIC PROGRAMMING Overview Quadratic programming (QP) problems are characterized by objective functions that are quadratic in the design variables, and linear constraints. In this sense,

More information

A Generalization of Principal Component Analysis to the Exponential Family

A Generalization of Principal Component Analysis to the Exponential Family A Generalization of Principal Component Analysis to the Exponential Family Michael Collins Sanjoy Dasgupta Robert E. Schapire AT&T Labs Research 8 Park Avenue, Florham Park, NJ 7932 mcollins, dasgupta,

More information

Local Modelling with A Priori Known Bounds Using Direct Weight Optimization

Local Modelling with A Priori Known Bounds Using Direct Weight Optimization Local Modelling with A Priori Known Bounds Using Direct Weight Optimization Jacob Roll, Alexander azin, Lennart Ljung Division of Automatic Control Department of Electrical Engineering Linköpings universitet,

More information

Nonlinear seismic imaging via reduced order model backprojection

Nonlinear seismic imaging via reduced order model backprojection Nonlinear seismic imaging via reduced order model backprojection Alexander V. Mamonov, Vladimir Druskin 2 and Mikhail Zaslavsky 2 University of Houston, 2 Schlumberger-Doll Research Center Mamonov, Druskin,

More information

Computational statistics

Computational statistics Computational statistics Lecture 3: Neural networks Thierry Denœux 5 March, 2016 Neural networks A class of learning methods that was developed separately in different fields statistics and artificial

More information

Introduction to Modelling and Simulation

Introduction to Modelling and Simulation Introduction to Modelling and Simulation Prof. Cesar de Prada Dpt. Systems Engineering and Automatic Control EII, University of Valladolid, Spain prada@autom.uva.es Digital simulation Methods and tools

More information

Nonlinear Identification of Backlash in Robot Transmissions

Nonlinear Identification of Backlash in Robot Transmissions Nonlinear Identification of Backlash in Robot Transmissions G. Hovland, S. Hanssen, S. Moberg, T. Brogårdh, S. Gunnarsson, M. Isaksson ABB Corporate Research, Control Systems Group, Switzerland ABB Automation

More information

Observability. Dynamic Systems. Lecture 2 Observability. Observability, continuous time: Observability, discrete time: = h (2) (x, u, u)

Observability. Dynamic Systems. Lecture 2 Observability. Observability, continuous time: Observability, discrete time: = h (2) (x, u, u) Observability Dynamic Systems Lecture 2 Observability Continuous time model: Discrete time model: ẋ(t) = f (x(t), u(t)), y(t) = h(x(t), u(t)) x(t + 1) = f (x(t), u(t)), y(t) = h(x(t)) Reglerteknik, ISY,

More information

MATH 680 Fall November 27, Homework 3

MATH 680 Fall November 27, Homework 3 MATH 680 Fall 208 November 27, 208 Homework 3 This homework is due on December 9 at :59pm. Provide both pdf, R files. Make an individual R file with proper comments for each sub-problem. Subgradients and

More information

Statistical Data Mining and Machine Learning Hilary Term 2016

Statistical Data Mining and Machine Learning Hilary Term 2016 Statistical Data Mining and Machine Learning Hilary Term 2016 Dino Sejdinovic Department of Statistics Oxford Slides and other materials available at: http://www.stats.ox.ac.uk/~sejdinov/sdmml Naïve Bayes

More information

Maps and differential equations

Maps and differential equations Maps and differential equations Marc R. Roussel November 8, 2005 Maps are algebraic rules for computing the next state of dynamical systems in discrete time. Differential equations and maps have a number

More information