A structured low-rank approximation approach to system identification. Ivan Markovsky

Size: px
Start display at page:

Download "A structured low-rank approximation approach to system identification. Ivan Markovsky"

Transcription

1 1 / 35 A structured low-rank approximation approach to system identification Ivan Markovsky

2 Main message: system identification SLRA minimize over B dist(a,b) subject to rank(b) r and B structured (SLRA) SLRA problem A dist(, ) r structure system identification observed data noise properties model complexity model class 2 / 35

3 3 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification

4 4 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification

5 5 / 35 Identification: finding models from data data D identification model B M aim: "accurate" and "simple" model "accurate" "simple" smallest approximation error Occam s razor principle: among equally accurate models, choose the simplest

6 6 / 35 Data D: set of vector-valued time series the data D is a set {w 1,...,w N } of vector valued w k = w k 1.. w k q time series w k i = ( w k i (1),...,w k i (T k ) ) N # of repeated experiments q # of variables T k # of time samples in kth exp.

7 7 / 35 Model B: subset of the data space behavioral definition of a model B = { w g(w)=0 holds } g(w)=0 representation of B model class M : set of models L set of linear models

8 8 / 35 dist(d,b):= min D B k w k ŵ k w B ŵ k w k ŵ k w k w 1 errors-in-variables model: data = true value + noise other error measures: output error, ARMAX,...

9 9 / 35 Model complexity = (# inputs, # states) simple = small ( B1 B 2 = B 1 is simpler than B 2 ) linear model is subspace, then size of B dimension of B linear time-invariant (LTI) dynamic model dimension of B ( # inputs, # }{{}} states {{} ) m l L m,l LTI systems of bounded complexity

10 10 / 35 Identification: error-complexity trade-off data D identification model B M minimize over B M [ ] dist(d,b) complexity(b)

11 11 / 35 Scalarization of the bi-objective problem 1. minimize dist(d, B) + λ complexity(b) minimize λ complexity(b) subject to dist(d, B) µ minimize dist(d, B) subject to complexity(b) (m, l) describe the same set of Pareto optimal solutions withmgiven, finding l is an order selection problem

12 12 / 35 LTI identification problem minimize over B dist(d,b) subject to B L m,l with distance measure dist(d,b)= min k w k ŵ k 2 2 = min D D D B D B the problem is minimize over B, D D D subject to D B Lm,l

13 13 / 35 w exact rank deficient Hankel matrix exact trajectory w B L m,l R 0 w(t)+r 1 w(t+ 1)+ +R l w(t+l)=0 w(1) w(2) w(t l) [ ] w(2) w(3) w(t l+1) R0 R 1 R l }{{}... = 0 R w(l+1) w(l+2) }{{ w(t) } H l+1 (w) rank ( H l+1 (w) ) = q(l+1) rank(r)=ql+m

14 14 / 35 D exact rank ( H l+1 (D) ) ql+m exact data D B L m,l w k B L m,l for all k = 1,...,N rank ([ H l+1 (w 1 ) H l+1 (w N ) ] ) ql+m }{{} mosaic-hankel matrix H l+1 (D)

15 15 / 35 LTI identification is mosaic-hankel SLRA minimize over B and D D D subject to D B Lm,l minimize over D D D subject to rank ( H l+1 ( D) ) ql+m

16 16 / 35 Summary: identification SLRA LTI model class bounded complexity Hankel structure rank constraint (m,l) ql+m

17 17 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification

18 Motivation goes beyond data corruption sensor failures measurements are accidentally corrupted compressive sensing measurements are intentionally skipped data-driven estimation and control missing data is what we aim to find examples: state estimation, control, and realization 18 / 35

19 1. state estimation given: system B, input u, and output y missing: initial conditions wini such that minimize over w ini, ŷ y ŷ }{{} s.t. estimation error w ini (u,ŷ) B 2. output tracking control given: B, wini, and reference output y ref missing: control input u such that minimize over û, ŷ y ref ŷ }{{} tracking error s.t. w ini (û,ŷ) B 3. realization given: impulse response h(1),...,h(t) missing: extension h(t + 1),h(T + 2), / 35

20 20 / 35 Exact, noisy, and missing data exact data kept fixed inexact / "noisy" data approximated (min error 2 ) missing data interpolated from w B the initial conditions w ini are the "past" of w w w ini w past future t

21 21 / state estimation past future input? u output? y 2. output tracking control past future input u ini? output y ini y ref 3. (noisy) realization past future input 0 δ output 0 (h,?) black exact blue inexact/noisy red missing

22 22 / 35 SLRA with element-wise weighted 2-norm minimize over D dist(d, D) subject to rank ( H l+1 (ŵ) ) ql+m weighted 2-norm approximation dist(d, D):= k,i,t vi k (t) ( wi k (t) ŵi k (t) ) 2 with element-wise weights vi k (t) (0, ) if wi k (t) is noisy vi k (t)=0 if wi k (t) is missing v k i (t)= if w k i (t) is exact approximate wi k (t) interpolate wi k (t) ŵi k (t)=wi k (t)

23 23 / 35 Summary: SLRA solves control problems the given data is exact or noisy what we want to compute is missing data exact/noisy/missing data is handled by /finite/0 weights

24 24 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification

25 25 / 35 Conic section fitting the points (x 1,y 1 ),...,(x N,y N ) lie on a conic section there are A=A, b, c, at least one of them nonzero, s.t. [ ] [ ] xi xi y i A + [ ] x y i y i b+ c = 0, for i = 1,...,N i there is θ = [ a 11 a 12 a 22 b 1 b 2 c ] 0, such that x1 2 x 2 N x 1 y 1 x N y N θ x 1 x N y1 2 yn 2 = 0 y 1 y N 1 1

26 26 / 35 Conic section fitting rank deficiency the points (x 1,y 1 ),...,(x N,y N ) lie on a conic section B(θ)={w w Aw+ w b+ c = 0} x1 2 x 2 N x 1 y 1 x N y N rank x 1 x N y1 2 yn 2 5 y 1 y N 1 1

27 27 / 35 Examples rank < 5 = nonunique fit rank = 5 = unique fit rank = 6 = no exact fit by a conic section

28 28 / 35 Nonlinear system identification discrete-time nonlinear system B := { w R ( w(t),w(t 1),...,w(t l) ) = 0 } special case: input/output NARX system B = { w = [ ] ( )} u y y(t)=f u(t),w(t 1),...,w(t l) linear parameterization: B θ φ model structure R(x)= θ i φ i (x)=θφ(x), θ parameter vector x(t):= ( w(t),w(t 1),...,w(t l) )

29 29 / 35 Link to SLRA parameter estimation problem ŵ B θ minimize over θ and ŵ w ŵ subject to ŵ B θ (NL SYSID) ( [φ rank ( x(1) ) φ ( x(t l) )] ) r }{{} polynomially structured matrix Φ(ŵ) (NL SYSID) polynomially structured LRA (NL SYSID) is nonconvex and yields biased estimator

30 30 / 35 Bias correction ignoring the structure of Φ(ŵ) leads to kernel PCA easy to compute, but biased in the EIV model w = w+ w, where w B and w N(0,σ 2 I) define Ψ:=Φ(w)Φ (w) and Ψ:=Φ( w)φ ( w) goal: construct corrected matrix Ψ c, such that E(Ψ c )= Ψ

31 31 / 35 Derivation of the correction Hermite polynomials h k (x) have the property E ( h k ( x+ x) ) = x k, where x N(0,σ 2 ) ( ) with w =(x,y), the (i,j)th element of Ψ=ΦΦ is ( x+ x) n x,i+n x,j (ȳ+ ỹ) n y,i+n y,j then, by ( ) φ c,ij := h nx,i +n x,j (x)h ny,i +n y,j (y) has the desired property E(ψ c,ij )= x n x,i+n x,jȳn y,i +n y,j =: ψ ij

32 32 / 35 Unbiased estimator the corrected Ψ c is an even polynomial in σ Ψ c (σ 2 )=Ψ c,0 + σ 2 Ψ c,1 + +σ 2n ψ Ψ c,nψ estimate: Ψ c (σ 2 )θ = 0 computing simultaneously σ and θ is polynomial EVP examples of static nonlinear model fitting KPCA dotted PLRA dashed-dotted bias corrected dashed

33 33 / 35 2 Example: x 3 + y 3 3xy = 0 1 y x

34 34 / 35 Conclusion: system identification SLRA LTI model class mosaic-hankel structure solving control problems as missing data estimation bias correction procedure for polynomial SLRA

35 35 / 35 Conclusion: system identification SLRA LTI model class mosaic-hankel structure solving control problems as missing data estimation bias correction procedure for polynomial SLRA papers, course materials, and code at:

Data-driven signal processing

Data-driven signal processing 1 / 35 Data-driven signal processing Ivan Markovsky 2 / 35 Modern signal processing is model-based 1. system identification prior information model structure 2. model-based design identification data parameter

More information

ELEC system identification workshop. Behavioral approach

ELEC system identification workshop. Behavioral approach 1 / 33 ELEC system identification workshop Behavioral approach Ivan Markovsky 2 / 33 The course consists of lectures and exercises Session 1: behavioral approach to data modeling Session 2: subspace identification

More information

A software package for system identification in the behavioral setting

A software package for system identification in the behavioral setting 1 / 20 A software package for system identification in the behavioral setting Ivan Markovsky Vrije Universiteit Brussel 2 / 20 Outline Introduction: system identification in the behavioral setting Solution

More information

ELEC system identification workshop. Subspace methods

ELEC system identification workshop. Subspace methods 1 / 33 ELEC system identification workshop Subspace methods Ivan Markovsky 2 / 33 Plan 1. Behavioral approach 2. Subspace methods 3. Optimization methods 3 / 33 Outline Exact modeling Algorithms 4 / 33

More information

Low-rank approximation and its applications for data fitting

Low-rank approximation and its applications for data fitting Low-rank approximation and its applications for data fitting Ivan Markovsky K.U.Leuven, ESAT-SISTA A line fitting example b 6 4 2 0 data points Classical problem: Fit the points d 1 = [ 0 6 ], d2 = [ 1

More information

Two well known examples. Applications of structured low-rank approximation. Approximate realisation = Model reduction. System realisation

Two well known examples. Applications of structured low-rank approximation. Approximate realisation = Model reduction. System realisation Two well known examples Applications of structured low-rank approximation Ivan Markovsky System realisation Discrete deconvolution School of Electronics and Computer Science University of Southampton The

More information

Sparsity in system identification and data-driven control

Sparsity in system identification and data-driven control 1 / 40 Sparsity in system identification and data-driven control Ivan Markovsky This signal is not sparse in the "time domain" 2 / 40 But it is sparse in the "frequency domain" (it is weighted sum of six

More information

A missing data approach to data-driven filtering and control

A missing data approach to data-driven filtering and control 1 A missing data approach to data-driven filtering and control Ivan Markovsky Abstract In filtering, control, and other mathematical engineering areas it is common to use a model-based approach, which

More information

Dynamic measurement: application of system identification in metrology

Dynamic measurement: application of system identification in metrology 1 / 25 Dynamic measurement: application of system identification in metrology Ivan Markovsky Dynamic measurement takes into account the dynamical properties of the sensor 2 / 25 model of sensor as dynamical

More information

Structured low-rank approximation with missing data

Structured low-rank approximation with missing data Structured low-rank approximation with missing data Ivan Markovsky and Konstantin Usevich Department ELEC Vrije Universiteit Brussel {imarkovs,kusevich}@vub.ac.be Abstract We consider low-rank approximation

More information

Improved initial approximation for errors-in-variables system identification

Improved initial approximation for errors-in-variables system identification Improved initial approximation for errors-in-variables system identification Konstantin Usevich Abstract Errors-in-variables system identification can be posed and solved as a Hankel structured low-rank

More information

Chapter 1 Nonlinearly structured low-rank approximation

Chapter 1 Nonlinearly structured low-rank approximation Chapter 1 onlinearly structured low-rank approximation Ivan Markovsky and Konstantin Usevich Abstract Polynomially structured low-rank approximation problems occur in algebraic curve fitting, e.g., conic

More information

A comparison between structured low-rank approximation and correlation approach for data-driven output tracking

A comparison between structured low-rank approximation and correlation approach for data-driven output tracking A comparison between structured low-rank approximation and correlation approach for data-driven output tracking Simone Formentin a and Ivan Markovsky b a Dipartimento di Elettronica, Informazione e Bioingegneria,

More information

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e.

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e. Least Squares with Eamples in Signal Processing Ivan Selesnick March 7, 3 NYU-Poly These notes address (approimate) solutions to linear equations by least squares We deal with the easy case wherein the

More information

Optimization on the Grassmann manifold: a case study

Optimization on the Grassmann manifold: a case study Optimization on the Grassmann manifold: a case study Konstantin Usevich and Ivan Markovsky Department ELEC, Vrije Universiteit Brussel 28 March 2013 32nd Benelux Meeting on Systems and Control, Houffalize,

More information

Using Hankel structured low-rank approximation for sparse signal recovery

Using Hankel structured low-rank approximation for sparse signal recovery Using Hankel structured low-rank approximation for sparse signal recovery Ivan Markovsky 1 and Pier Luigi Dragotti 2 Department ELEC Vrije Universiteit Brussel (VUB) Pleinlaan 2, Building K, B-1050 Brussels,

More information

Outline lecture 2 2(30)

Outline lecture 2 2(30) Outline lecture 2 2(3), Lecture 2 Linear Regression it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic Control

More information

AN IDENTIFICATION ALGORITHM FOR ARMAX SYSTEMS

AN IDENTIFICATION ALGORITHM FOR ARMAX SYSTEMS AN IDENTIFICATION ALGORITHM FOR ARMAX SYSTEMS First the X, then the AR, finally the MA Jan C. Willems, K.U. Leuven Workshop on Observation and Estimation Ben Gurion University, July 3, 2004 p./2 Joint

More information

Generalization theory

Generalization theory Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1

More information

Differential and Difference LTI systems

Differential and Difference LTI systems Signals and Systems Lecture: 6 Differential and Difference LTI systems Differential and difference linear time-invariant (LTI) systems constitute an extremely important class of systems in engineering.

More information

4 Bias-Variance for Ridge Regression (24 points)

4 Bias-Variance for Ridge Regression (24 points) Implement Ridge Regression with λ = 0.00001. Plot the Squared Euclidean test error for the following values of k (the dimensions you reduce to): k = {0, 50, 100, 150, 200, 250, 300, 350, 400, 450, 500,

More information

LECTURE NOTE #11 PROF. ALAN YUILLE

LECTURE NOTE #11 PROF. ALAN YUILLE LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform

More information

Low-Rank Approximation

Low-Rank Approximation Ivan Markovsky Low-Rank Approximation Algorithms, Implementation, Applications September 2, 2014 Springer Preface Mathematical models are obtained from first principles (natural laws, interconnection,

More information

Realization and identification of autonomous linear periodically time-varying systems

Realization and identification of autonomous linear periodically time-varying systems Realization and identification of autonomous linear periodically time-varying systems Ivan Markovsky, Jan Goos, Konstantin Usevich, and Rik Pintelon Department ELEC, Vrije Universiteit Brussel, Pleinlaan

More information

Geometric Interpolation by Planar Cubic Polynomials

Geometric Interpolation by Planar Cubic Polynomials 1 / 20 Geometric Interpolation by Planar Cubic Polynomials Jernej Kozak, Marjeta Krajnc Faculty of Mathematics and Physics University of Ljubljana Institute of Mathematics, Physics and Mechanics Avignon,

More information

Lecture 5 Least-squares

Lecture 5 Least-squares EE263 Autumn 2008-09 Stephen Boyd Lecture 5 Least-squares least-squares (approximate) solution of overdetermined equations projection and orthogonality principle least-squares estimation BLUE property

More information

MCE693/793: Analysis and Control of Nonlinear Systems

MCE693/793: Analysis and Control of Nonlinear Systems MCE693/793: Analysis and Control of Nonlinear Systems Input-Output and Input-State Linearization Zero Dynamics of Nonlinear Systems Hanz Richter Mechanical Engineering Department Cleveland State University

More information

14 - Gaussian Stochastic Processes

14 - Gaussian Stochastic Processes 14-1 Gaussian Stochastic Processes S. Lall, Stanford 211.2.24.1 14 - Gaussian Stochastic Processes Linear systems driven by IID noise Evolution of mean and covariance Example: mass-spring system Steady-state

More information

3 Gramians and Balanced Realizations

3 Gramians and Balanced Realizations 3 Gramians and Balanced Realizations In this lecture, we use an optimization approach to find suitable realizations for truncation and singular perturbation of G. It turns out that the recommended realizations

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Lecture Notes of EE 714

Lecture Notes of EE 714 Lecture Notes of EE 714 Lecture 1 Motivation Systems theory that we have studied so far deals with the notion of specified input and output spaces. But there are systems which do not have a clear demarcation

More information

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.

Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. ISS0031 Modeling and Identification Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. Aleksei Tepljakov, Ph.D. October 21, 2015 Discrete-time Transfer Functions

More information

Lecture 6. Regularized least-squares and minimum-norm methods 6 1

Lecture 6. Regularized least-squares and minimum-norm methods 6 1 Regularized least-squares and minimum-norm methods 6 1 Lecture 6 Regularized least-squares and minimum-norm methods EE263 Autumn 2004 multi-objective least-squares regularized least-squares nonlinear least-squares

More information

A Behavioral Approach to GNSS Positioning and DOP Determination

A Behavioral Approach to GNSS Positioning and DOP Determination A Behavioral Approach to GNSS Positioning and DOP Determination Department of Communications and Guidance Engineering National Taiwan Ocean University, Keelung, TAIWAN Phone: +886-96654, FAX: +886--463349

More information

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION M. Schwab, P. Noll, and T. Sikora Technical University Berlin, Germany Communication System Group Einsteinufer 17, 1557 Berlin (Germany) {schwab noll

More information

EECE Adaptive Control

EECE Adaptive Control EECE 574 - Adaptive Control Basics of System Identification Guy Dumont Department of Electrical and Computer Engineering University of British Columbia January 2010 Guy Dumont (UBC) EECE574 - Basics of

More information

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am ECS130 Scientific Computing Lecture 1: Introduction Monday, January 7, 10:00 10:50 am About Course: ECS130 Scientific Computing Professor: Zhaojun Bai Webpage: http://web.cs.ucdavis.edu/~bai/ecs130/ Today

More information

Outline Lecture 2 2(32)

Outline Lecture 2 2(32) Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic

More information

Stochastic gradient descent on Riemannian manifolds

Stochastic gradient descent on Riemannian manifolds Stochastic gradient descent on Riemannian manifolds Silvère Bonnabel 1 Robotics lab - Mathématiques et systèmes Mines ParisTech Gipsa-lab, Grenoble June 20th, 2013 1 silvere.bonnabel@mines-paristech Introduction

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

S. Saboktakin and B. Kordi * Department of Electrical and Computer Engineering, University of Manitoba Winnipeg, MB R3T 5V6, Canada

S. Saboktakin and B. Kordi * Department of Electrical and Computer Engineering, University of Manitoba Winnipeg, MB R3T 5V6, Canada Progress In Electromagnetics Research Letters, Vol. 32, 109 118, 2012 DISTORTION ANALYSIS OF ELECTROMAGNETIC FIELD SENSORS IN LAGUERRE FUNCTIONS SUBSPACE S. Saboktakin and B. Kordi * Department of Electrical

More information

4 Bias-Variance for Ridge Regression (24 points)

4 Bias-Variance for Ridge Regression (24 points) 2 count = 0 3 for x in self.x_test_ridge: 4 5 prediction = np.matmul(self.w_ridge,x) 6 ###ADD THE COMPUTED MEAN BACK TO THE PREDICTED VECTOR### 7 prediction = self.ss_y.inverse_transform(prediction) 8

More information

Computational Methods. Least Squares Approximation/Optimization

Computational Methods. Least Squares Approximation/Optimization Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the

More information

Least-squares data fitting

Least-squares data fitting EE263 Autumn 2015 S. Boyd and S. Lall Least-squares data fitting 1 Least-squares data fitting we are given: functions f 1,..., f n : S R, called regressors or basis functions data or measurements (s i,

More information

Lecture 11 FIR Filters

Lecture 11 FIR Filters Lecture 11 FIR Filters Fundamentals of Digital Signal Processing Spring, 2012 Wei-Ta Chu 2012/4/12 1 The Unit Impulse Sequence Any sequence can be represented in this way. The equation is true if k ranges

More information

Closed-Loop Identification of Unstable Systems Using Noncausal FIR Models

Closed-Loop Identification of Unstable Systems Using Noncausal FIR Models 23 American Control Conference (ACC) Washington, DC, USA, June 7-9, 23 Closed-Loop Identification of Unstable Systems Using Noncausal FIR Models Khaled Aljanaideh, Benjamin J. Coffer, and Dennis S. Bernstein

More information

Set-Membership Identification of Wiener models with noninvertible nonlinearity

Set-Membership Identification of Wiener models with noninvertible nonlinearity Set-Membership Identification of Wiener models with noninvertible nonlinearity V. Cerone, Dipartimento di Automatica e Informatica Politecnico di Torino (Italy) e-mail: vito.cerone@polito.it, diego.regruto@polito.it

More information

Parameter Estimation in a Moving Horizon Perspective

Parameter Estimation in a Moving Horizon Perspective Parameter Estimation in a Moving Horizon Perspective State and Parameter Estimation in Dynamical Systems Reglerteknik, ISY, Linköpings Universitet State and Parameter Estimation in Dynamical Systems OUTLINE

More information

Linear model selection and regularization

Linear model selection and regularization Linear model selection and regularization Problems with linear regression with least square 1. Prediction Accuracy: linear regression has low bias but suffer from high variance, especially when n p. It

More information

Stochastic gradient descent on Riemannian manifolds

Stochastic gradient descent on Riemannian manifolds Stochastic gradient descent on Riemannian manifolds Silvère Bonnabel 1 Centre de Robotique - Mathématiques et systèmes Mines ParisTech SMILE Seminar Mines ParisTech Novembre 14th, 2013 1 silvere.bonnabel@mines-paristech

More information

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY

More information

Regression Estimation Least Squares and Maximum Likelihood

Regression Estimation Least Squares and Maximum Likelihood Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize

More information

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives

More information

Chapter 4 Neural Networks in System Identification

Chapter 4 Neural Networks in System Identification Chapter 4 Neural Networks in System Identification Gábor HORVÁTH Department of Measurement and Information Systems Budapest University of Technology and Economics Magyar tudósok körútja 2, 52 Budapest,

More information

Accelerated Block-Coordinate Relaxation for Regularized Optimization

Accelerated Block-Coordinate Relaxation for Regularized Optimization Accelerated Block-Coordinate Relaxation for Regularized Optimization Stephen J. Wright Computer Sciences University of Wisconsin, Madison October 09, 2012 Problem descriptions Consider where f is smooth

More information

Notes for System Identification: Impulse Response Functions via Wavelet

Notes for System Identification: Impulse Response Functions via Wavelet Notes for System Identification: Impulse Response Functions via Wavelet 1 Basic Wavelet Algorithm for IRF Extraction In contrast to the FFT-based extraction procedure which must process the data both in

More information

10. Linear Models and Maximum Likelihood Estimation

10. Linear Models and Maximum Likelihood Estimation 10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that

More information

Probabilistic & Bayesian deep learning. Andreas Damianou

Probabilistic & Bayesian deep learning. Andreas Damianou Probabilistic & Bayesian deep learning Andreas Damianou Amazon Research Cambridge, UK Talk at University of Sheffield, 19 March 2019 In this talk Not in this talk: CRFs, Boltzmann machines,... In this

More information

Identification of ARX, OE, FIR models with the least squares method

Identification of ARX, OE, FIR models with the least squares method Identification of ARX, OE, FIR models with the least squares method CHEM-E7145 Advanced Process Control Methods Lecture 2 Contents Identification of ARX model with the least squares minimizing the equation

More information

Positive systems in the behavioral approach: main issues and recent results

Positive systems in the behavioral approach: main issues and recent results Positive systems in the behavioral approach: main issues and recent results Maria Elena Valcher Dipartimento di Elettronica ed Informatica Università dipadova via Gradenigo 6A 35131 Padova Italy Abstract

More information

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction Estimation Tasks Short Course on Image Quality Matthew A. Kupinski Introduction Section 13.3 in B&M Keep in mind the similarities between estimation and classification Image-quality is a statistical concept

More information

EL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1)

EL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1) EL 625 Lecture 0 EL 625 Lecture 0 Pole Placement and Observer Design Pole Placement Consider the system ẋ Ax () The solution to this system is x(t) e At x(0) (2) If the eigenvalues of A all lie in the

More information

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A. A EW IFORMATIO THEORETIC APPROACH TO ORDER ESTIMATIO PROBLEM Soosan Beheshti Munther A. Dahleh Massachusetts Institute of Technology, Cambridge, MA 0239, U.S.A. Abstract: We introduce a new method of model

More information

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing

More information

GAUSSIAN PROCESS REGRESSION

GAUSSIAN PROCESS REGRESSION GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The

More information

Manifold Learning: Theory and Applications to HRI

Manifold Learning: Theory and Applications to HRI Manifold Learning: Theory and Applications to HRI Seungjin Choi Department of Computer Science Pohang University of Science and Technology, Korea seungjin@postech.ac.kr August 19, 2008 1 / 46 Greek Philosopher

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed

More information

Math 2433 Notes Week The Dot Product. The angle between two vectors is found with this formula: cosθ = a b

Math 2433 Notes Week The Dot Product. The angle between two vectors is found with this formula: cosθ = a b Math 2433 Notes Week 2 11.3 The Dot Product The angle between two vectors is found with this formula: cosθ = a b a b 3) Given, a = 4i + 4j, b = i - 2j + 3k, c = 2i + 2k Find the angle between a and c Projection

More information

CME 345: MODEL REDUCTION

CME 345: MODEL REDUCTION CME 345: MODEL REDUCTION Proper Orthogonal Decomposition (POD) Charbel Farhat & David Amsallem Stanford University cfarhat@stanford.edu 1 / 43 Outline 1 Time-continuous Formulation 2 Method of Snapshots

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Kernel Principal Component Analysis

Kernel Principal Component Analysis Kernel Principal Component Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

A rational interpolation scheme with super-polynomial rate of convergence

A rational interpolation scheme with super-polynomial rate of convergence Center for Turbulence Research Annual Research Briefs 2008 31 A rational interpolation scheme with super-polynomial rate of convergence By Q. Wang, P. Moin AND G. Iaccarino 1. Motivation and objectives

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

8. Diagonalization.

8. Diagonalization. 8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard

More information

Chapter 8: Estimation 1

Chapter 8: Estimation 1 Chapter 8: Estimation 1 Jae-Kwang Kim Iowa State University Fall, 2014 Kim (ISU) Ch. 8: Estimation 1 Fall, 2014 1 / 33 Introduction 1 Introduction 2 Ratio estimation 3 Regression estimator Kim (ISU) Ch.

More information

SPARSE SIGNAL RESTORATION. 1. Introduction

SPARSE SIGNAL RESTORATION. 1. Introduction SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful

More information

Lecture 3 January 23

Lecture 3 January 23 EE 123: Digital Signal Processing Spring 2007 Lecture 3 January 23 Lecturer: Prof. Anant Sahai Scribe: Dominic Antonelli 3.1 Outline These notes cover the following topics: Eigenvectors and Eigenvalues

More information

arxiv: v2 [math.oc] 18 Jan 2015

arxiv: v2 [math.oc] 18 Jan 2015 Regularization for Design Nikolai Matni and Venkat Chandrasekaran January 20, 2015 arxiv:1404.1972v2 [math.oc] 18 Jan 2015 bstr When designing controllers for large-scale systems, the architectural aspects

More information

CSE446: non-parametric methods Spring 2017

CSE446: non-parametric methods Spring 2017 CSE446: non-parametric methods Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin and Luke Zettlemoyer Linear Regression: What can go wrong? What do we do if the bias is too strong? Might want

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 18 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 31, 2012 Andre Tkacenko

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

Advanced Machine Learning Practical 4b Solution: Regression (BLR, GPR & Gradient Boosting)

Advanced Machine Learning Practical 4b Solution: Regression (BLR, GPR & Gradient Boosting) Advanced Machine Learning Practical 4b Solution: Regression (BLR, GPR & Gradient Boosting) Professor: Aude Billard Assistants: Nadia Figueroa, Ilaria Lauzana and Brice Platerrier E-mails: aude.billard@epfl.ch,

More information

CS 340 Lec. 15: Linear Regression

CS 340 Lec. 15: Linear Regression CS 340 Lec. 15: Linear Regression AD February 2011 AD () February 2011 1 / 31 Regression Assume you are given some training data { x i, y i } N where x i R d and y i R c. Given an input test data x, you

More information

Lecture 6: Linear models and Gauss-Markov theorem

Lecture 6: Linear models and Gauss-Markov theorem Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables

More information

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random

More information

Identification of Non-linear Dynamical Systems

Identification of Non-linear Dynamical Systems Identification of Non-linear Dynamical Systems Linköping University Sweden Prologue Prologue The PI, the Customer and the Data Set C: I have this data set. I have collected it from a cell metabolism experiment.

More information

and u and v are orthogonal if and only if u v = 0. u v = x1x2 + y1y2 + z1z2. 1. In R 3 the dot product is defined by

and u and v are orthogonal if and only if u v = 0. u v = x1x2 + y1y2 + z1z2. 1. In R 3 the dot product is defined by Linear Algebra [] 4.2 The Dot Product and Projections. In R 3 the dot product is defined by u v = u v cos θ. 2. For u = (x, y, z) and v = (x2, y2, z2), we have u v = xx2 + yy2 + zz2. 3. cos θ = u v u v,

More information

Event-Triggered Decentralized Dynamic Output Feedback Control for LTI Systems

Event-Triggered Decentralized Dynamic Output Feedback Control for LTI Systems Event-Triggered Decentralized Dynamic Output Feedback Control for LTI Systems Pavankumar Tallapragada Nikhil Chopra Department of Mechanical Engineering, University of Maryland, College Park, 2742 MD,

More information

(a). W contains the zero vector in R n. (b). W is closed under addition. (c). W is closed under scalar multiplication.

(a). W contains the zero vector in R n. (b). W is closed under addition. (c). W is closed under scalar multiplication. . Subspaces of R n Bases and Linear Independence Definition. Subspaces of R n A subset W of R n is called a subspace of R n if it has the following properties: (a). W contains the zero vector in R n. (b).

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

System Identification: An Inverse Problem in Control

System Identification: An Inverse Problem in Control System Identification: An Inverse Problem in Control Potentials and Possibilities of Regularization Lennart Ljung Linköping University, Sweden Inverse Problems, MAI, Linköping, April 4, 2013 System Identification

More information

4. DATA ASSIMILATION FUNDAMENTALS

4. DATA ASSIMILATION FUNDAMENTALS 4. DATA ASSIMILATION FUNDAMENTALS... [the atmosphere] "is a chaotic system in which errors introduced into the system can grow with time... As a consequence, data assimilation is a struggle between chaotic

More information

Input: A set (x i -yy i ) data. Output: Function value at arbitrary point x. What for x = 1.2?

Input: A set (x i -yy i ) data. Output: Function value at arbitrary point x. What for x = 1.2? Applied Numerical Analysis Interpolation Lecturer: Emad Fatemizadeh Interpolation Input: A set (x i -yy i ) data. Output: Function value at arbitrary point x. 0 1 4 1-3 3 9 What for x = 1.? Interpolation

More information

REGULARIZED STRUCTURED LOW-RANK APPROXIMATION WITH APPLICATIONS

REGULARIZED STRUCTURED LOW-RANK APPROXIMATION WITH APPLICATIONS REGULARIZED STRUCTURED LOW-RANK APPROXIMATION WITH APPLICATIONS MARIYA ISHTEVA, KONSTANTIN USEVICH, AND IVAN MARKOVSKY Abstract. We consider the problem of approximating an affinely structured matrix,

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil

Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil Yuejie Chi Departments of ECE and BMI The Ohio State University Colorado School of Mines December 9, 24 Page Acknowledgement Joint work

More information

EL1820 Modeling of Dynamical Systems

EL1820 Modeling of Dynamical Systems EL1820 Modeling of Dynamical Systems Lecture 10 - System identification as a model building tool Experiment design Examination and prefiltering of data Model structure selection Model validation Lecture

More information

10. Multi-objective least squares

10. Multi-objective least squares L Vandenberghe ECE133A (Winter 2018) 10 Multi-objective least squares multi-objective least squares regularized data fitting control estimation and inversion 10-1 Multi-objective least squares we have

More information

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane

More information