A structured low-rank approximation approach to system identification. Ivan Markovsky
|
|
- Ashley Richards
- 5 years ago
- Views:
Transcription
1 1 / 35 A structured low-rank approximation approach to system identification Ivan Markovsky
2 Main message: system identification SLRA minimize over B dist(a,b) subject to rank(b) r and B structured (SLRA) SLRA problem A dist(, ) r structure system identification observed data noise properties model complexity model class 2 / 35
3 3 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification
4 4 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification
5 5 / 35 Identification: finding models from data data D identification model B M aim: "accurate" and "simple" model "accurate" "simple" smallest approximation error Occam s razor principle: among equally accurate models, choose the simplest
6 6 / 35 Data D: set of vector-valued time series the data D is a set {w 1,...,w N } of vector valued w k = w k 1.. w k q time series w k i = ( w k i (1),...,w k i (T k ) ) N # of repeated experiments q # of variables T k # of time samples in kth exp.
7 7 / 35 Model B: subset of the data space behavioral definition of a model B = { w g(w)=0 holds } g(w)=0 representation of B model class M : set of models L set of linear models
8 8 / 35 dist(d,b):= min D B k w k ŵ k w B ŵ k w k ŵ k w k w 1 errors-in-variables model: data = true value + noise other error measures: output error, ARMAX,...
9 9 / 35 Model complexity = (# inputs, # states) simple = small ( B1 B 2 = B 1 is simpler than B 2 ) linear model is subspace, then size of B dimension of B linear time-invariant (LTI) dynamic model dimension of B ( # inputs, # }{{}} states {{} ) m l L m,l LTI systems of bounded complexity
10 10 / 35 Identification: error-complexity trade-off data D identification model B M minimize over B M [ ] dist(d,b) complexity(b)
11 11 / 35 Scalarization of the bi-objective problem 1. minimize dist(d, B) + λ complexity(b) minimize λ complexity(b) subject to dist(d, B) µ minimize dist(d, B) subject to complexity(b) (m, l) describe the same set of Pareto optimal solutions withmgiven, finding l is an order selection problem
12 12 / 35 LTI identification problem minimize over B dist(d,b) subject to B L m,l with distance measure dist(d,b)= min k w k ŵ k 2 2 = min D D D B D B the problem is minimize over B, D D D subject to D B Lm,l
13 13 / 35 w exact rank deficient Hankel matrix exact trajectory w B L m,l R 0 w(t)+r 1 w(t+ 1)+ +R l w(t+l)=0 w(1) w(2) w(t l) [ ] w(2) w(3) w(t l+1) R0 R 1 R l }{{}... = 0 R w(l+1) w(l+2) }{{ w(t) } H l+1 (w) rank ( H l+1 (w) ) = q(l+1) rank(r)=ql+m
14 14 / 35 D exact rank ( H l+1 (D) ) ql+m exact data D B L m,l w k B L m,l for all k = 1,...,N rank ([ H l+1 (w 1 ) H l+1 (w N ) ] ) ql+m }{{} mosaic-hankel matrix H l+1 (D)
15 15 / 35 LTI identification is mosaic-hankel SLRA minimize over B and D D D subject to D B Lm,l minimize over D D D subject to rank ( H l+1 ( D) ) ql+m
16 16 / 35 Summary: identification SLRA LTI model class bounded complexity Hankel structure rank constraint (m,l) ql+m
17 17 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification
18 Motivation goes beyond data corruption sensor failures measurements are accidentally corrupted compressive sensing measurements are intentionally skipped data-driven estimation and control missing data is what we aim to find examples: state estimation, control, and realization 18 / 35
19 1. state estimation given: system B, input u, and output y missing: initial conditions wini such that minimize over w ini, ŷ y ŷ }{{} s.t. estimation error w ini (u,ŷ) B 2. output tracking control given: B, wini, and reference output y ref missing: control input u such that minimize over û, ŷ y ref ŷ }{{} tracking error s.t. w ini (û,ŷ) B 3. realization given: impulse response h(1),...,h(t) missing: extension h(t + 1),h(T + 2), / 35
20 20 / 35 Exact, noisy, and missing data exact data kept fixed inexact / "noisy" data approximated (min error 2 ) missing data interpolated from w B the initial conditions w ini are the "past" of w w w ini w past future t
21 21 / state estimation past future input? u output? y 2. output tracking control past future input u ini? output y ini y ref 3. (noisy) realization past future input 0 δ output 0 (h,?) black exact blue inexact/noisy red missing
22 22 / 35 SLRA with element-wise weighted 2-norm minimize over D dist(d, D) subject to rank ( H l+1 (ŵ) ) ql+m weighted 2-norm approximation dist(d, D):= k,i,t vi k (t) ( wi k (t) ŵi k (t) ) 2 with element-wise weights vi k (t) (0, ) if wi k (t) is noisy vi k (t)=0 if wi k (t) is missing v k i (t)= if w k i (t) is exact approximate wi k (t) interpolate wi k (t) ŵi k (t)=wi k (t)
23 23 / 35 Summary: SLRA solves control problems the given data is exact or noisy what we want to compute is missing data exact/noisy/missing data is handled by /finite/0 weights
24 24 / 35 Plan of the presentation System identification low-rank approximation Missing data estimation Nonlinear system identification
25 25 / 35 Conic section fitting the points (x 1,y 1 ),...,(x N,y N ) lie on a conic section there are A=A, b, c, at least one of them nonzero, s.t. [ ] [ ] xi xi y i A + [ ] x y i y i b+ c = 0, for i = 1,...,N i there is θ = [ a 11 a 12 a 22 b 1 b 2 c ] 0, such that x1 2 x 2 N x 1 y 1 x N y N θ x 1 x N y1 2 yn 2 = 0 y 1 y N 1 1
26 26 / 35 Conic section fitting rank deficiency the points (x 1,y 1 ),...,(x N,y N ) lie on a conic section B(θ)={w w Aw+ w b+ c = 0} x1 2 x 2 N x 1 y 1 x N y N rank x 1 x N y1 2 yn 2 5 y 1 y N 1 1
27 27 / 35 Examples rank < 5 = nonunique fit rank = 5 = unique fit rank = 6 = no exact fit by a conic section
28 28 / 35 Nonlinear system identification discrete-time nonlinear system B := { w R ( w(t),w(t 1),...,w(t l) ) = 0 } special case: input/output NARX system B = { w = [ ] ( )} u y y(t)=f u(t),w(t 1),...,w(t l) linear parameterization: B θ φ model structure R(x)= θ i φ i (x)=θφ(x), θ parameter vector x(t):= ( w(t),w(t 1),...,w(t l) )
29 29 / 35 Link to SLRA parameter estimation problem ŵ B θ minimize over θ and ŵ w ŵ subject to ŵ B θ (NL SYSID) ( [φ rank ( x(1) ) φ ( x(t l) )] ) r }{{} polynomially structured matrix Φ(ŵ) (NL SYSID) polynomially structured LRA (NL SYSID) is nonconvex and yields biased estimator
30 30 / 35 Bias correction ignoring the structure of Φ(ŵ) leads to kernel PCA easy to compute, but biased in the EIV model w = w+ w, where w B and w N(0,σ 2 I) define Ψ:=Φ(w)Φ (w) and Ψ:=Φ( w)φ ( w) goal: construct corrected matrix Ψ c, such that E(Ψ c )= Ψ
31 31 / 35 Derivation of the correction Hermite polynomials h k (x) have the property E ( h k ( x+ x) ) = x k, where x N(0,σ 2 ) ( ) with w =(x,y), the (i,j)th element of Ψ=ΦΦ is ( x+ x) n x,i+n x,j (ȳ+ ỹ) n y,i+n y,j then, by ( ) φ c,ij := h nx,i +n x,j (x)h ny,i +n y,j (y) has the desired property E(ψ c,ij )= x n x,i+n x,jȳn y,i +n y,j =: ψ ij
32 32 / 35 Unbiased estimator the corrected Ψ c is an even polynomial in σ Ψ c (σ 2 )=Ψ c,0 + σ 2 Ψ c,1 + +σ 2n ψ Ψ c,nψ estimate: Ψ c (σ 2 )θ = 0 computing simultaneously σ and θ is polynomial EVP examples of static nonlinear model fitting KPCA dotted PLRA dashed-dotted bias corrected dashed
33 33 / 35 2 Example: x 3 + y 3 3xy = 0 1 y x
34 34 / 35 Conclusion: system identification SLRA LTI model class mosaic-hankel structure solving control problems as missing data estimation bias correction procedure for polynomial SLRA
35 35 / 35 Conclusion: system identification SLRA LTI model class mosaic-hankel structure solving control problems as missing data estimation bias correction procedure for polynomial SLRA papers, course materials, and code at:
Data-driven signal processing
1 / 35 Data-driven signal processing Ivan Markovsky 2 / 35 Modern signal processing is model-based 1. system identification prior information model structure 2. model-based design identification data parameter
More informationELEC system identification workshop. Behavioral approach
1 / 33 ELEC system identification workshop Behavioral approach Ivan Markovsky 2 / 33 The course consists of lectures and exercises Session 1: behavioral approach to data modeling Session 2: subspace identification
More informationA software package for system identification in the behavioral setting
1 / 20 A software package for system identification in the behavioral setting Ivan Markovsky Vrije Universiteit Brussel 2 / 20 Outline Introduction: system identification in the behavioral setting Solution
More informationELEC system identification workshop. Subspace methods
1 / 33 ELEC system identification workshop Subspace methods Ivan Markovsky 2 / 33 Plan 1. Behavioral approach 2. Subspace methods 3. Optimization methods 3 / 33 Outline Exact modeling Algorithms 4 / 33
More informationLow-rank approximation and its applications for data fitting
Low-rank approximation and its applications for data fitting Ivan Markovsky K.U.Leuven, ESAT-SISTA A line fitting example b 6 4 2 0 data points Classical problem: Fit the points d 1 = [ 0 6 ], d2 = [ 1
More informationTwo well known examples. Applications of structured low-rank approximation. Approximate realisation = Model reduction. System realisation
Two well known examples Applications of structured low-rank approximation Ivan Markovsky System realisation Discrete deconvolution School of Electronics and Computer Science University of Southampton The
More informationSparsity in system identification and data-driven control
1 / 40 Sparsity in system identification and data-driven control Ivan Markovsky This signal is not sparse in the "time domain" 2 / 40 But it is sparse in the "frequency domain" (it is weighted sum of six
More informationA missing data approach to data-driven filtering and control
1 A missing data approach to data-driven filtering and control Ivan Markovsky Abstract In filtering, control, and other mathematical engineering areas it is common to use a model-based approach, which
More informationDynamic measurement: application of system identification in metrology
1 / 25 Dynamic measurement: application of system identification in metrology Ivan Markovsky Dynamic measurement takes into account the dynamical properties of the sensor 2 / 25 model of sensor as dynamical
More informationStructured low-rank approximation with missing data
Structured low-rank approximation with missing data Ivan Markovsky and Konstantin Usevich Department ELEC Vrije Universiteit Brussel {imarkovs,kusevich}@vub.ac.be Abstract We consider low-rank approximation
More informationImproved initial approximation for errors-in-variables system identification
Improved initial approximation for errors-in-variables system identification Konstantin Usevich Abstract Errors-in-variables system identification can be posed and solved as a Hankel structured low-rank
More informationChapter 1 Nonlinearly structured low-rank approximation
Chapter 1 onlinearly structured low-rank approximation Ivan Markovsky and Konstantin Usevich Abstract Polynomially structured low-rank approximation problems occur in algebraic curve fitting, e.g., conic
More informationA comparison between structured low-rank approximation and correlation approach for data-driven output tracking
A comparison between structured low-rank approximation and correlation approach for data-driven output tracking Simone Formentin a and Ivan Markovsky b a Dipartimento di Elettronica, Informazione e Bioingegneria,
More informationLeast Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e.
Least Squares with Eamples in Signal Processing Ivan Selesnick March 7, 3 NYU-Poly These notes address (approimate) solutions to linear equations by least squares We deal with the easy case wherein the
More informationOptimization on the Grassmann manifold: a case study
Optimization on the Grassmann manifold: a case study Konstantin Usevich and Ivan Markovsky Department ELEC, Vrije Universiteit Brussel 28 March 2013 32nd Benelux Meeting on Systems and Control, Houffalize,
More informationUsing Hankel structured low-rank approximation for sparse signal recovery
Using Hankel structured low-rank approximation for sparse signal recovery Ivan Markovsky 1 and Pier Luigi Dragotti 2 Department ELEC Vrije Universiteit Brussel (VUB) Pleinlaan 2, Building K, B-1050 Brussels,
More informationOutline lecture 2 2(30)
Outline lecture 2 2(3), Lecture 2 Linear Regression it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic Control
More informationAN IDENTIFICATION ALGORITHM FOR ARMAX SYSTEMS
AN IDENTIFICATION ALGORITHM FOR ARMAX SYSTEMS First the X, then the AR, finally the MA Jan C. Willems, K.U. Leuven Workshop on Observation and Estimation Ben Gurion University, July 3, 2004 p./2 Joint
More informationGeneralization theory
Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1
More informationDifferential and Difference LTI systems
Signals and Systems Lecture: 6 Differential and Difference LTI systems Differential and difference linear time-invariant (LTI) systems constitute an extremely important class of systems in engineering.
More information4 Bias-Variance for Ridge Regression (24 points)
Implement Ridge Regression with λ = 0.00001. Plot the Squared Euclidean test error for the following values of k (the dimensions you reduce to): k = {0, 50, 100, 150, 200, 250, 300, 350, 400, 450, 500,
More informationLECTURE NOTE #11 PROF. ALAN YUILLE
LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform
More informationLow-Rank Approximation
Ivan Markovsky Low-Rank Approximation Algorithms, Implementation, Applications September 2, 2014 Springer Preface Mathematical models are obtained from first principles (natural laws, interconnection,
More informationRealization and identification of autonomous linear periodically time-varying systems
Realization and identification of autonomous linear periodically time-varying systems Ivan Markovsky, Jan Goos, Konstantin Usevich, and Rik Pintelon Department ELEC, Vrije Universiteit Brussel, Pleinlaan
More informationGeometric Interpolation by Planar Cubic Polynomials
1 / 20 Geometric Interpolation by Planar Cubic Polynomials Jernej Kozak, Marjeta Krajnc Faculty of Mathematics and Physics University of Ljubljana Institute of Mathematics, Physics and Mechanics Avignon,
More informationLecture 5 Least-squares
EE263 Autumn 2008-09 Stephen Boyd Lecture 5 Least-squares least-squares (approximate) solution of overdetermined equations projection and orthogonality principle least-squares estimation BLUE property
More informationMCE693/793: Analysis and Control of Nonlinear Systems
MCE693/793: Analysis and Control of Nonlinear Systems Input-Output and Input-State Linearization Zero Dynamics of Nonlinear Systems Hanz Richter Mechanical Engineering Department Cleveland State University
More information14 - Gaussian Stochastic Processes
14-1 Gaussian Stochastic Processes S. Lall, Stanford 211.2.24.1 14 - Gaussian Stochastic Processes Linear systems driven by IID noise Evolution of mean and covariance Example: mass-spring system Steady-state
More information3 Gramians and Balanced Realizations
3 Gramians and Balanced Realizations In this lecture, we use an optimization approach to find suitable realizations for truncation and singular perturbation of G. It turns out that the recommended realizations
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationLecture Notes of EE 714
Lecture Notes of EE 714 Lecture 1 Motivation Systems theory that we have studied so far deals with the notion of specified input and output spaces. But there are systems which do not have a clear demarcation
More informationLecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data.
ISS0031 Modeling and Identification Lecture 7: Discrete-time Models. Modeling of Physical Systems. Preprocessing Experimental Data. Aleksei Tepljakov, Ph.D. October 21, 2015 Discrete-time Transfer Functions
More informationLecture 6. Regularized least-squares and minimum-norm methods 6 1
Regularized least-squares and minimum-norm methods 6 1 Lecture 6 Regularized least-squares and minimum-norm methods EE263 Autumn 2004 multi-objective least-squares regularized least-squares nonlinear least-squares
More informationA Behavioral Approach to GNSS Positioning and DOP Determination
A Behavioral Approach to GNSS Positioning and DOP Determination Department of Communications and Guidance Engineering National Taiwan Ocean University, Keelung, TAIWAN Phone: +886-96654, FAX: +886--463349
More informationNOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group
NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION M. Schwab, P. Noll, and T. Sikora Technical University Berlin, Germany Communication System Group Einsteinufer 17, 1557 Berlin (Germany) {schwab noll
More informationEECE Adaptive Control
EECE 574 - Adaptive Control Basics of System Identification Guy Dumont Department of Electrical and Computer Engineering University of British Columbia January 2010 Guy Dumont (UBC) EECE574 - Basics of
More informationECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am
ECS130 Scientific Computing Lecture 1: Introduction Monday, January 7, 10:00 10:50 am About Course: ECS130 Scientific Computing Professor: Zhaojun Bai Webpage: http://web.cs.ucdavis.edu/~bai/ecs130/ Today
More informationOutline Lecture 2 2(32)
Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic
More informationStochastic gradient descent on Riemannian manifolds
Stochastic gradient descent on Riemannian manifolds Silvère Bonnabel 1 Robotics lab - Mathématiques et systèmes Mines ParisTech Gipsa-lab, Grenoble June 20th, 2013 1 silvere.bonnabel@mines-paristech Introduction
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationS. Saboktakin and B. Kordi * Department of Electrical and Computer Engineering, University of Manitoba Winnipeg, MB R3T 5V6, Canada
Progress In Electromagnetics Research Letters, Vol. 32, 109 118, 2012 DISTORTION ANALYSIS OF ELECTROMAGNETIC FIELD SENSORS IN LAGUERRE FUNCTIONS SUBSPACE S. Saboktakin and B. Kordi * Department of Electrical
More information4 Bias-Variance for Ridge Regression (24 points)
2 count = 0 3 for x in self.x_test_ridge: 4 5 prediction = np.matmul(self.w_ridge,x) 6 ###ADD THE COMPUTED MEAN BACK TO THE PREDICTED VECTOR### 7 prediction = self.ss_y.inverse_transform(prediction) 8
More informationComputational Methods. Least Squares Approximation/Optimization
Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the
More informationLeast-squares data fitting
EE263 Autumn 2015 S. Boyd and S. Lall Least-squares data fitting 1 Least-squares data fitting we are given: functions f 1,..., f n : S R, called regressors or basis functions data or measurements (s i,
More informationLecture 11 FIR Filters
Lecture 11 FIR Filters Fundamentals of Digital Signal Processing Spring, 2012 Wei-Ta Chu 2012/4/12 1 The Unit Impulse Sequence Any sequence can be represented in this way. The equation is true if k ranges
More informationClosed-Loop Identification of Unstable Systems Using Noncausal FIR Models
23 American Control Conference (ACC) Washington, DC, USA, June 7-9, 23 Closed-Loop Identification of Unstable Systems Using Noncausal FIR Models Khaled Aljanaideh, Benjamin J. Coffer, and Dennis S. Bernstein
More informationSet-Membership Identification of Wiener models with noninvertible nonlinearity
Set-Membership Identification of Wiener models with noninvertible nonlinearity V. Cerone, Dipartimento di Automatica e Informatica Politecnico di Torino (Italy) e-mail: vito.cerone@polito.it, diego.regruto@polito.it
More informationParameter Estimation in a Moving Horizon Perspective
Parameter Estimation in a Moving Horizon Perspective State and Parameter Estimation in Dynamical Systems Reglerteknik, ISY, Linköpings Universitet State and Parameter Estimation in Dynamical Systems OUTLINE
More informationLinear model selection and regularization
Linear model selection and regularization Problems with linear regression with least square 1. Prediction Accuracy: linear regression has low bias but suffer from high variance, especially when n p. It
More informationStochastic gradient descent on Riemannian manifolds
Stochastic gradient descent on Riemannian manifolds Silvère Bonnabel 1 Centre de Robotique - Mathématiques et systèmes Mines ParisTech SMILE Seminar Mines ParisTech Novembre 14th, 2013 1 silvere.bonnabel@mines-paristech
More informationCOMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017
COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY
More informationRegression Estimation Least Squares and Maximum Likelihood
Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize
More informationRestricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model
Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives
More informationChapter 4 Neural Networks in System Identification
Chapter 4 Neural Networks in System Identification Gábor HORVÁTH Department of Measurement and Information Systems Budapest University of Technology and Economics Magyar tudósok körútja 2, 52 Budapest,
More informationAccelerated Block-Coordinate Relaxation for Regularized Optimization
Accelerated Block-Coordinate Relaxation for Regularized Optimization Stephen J. Wright Computer Sciences University of Wisconsin, Madison October 09, 2012 Problem descriptions Consider where f is smooth
More informationNotes for System Identification: Impulse Response Functions via Wavelet
Notes for System Identification: Impulse Response Functions via Wavelet 1 Basic Wavelet Algorithm for IRF Extraction In contrast to the FFT-based extraction procedure which must process the data both in
More information10. Linear Models and Maximum Likelihood Estimation
10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that
More informationProbabilistic & Bayesian deep learning. Andreas Damianou
Probabilistic & Bayesian deep learning Andreas Damianou Amazon Research Cambridge, UK Talk at University of Sheffield, 19 March 2019 In this talk Not in this talk: CRFs, Boltzmann machines,... In this
More informationIdentification of ARX, OE, FIR models with the least squares method
Identification of ARX, OE, FIR models with the least squares method CHEM-E7145 Advanced Process Control Methods Lecture 2 Contents Identification of ARX model with the least squares minimizing the equation
More informationPositive systems in the behavioral approach: main issues and recent results
Positive systems in the behavioral approach: main issues and recent results Maria Elena Valcher Dipartimento di Elettronica ed Informatica Università dipadova via Gradenigo 6A 35131 Padova Italy Abstract
More informationEstimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction
Estimation Tasks Short Course on Image Quality Matthew A. Kupinski Introduction Section 13.3 in B&M Keep in mind the similarities between estimation and classification Image-quality is a statistical concept
More informationEL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1)
EL 625 Lecture 0 EL 625 Lecture 0 Pole Placement and Observer Design Pole Placement Consider the system ẋ Ax () The solution to this system is x(t) e At x(0) (2) If the eigenvalues of A all lie in the
More informationA NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.
A EW IFORMATIO THEORETIC APPROACH TO ORDER ESTIMATIO PROBLEM Soosan Beheshti Munther A. Dahleh Massachusetts Institute of Technology, Cambridge, MA 0239, U.S.A. Abstract: We introduce a new method of model
More informationRegression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood
Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing
More informationGAUSSIAN PROCESS REGRESSION
GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The
More informationManifold Learning: Theory and Applications to HRI
Manifold Learning: Theory and Applications to HRI Seungjin Choi Department of Computer Science Pohang University of Science and Technology, Korea seungjin@postech.ac.kr August 19, 2008 1 / 46 Greek Philosopher
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationLecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012
Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed
More informationMath 2433 Notes Week The Dot Product. The angle between two vectors is found with this formula: cosθ = a b
Math 2433 Notes Week 2 11.3 The Dot Product The angle between two vectors is found with this formula: cosθ = a b a b 3) Given, a = 4i + 4j, b = i - 2j + 3k, c = 2i + 2k Find the angle between a and c Projection
More informationCME 345: MODEL REDUCTION
CME 345: MODEL REDUCTION Proper Orthogonal Decomposition (POD) Charbel Farhat & David Amsallem Stanford University cfarhat@stanford.edu 1 / 43 Outline 1 Time-continuous Formulation 2 Method of Snapshots
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationKernel Principal Component Analysis
Kernel Principal Component Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationA rational interpolation scheme with super-polynomial rate of convergence
Center for Turbulence Research Annual Research Briefs 2008 31 A rational interpolation scheme with super-polynomial rate of convergence By Q. Wang, P. Moin AND G. Iaccarino 1. Motivation and objectives
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More information8. Diagonalization.
8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard
More informationChapter 8: Estimation 1
Chapter 8: Estimation 1 Jae-Kwang Kim Iowa State University Fall, 2014 Kim (ISU) Ch. 8: Estimation 1 Fall, 2014 1 / 33 Introduction 1 Introduction 2 Ratio estimation 3 Regression estimator Kim (ISU) Ch.
More informationSPARSE SIGNAL RESTORATION. 1. Introduction
SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful
More informationLecture 3 January 23
EE 123: Digital Signal Processing Spring 2007 Lecture 3 January 23 Lecturer: Prof. Anant Sahai Scribe: Dominic Antonelli 3.1 Outline These notes cover the following topics: Eigenvectors and Eigenvalues
More informationarxiv: v2 [math.oc] 18 Jan 2015
Regularization for Design Nikolai Matni and Venkat Chandrasekaran January 20, 2015 arxiv:1404.1972v2 [math.oc] 18 Jan 2015 bstr When designing controllers for large-scale systems, the architectural aspects
More informationCSE446: non-parametric methods Spring 2017
CSE446: non-parametric methods Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin and Luke Zettlemoyer Linear Regression: What can go wrong? What do we do if the bias is too strong? Might want
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 18 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 31, 2012 Andre Tkacenko
More informationVector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.
Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar
More informationAdvanced Machine Learning Practical 4b Solution: Regression (BLR, GPR & Gradient Boosting)
Advanced Machine Learning Practical 4b Solution: Regression (BLR, GPR & Gradient Boosting) Professor: Aude Billard Assistants: Nadia Figueroa, Ilaria Lauzana and Brice Platerrier E-mails: aude.billard@epfl.ch,
More informationCS 340 Lec. 15: Linear Regression
CS 340 Lec. 15: Linear Regression AD February 2011 AD () February 2011 1 / 31 Regression Assume you are given some training data { x i, y i } N where x i R d and y i R c. Given an input test data x, you
More informationLecture 6: Linear models and Gauss-Markov theorem
Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables
More informationLecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel
Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random
More informationIdentification of Non-linear Dynamical Systems
Identification of Non-linear Dynamical Systems Linköping University Sweden Prologue Prologue The PI, the Customer and the Data Set C: I have this data set. I have collected it from a cell metabolism experiment.
More informationand u and v are orthogonal if and only if u v = 0. u v = x1x2 + y1y2 + z1z2. 1. In R 3 the dot product is defined by
Linear Algebra [] 4.2 The Dot Product and Projections. In R 3 the dot product is defined by u v = u v cos θ. 2. For u = (x, y, z) and v = (x2, y2, z2), we have u v = xx2 + yy2 + zz2. 3. cos θ = u v u v,
More informationEvent-Triggered Decentralized Dynamic Output Feedback Control for LTI Systems
Event-Triggered Decentralized Dynamic Output Feedback Control for LTI Systems Pavankumar Tallapragada Nikhil Chopra Department of Mechanical Engineering, University of Maryland, College Park, 2742 MD,
More information(a). W contains the zero vector in R n. (b). W is closed under addition. (c). W is closed under scalar multiplication.
. Subspaces of R n Bases and Linear Independence Definition. Subspaces of R n A subset W of R n is called a subspace of R n if it has the following properties: (a). W contains the zero vector in R n. (b).
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationSystem Identification: An Inverse Problem in Control
System Identification: An Inverse Problem in Control Potentials and Possibilities of Regularization Lennart Ljung Linköping University, Sweden Inverse Problems, MAI, Linköping, April 4, 2013 System Identification
More information4. DATA ASSIMILATION FUNDAMENTALS
4. DATA ASSIMILATION FUNDAMENTALS... [the atmosphere] "is a chaotic system in which errors introduced into the system can grow with time... As a consequence, data assimilation is a struggle between chaotic
More informationInput: A set (x i -yy i ) data. Output: Function value at arbitrary point x. What for x = 1.2?
Applied Numerical Analysis Interpolation Lecturer: Emad Fatemizadeh Interpolation Input: A set (x i -yy i ) data. Output: Function value at arbitrary point x. 0 1 4 1-3 3 9 What for x = 1.? Interpolation
More informationREGULARIZED STRUCTURED LOW-RANK APPROXIMATION WITH APPLICATIONS
REGULARIZED STRUCTURED LOW-RANK APPROXIMATION WITH APPLICATIONS MARIYA ISHTEVA, KONSTANTIN USEVICH, AND IVAN MARKOVSKY Abstract. We consider the problem of approximating an affinely structured matrix,
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationSparse Parameter Estimation: Compressed Sensing meets Matrix Pencil
Sparse Parameter Estimation: Compressed Sensing meets Matrix Pencil Yuejie Chi Departments of ECE and BMI The Ohio State University Colorado School of Mines December 9, 24 Page Acknowledgement Joint work
More informationEL1820 Modeling of Dynamical Systems
EL1820 Modeling of Dynamical Systems Lecture 10 - System identification as a model building tool Experiment design Examination and prefiltering of data Model structure selection Model validation Lecture
More information10. Multi-objective least squares
L Vandenberghe ECE133A (Winter 2018) 10 Multi-objective least squares multi-objective least squares regularized data fitting control estimation and inversion 10-1 Multi-objective least squares we have
More informationSupport Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar
Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane
More information