Multivariate time series estimation via projections and matrix equations
|
|
- Corey Jones
- 5 years ago
- Views:
Transcription
1 Multivariate time series estimation via projections and matrix equations Federico Poloni 1 Giacomo Sbrana 2 1 U Pisa, Italy, Dept of Computer Science 2 NEOMA Business School, Rouen, France Structured Numerical Linear and Multilinear Algebra Analysis, Algorithms and Applications September 2014 Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
2 Moving average models Moving average models [Lütkepohl 06 book, Tsay 14 book] {u t R n } Gaussian independent N(0, I) y t = A 0 u t + A 1 u t 1 y t = A 0 u t + A 1 u t 1 + A 2 u t A q u t q MA(1) MA(q) (A i real square matrices, y t, u t vectors) Example: Yearly inflation data looks a lot like a y t Our problem: estimation Given output y t for t = 1, 2,..., T, find the A i s Scalar models popular for macroeconomics; multivariate ones less used because of estimation difficulties Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
3 Transfer function formalism To A 0 u t + A 1 u t 1 + A 2 u t A q u t q we associate G(λ) = A 0 + A 1 λ + A 2 λ A q λ q G(λ) assumed nonsingular (det(a 0 ) 0), and all eigvals outside unit circle Definition G (λ) := A 0 + A 1 λ 1 + A 2 λ A q λ q λ behaves like a complex number on the unit circle G(λ)G (λ) = M q λ q + + M 1 λ 1 + M 0 + M 1 λ + + M q λ q is palindromic (M 0 = M 0 and M i = M i for each i) Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
4 Autocovariance generating function G(λ)G (λ) = M q λ q + + M 1 λ 1 + M 0 + M 1 λ + + M q λ q Theorem [Box, Jenkins, 76 book] [ ] M i = E y t yt i for each i = q,..., q 1 Estimate M i := 1 T 1 Ti=1 y t y t i 2 Factorize M(λ) = G(λ) G (λ) Point 2. is spectral factorization [Wiener, Kailath, Kucera, Ran et al, Ephremidze et al; many more for the scalar case] Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
5 Spectral factorization and matrix equations M 1 λ 1 + M 0 + M 1 λ = (A 1 λ + A 0 )(A 1 λ + A 0 ) Several approaches to solve it: Y = A 0 A 0 solves M 1Y 1 M 1 + Y = M 0 [Ran et al., Meini, CH Guo... ] Z = A 0 A 1 solves M 1 + M 0 Z + M 1 Z 2 [Bini et al, Higham and Kim, Meerbergen and Tisseur... ] G(λ) left spectral divisor of M(λ) [Gohberg Lancaster Rodman]: linearization (palindromic?)+ Schur form Extreme accuracy not necessary here larger error from approximations M i already Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
6 What can go wrong 2 Factorize M(λ) = G(λ) G (λ) Factorization M(λ) = G(λ) G (λ) exists iff M(z) 0 for each z C, z = 1 M i are only approximate, so this property may fail Possible solutions... 1 Apply a correction to M(z) 2 Robust factorization algorithm do something in case of error (something like a robust chol ) 3 Get better M i s Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
7 Possible solutions Apply a correction to M(z) [Brüll, P, Sbrana, Schröder, preprint] regularization approach: perturb polynomial to move eigenvalues Other approaches for similar structures around [Alam et al, Benner and Voigt, Grivet-Talocia, Guglielmi-Overton... ] Robust factorization algorithm Not much usable around! [Kucera, 91] Trying to modify some (Riccati-type eigensolvers, symbolic Cholesky factorization [Janashia, Lagvilava, Ephremidze 11]) Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
8 Get better M i s Other (slower) estimation methods typically get better M i s, such as Maximum Likelihood (ML) Our idea: get the M i s from projection + scalar ML Algorithm 1 Choose weight row vector w, build scalar process z t = {wy t } t=1,...,t 2 Use scalar ML to get g(λ) 3 g(λ)g (λ) = m(λ) = wm(λ)w 4 Repeat with many different w s to reconstruct M(λ) Convergence properties to correct solution when T (technical) [P, Sbrana, submitted] Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
9 The algorithm in a picture Observed data y t ML-estimate Parameters A i of G(λ) compute Autocovariances [ ] M i E y t yt k aggregate compute Scalars z t = w T y t ML-estimate Parameters a i of g(λ) compute Autocovariances m k E [z t z t k ] We take the blue road rather than the red Many (easier) scalar problems rather than a (difficult) multivariate one Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
10 A 2 2 example [ ] w = 1 0 : [ ] ([ ] [ ] [ ] ) [ ] wm(λ)w T a b = 1 0 λ 1 e f a c λ c d f g b d 0 = aλ 1 + e + aλ [ ] w = 0 1 : [ ] ([ ] [ ] [ ] ) [ ] wm(λ)w T a b = 0 1 λ 1 e f a c λ c d f g b d 1 w = [ ] 1 1 : = dλ 1 + g + dλ m(λ) = (a + b + c + d)λ 1 + (e + 2f + g) + (a + b + c + d)λ If b = c, enough to get the remaining entries Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
11 Symmetric case For a symmetric MA(1), w = e i for all i and w = e j + e k for all j k enough to get M(λ) Aλ 1 + B + Aλ, with A = A, B = B can be simultaneously diagonalized (congruence) easier to find spectral factorization This is enough to solve some easier models: Exponentially weighted moving average (random-walk-plus-noise) MA(1) with symmetric matrices Multiple trends MA(2) with symmetric matrices and M 1 = 4M 2 Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
12 Nonsymmetric case Idea Use w(λ), not constant w! [ ] [ ] y 1 λ t,1 = y y t,1 + y t 1,2 t,2 [ ] w = 1 λ [ ] ([ ] [ ] [ ] ) [ ] w(λ)m(λ)w a b = 1 λ λ 1 e f a c λ c d f g b d λ 1 = cλ 2 + (a + d + f )λ 1 + e + 2b + g + (a + d + f )λ + cλ 2 MA(q + 1): more difficult to estimate, but still scalar Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
13 Least squares and more weights Each aggregation vector w(λ) gives us several equations Combine them using least squares! But... joint covariance? W = Same-aggregation values (in red) easy to compute (independence) Black ones are hard set them to zero Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
14 Some results LongAR Quicker but less accurate method META Our method (Moment Estimation Through Aggregation) CML Conditional Maximum Likelihood. Scales badly, basically impossible to beat Z LongAR Z META Z CML Frobenius norm relative errors, 300 trials, 2 2 problem, excluding 48 non-passive samples. eigvals(theta)=[ , ] Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
15 Some more results Symmetric model (EWMA), against out-of-the-box ML on the MA(1) representation (larger parameter space, difficult to enforce symmetry) d = 2, easy d = 2, harder d = 3, easy d = 3, harder T Θ META Θ ML Σ u META Σ u ML Time META Time ML Frobenius norm parameter errors 10 3, average on 500 trials + real-life inflation test by Bankitalia Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
16 Parallelizability for i = 1:nWeights # Estimate g i (λ) with weight w i end # Put together estimates to form M(λ) (fast) # Spectral factorization (fast) while (convergence) #compute likelihood(g(λ)) #compute G i+1 (λ) from G i (λ) using BFGS-like search end function likelihood(g) for t = 1:T # update state[t] from state[t-1], compute likelihood end end Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
17 Why isn t this... Why isn t this model reduction/system identification? Similar to tangential interpolation [Antoulas, Beattie, Gugercin 10 inbook] but... No access to input u t No possibility to get G(λ) exactly: G(λ)Q equally good We work on a squared version: not clear how to relate g(λ) and G(λ) in our setting Why isn t this compressed sensing? Similar to phase retrieval [Candès, Strohmer, Voroninski 11], but it is a matrix version, uncertainty by an orthogonal matrix no sparsity/rank constraint, we want full reconstruction with a full set of measurements Why isn t this Kalman filtering? KF assumes model known, our goal is estimating it here Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
18 Conclusions and issues Interesting idea to explore, new application area Works well in some cases (including symmetric problems) Still missing ideas: fit G(λ) directly without squaring, find optimal weights and equations in LS Tricky to develop with the current software ecosystem Open problem: better regularizing spectral factorization algorithm: We know well how to solve Riccati-like equations, now we want to solve the ones that have no solution! Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
19 Conclusions and issues Interesting idea to explore, new application area Works well in some cases (including symmetric problems) Still missing ideas: fit G(λ) directly without squaring, find optimal weights and equations in LS Tricky to develop with the current software ecosystem Open problem: better regularizing spectral factorization algorithm: We know well how to solve Riccati-like equations, now we want to solve the ones that have no solution! Thanks for your attention Poloni, Sbrana (Pisa, Rouen) Matrix eqns estimators SLA / 18
Perturbing Palindromic Matrix Equations to Make Them Solvable
Perturbing Palindromic Matrix Equations to Make Them Solvable Federico Poloni 1 Joint work with Tobias Brüll 2, Giacomo Sbrana 3, Christian Schröder 4 1 U Pisa, Dept of Computer Science 2 Computer Simulation
More informationEstimating Econometric Models through Matrix Equations
Estimating Econometric Models through Matrix Equations Federico Poloni 1 Giacomo Sbrana 2 1 U Pisa, Dept of Computer Science 2 Rouen Business School, France No Free Lunch Seminar SNS, Pisa, February 2013
More informationModel estimation through matrix equations in financial econometrics
Model estimation through matrix equations in financial econometrics Federico Poloni 1 Joint work with Giacomo Sbrana 2 1 Technische Universität Berlin (A. Von Humboldt postdoctoral fellow) 2 Rouen Business
More informationChapter 3 - Temporal processes
STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect
More informationCyclic reduction and index reduction/shifting for a second-order probabilistic problem
Cyclic reduction and index reduction/shifting for a second-order probabilistic problem Federico Poloni 1 Joint work with Giang T. Nguyen 2 1 U Pisa, Computer Science Dept. 2 U Adelaide, School of Math.
More informationDiagonalizing Matrices
Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B
More informationGaussian Filtering Strategies for Nonlinear Systems
Gaussian Filtering Strategies for Nonlinear Systems Canonical Nonlinear Filtering Problem ~u m+1 = ~ f (~u m )+~ m+1 ~v m+1 = ~g(~u m+1 )+~ o m+1 I ~ f and ~g are nonlinear & deterministic I Noise/Errors
More informationEL2520 Control Theory and Practice
EL2520 Control Theory and Practice Lecture 8: Linear quadratic control Mikael Johansson School of Electrical Engineering KTH, Stockholm, Sweden Linear quadratic control Allows to compute the controller
More informationRecent Advances in the Numerical Solution of Quadratic Eigenvalue Problems
Recent Advances in the Numerical Solution of Quadratic Eigenvalue Problems Françoise Tisseur School of Mathematics The University of Manchester ftisseur@ma.man.ac.uk http://www.ma.man.ac.uk/~ftisseur/
More informationLecture 10 - Eigenvalues problem
Lecture 10 - Eigenvalues problem Department of Computer Science University of Houston February 28, 2008 1 Lecture 10 - Eigenvalues problem Introduction Eigenvalue problems form an important class of problems
More informationHere represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities.
19 KALMAN FILTER 19.1 Introduction In the previous section, we derived the linear quadratic regulator as an optimal solution for the fullstate feedback control problem. The inherent assumption was that
More informationNotes on Linear Algebra and Matrix Theory
Massimo Franceschet featuring Enrico Bozzo Scalar product The scalar product (a.k.a. dot product or inner product) of two real vectors x = (x 1,..., x n ) and y = (y 1,..., y n ) is not a vector but a
More informationRecursive Solutions of the Matrix Equations X + A T X 1 A = Q and X A T X 1 A = Q
Applied Mathematical Sciences, Vol. 2, 2008, no. 38, 1855-1872 Recursive Solutions of the Matrix Equations X + A T X 1 A = Q and X A T X 1 A = Q Nicholas Assimakis Department of Electronics, Technological
More informationDerivation of the Kalman Filter
Derivation of the Kalman Filter Kai Borre Danish GPS Center, Denmark Block Matrix Identities The key formulas give the inverse of a 2 by 2 block matrix, assuming T is invertible: T U 1 L M. (1) V W N P
More informationCh. 12 Linear Bayesian Estimators
Ch. 1 Linear Bayesian Estimators 1 In chapter 11 we saw: the MMSE estimator takes a simple form when and are jointly Gaussian it is linear and used only the 1 st and nd order moments (means and covariances).
More informationLinear Least-Squares Data Fitting
CHAPTER 6 Linear Least-Squares Data Fitting 61 Introduction Recall that in chapter 3 we were discussing linear systems of equations, written in shorthand in the form Ax = b In chapter 3, we just considered
More informationReview of Basic Concepts in Linear Algebra
Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra
More informationNew Introduction to Multiple Time Series Analysis
Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2
More informationComputing Unstructured and Structured Polynomial Pseudospectrum Approximations
Computing Unstructured and Structured Polynomial Pseudospectrum Approximations Silvia Noschese 1 and Lothar Reichel 2 1 Dipartimento di Matematica, SAPIENZA Università di Roma, P.le Aldo Moro 5, 185 Roma,
More informationStatistical Signal Processing Detection, Estimation, and Time Series Analysis
Statistical Signal Processing Detection, Estimation, and Time Series Analysis Louis L. Scharf University of Colorado at Boulder with Cedric Demeure collaborating on Chapters 10 and 11 A TT ADDISON-WESLEY
More informationThe Eigenvalue Problem: Perturbation Theory
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just
More informationReview of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationDescriptor system techniques in solving H 2 -optimal fault detection problems
Descriptor system techniques in solving H 2 -optimal fault detection problems Andras Varga German Aerospace Center (DLR) DAE 10 Workshop Banff, Canada, October 25-29, 2010 Outline approximate fault detection
More informationFinite dimensional indefinite inner product spaces and applications in Numerical Analysis
Finite dimensional indefinite inner product spaces and applications in Numerical Analysis Christian Mehl Technische Universität Berlin, Institut für Mathematik, MA 4-5, 10623 Berlin, Germany, Email: mehl@math.tu-berlin.de
More informationarxiv: v1 [stat.co] 15 Nov 2014
A note on forecasting demand using the multivariate exponential smoothing framework Federico Poloni and Giacomo Sbrana arxiv:4.434v stat.co 5 Nov 04 Abstract Simple exponential smoothing is widely used
More informationChapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors
Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there
More informationA Framework for Structured Linearizations of Matrix Polynomials in Various Bases
A Framework for Structured Linearizations of Matrix Polynomials in Various Bases Leonardo Robol Joint work with Raf Vandebril and Paul Van Dooren, KU Leuven and Université
More informationEigenvalues and Eigenvectors
5 Eigenvalues and Eigenvectors 5.2 THE CHARACTERISTIC EQUATION DETERMINANATS nn Let A be an matrix, let U be any echelon form obtained from A by row replacements and row interchanges (without scaling),
More informationIntroduction. Chapter One
Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and
More informationLecture 2: Computing functions of dense matrices
Lecture 2: Computing functions of dense matrices Paola Boito and Federico Poloni Università di Pisa Pisa - Hokkaido - Roma2 Summer School Pisa, August 27 - September 8, 2018 Introduction In this lecture
More informationConvergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit
Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Evan Kwiatkowski, Jan Mandel University of Colorado Denver December 11, 2014 OUTLINE 2 Data Assimilation Bayesian Estimation
More informationLinear Algebra (MATH ) Spring 2011 Final Exam Practice Problem Solutions
Linear Algebra (MATH 4) Spring 2 Final Exam Practice Problem Solutions Instructions: Try the following on your own, then use the book and notes where you need help. Afterwards, check your solutions with
More informationBindel, Fall 2011 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Jan 9
Problem du jour Week 3: Wednesday, Jan 9 1. As a function of matrix dimension, what is the asymptotic complexity of computing a determinant using the Laplace expansion (cofactor expansion) that you probably
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7
More informationVector autoregressive Moving Average Process. Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem
Vector autoregressive Moving Average Process Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem Road Map 1. Introduction 2. Properties of MA Finite Process 3. Stationarity of MA Process 4. VARMA
More informationFast Angular Synchronization for Phase Retrieval via Incomplete Information
Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department
More informationRemark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called
More informationLeast Squares. Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Winter UCSD
Least Squares Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 75A Winter 0 - UCSD (Unweighted) Least Squares Assume linearity in the unnown, deterministic model parameters Scalar, additive noise model: y f (
More informationExamples True or false: 3. Let A be a 3 3 matrix. Then there is a pattern in A with precisely 4 inversions.
The exam will cover Sections 6.-6.2 and 7.-7.4: True/False 30% Definitions 0% Computational 60% Skip Minors and Laplace Expansion in Section 6.2 and p. 304 (trajectories and phase portraits) in Section
More informationBasic Concepts in Linear Algebra
Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear
More informationApplied Linear Algebra
Applied Linear Algebra Peter J. Olver School of Mathematics University of Minnesota Minneapolis, MN 55455 olver@math.umn.edu http://www.math.umn.edu/ olver Chehrzad Shakiban Department of Mathematics University
More informationMath 4A Notes. Written by Victoria Kala Last updated June 11, 2017
Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...
More informationApplied Time. Series Analysis. Wayne A. Woodward. Henry L. Gray. Alan C. Elliott. Dallas, Texas, USA
Applied Time Series Analysis Wayne A. Woodward Southern Methodist University Dallas, Texas, USA Henry L. Gray Southern Methodist University Dallas, Texas, USA Alan C. Elliott University of Texas Southwestern
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationEnforcing solvability of a nonlinear matrix equation and estimation of multivariate ARMA time series
Enforcing solvability of a nonlinear matrix equation and estimation of multivariate ARMA time series Tobias Brüll and Federico Poloni and Giacomo Sbrana and Christian Schröder July 18, 2013 Abstract The
More information1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?
. Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in
More information9 Multi-Model State Estimation
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State
More information(Mathematical Operations with Arrays) Applied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB (Mathematical Operations with Arrays) Contents Getting Started Matrices Creating Arrays Linear equations Mathematical Operations with Arrays Using Script
More informationDot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.
Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................
More informationEigenvector error bound and perturbation for nonlinear eigenvalue problems
Eigenvector error bound and perturbation for nonlinear eigenvalue problems Yuji Nakatsukasa School of Mathematics University of Tokyo Joint work with Françoise Tisseur Workshop on Nonlinear Eigenvalue
More informationPrinciples of the Global Positioning System Lecture 11
12.540 Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring http://geoweb.mit.edu/~tah/12.540 Statistical approach to estimation Summary Look at estimation from statistical point
More informationThe quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying
I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given
More informationAlgorithms for Solving the Polynomial Eigenvalue Problem
Algorithms for Solving the Polynomial Eigenvalue Problem Nick Higham School of Mathematics The University of Manchester higham@ma.man.ac.uk http://www.ma.man.ac.uk/~higham/ Joint work with D. Steven Mackey
More informationVectors for Beginners
Vectors for Beginners Leo Dorst September 6, 2007 1 Three ways of looking at linear algebra We will always try to look at what we do in linear algebra at three levels: geometric: drawing a picture. This
More informationRemark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.
Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial
More informationThe Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix
The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix Chun-Yueh Chiang Center for General Education, National Formosa University, Huwei 632, Taiwan. Matthew M. Lin 2, Department of
More informationSparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices
Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Jaehyun Park June 1 2016 Abstract We consider the problem of writing an arbitrary symmetric matrix as
More informationTheorem A.1. If A is any nonzero m x n matrix, then A is equivalent to a partitioned matrix of the form. k k n-k. m-k k m-k n-k
I. REVIEW OF LINEAR ALGEBRA A. Equivalence Definition A1. If A and B are two m x n matrices, then A is equivalent to B if we can obtain B from A by a finite sequence of elementary row or elementary column
More informationThis appendix provides a very basic introduction to linear algebra concepts.
APPENDIX Basic Linear Algebra Concepts This appendix provides a very basic introduction to linear algebra concepts. Some of these concepts are intentionally presented here in a somewhat simplified (not
More informationMath 18.6, Spring 213 Problem Set #6 April 5, 213 Problem 1 ( 5.2, 4). Identify all the nonzero terms in the big formula for the determinants of the following matrices: 1 1 1 2 A = 1 1 1 1 1 1, B = 3 4
More informationA time series is called strictly stationary if the joint distribution of every collection (Y t
5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a
More informationAdvanced Digital Signal Processing -Introduction
Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary
More informationExercise Set 7.2. Skills
Orthogonally diagonalizable matrix Spectral decomposition (or eigenvalue decomposition) Schur decomposition Subdiagonal Upper Hessenburg form Upper Hessenburg decomposition Skills Be able to recognize
More informationPerturbation theory for eigenvalues of Hermitian pencils. Christian Mehl Institut für Mathematik TU Berlin, Germany. 9th Elgersburg Workshop
Perturbation theory for eigenvalues of Hermitian pencils Christian Mehl Institut für Mathematik TU Berlin, Germany 9th Elgersburg Workshop Elgersburg, March 3, 2014 joint work with Shreemayee Bora, Michael
More informationOPTIMAL CONTROL AND ESTIMATION
OPTIMAL CONTROL AND ESTIMATION Robert F. Stengel Department of Mechanical and Aerospace Engineering Princeton University, Princeton, New Jersey DOVER PUBLICATIONS, INC. New York CONTENTS 1. INTRODUCTION
More informationSynthesis of Gaussian and non-gaussian stationary time series using circulant matrix embedding
Synthesis of Gaussian and non-gaussian stationary time series using circulant matrix embedding Vladas Pipiras University of North Carolina at Chapel Hill UNC Graduate Seminar, November 10, 2010 (joint
More informationEstimation, Detection, and Identification CMU 18752
Estimation, Detection, and Identification CMU 18752 Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt Phone: +351 21 8418053
More informationDifferentiation of functions of covariance
Differentiation of log X May 5, 2005 1 Differentiation of functions of covariance matrices or: Why you can forget you ever read this Richard Turner Covariance matrices are symmetric, but we often conveniently
More informationLecture 16: State Space Model and Kalman Filter Bus 41910, Time Series Analysis, Mr. R. Tsay
Lecture 6: State Space Model and Kalman Filter Bus 490, Time Series Analysis, Mr R Tsay A state space model consists of two equations: S t+ F S t + Ge t+, () Z t HS t + ɛ t (2) where S t is a state vector
More information11. Further Issues in Using OLS with TS Data
11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,
More informationFactor Analysis (10/2/13)
STA561: Probabilistic machine learning Factor Analysis (10/2/13) Lecturer: Barbara Engelhardt Scribes: Li Zhu, Fan Li, Ni Guan Factor Analysis Factor analysis is related to the mixture models we have studied.
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationMatrix decompositions
Matrix decompositions Zdeněk Dvořák May 19, 2015 Lemma 1 (Schur decomposition). If A is a symmetric real matrix, then there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ T. The
More information15 Singular Value Decomposition
15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationLinear Algebra for Machine Learning. Sargur N. Srihari
Linear Algebra for Machine Learning Sargur N. srihari@cedar.buffalo.edu 1 Overview Linear Algebra is based on continuous math rather than discrete math Computer scientists have little experience with it
More informationLecture 2: Univariate Time Series
Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:
More informationChapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes
Chapter 2 Some basic tools 2.1 Time series: Theory 2.1.1 Stochastic processes A stochastic process is a sequence of random variables..., x 0, x 1, x 2,.... In this class, the subscript always means time.
More informationFactoring block Fiedler Companion Matrices
Chapter Factoring block Fiedler Companion Matrices Gianna M. Del Corso, Federico Poloni, Leonardo Robol, and Raf Vandebril Abstract When Fiedler published his A note on Companion matrices in 00 in Linear
More informationReview of some mathematical tools
MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical
More informationSubspace Identification
Chapter 10 Subspace Identification Given observations of m 1 input signals, and p 1 signals resulting from those when fed into a dynamical system under study, can we estimate the internal dynamics regulating
More informationExercise Sheet 1. 1 Probability revision 1: Student-t as an infinite mixture of Gaussians
Exercise Sheet 1 1 Probability revision 1: Student-t as an infinite mixture of Gaussians Show that an infinite mixture of Gaussian distributions, with Gamma distributions as mixing weights in the following
More informationStructured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries
Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries Fakultät für Mathematik TU Chemnitz, Germany Peter Benner benner@mathematik.tu-chemnitz.de joint work with Heike Faßbender
More informationCONSTRUCTION OF A COMPLEX JACOBI MATRIX FROM TWO-SPECTRA
Hacettepe Journal of Mathematics and Statistics Volume 40(2) (2011), 297 303 CONSTRUCTION OF A COMPLEX JACOBI MATRIX FROM TWO-SPECTRA Gusein Sh Guseinov Received 21:06 :2010 : Accepted 25 :04 :2011 Abstract
More informationDigital Image Processing
Digital Image Processing 2D SYSTEMS & PRELIMINARIES Hamid R. Rabiee Fall 2015 Outline 2 Two Dimensional Fourier & Z-transform Toeplitz & Circulant Matrices Orthogonal & Unitary Matrices Block Matrices
More informationCity Suburbs. : population distribution after m years
Section 5.3 Diagonalization of Matrices Definition Example: stochastic matrix To City Suburbs From City Suburbs.85.03 = A.15.97 City.15.85 Suburbs.97.03 probability matrix of a sample person s residence
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference
More informationQuiz ) Locate your 1 st order neighbors. 1) Simplify. Name Hometown. Name Hometown. Name Hometown.
Quiz 1) Simplify 9999 999 9999 998 9999 998 2) Locate your 1 st order neighbors Name Hometown Me Name Hometown Name Hometown Name Hometown Solving Linear Algebraic Equa3ons Basic Concepts Here only real
More informationQuestion: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?
Section 5. The Characteristic Polynomial Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI? Property The eigenvalues
More informationQuadratic Matrix Polynomials
Research Triangularization Matters of Quadratic Matrix Polynomials February 25, 2009 Nick Françoise Higham Tisseur Director School of of Research Mathematics The University of Manchester School of Mathematics
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationLinear Models 1. Isfahan University of Technology Fall Semester, 2014
Linear Models 1 Isfahan University of Technology Fall Semester, 2014 References: [1] G. A. F., Seber and A. J. Lee (2003). Linear Regression Analysis (2nd ed.). Hoboken, NJ: Wiley. [2] A. C. Rencher and
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationOn a root-finding approach to the polynomial eigenvalue problem
On a root-finding approach to the polynomial eigenvalue problem Dipartimento di Matematica, Università di Pisa www.dm.unipi.it/ bini Joint work with Vanni Noferini and Meisam Sharify Limoges, May, 10,
More informationState Observers and the Kalman filter
Modelling and Control of Dynamic Systems State Observers and the Kalman filter Prof. Oreste S. Bursi University of Trento Page 1 Feedback System State variable feedback system: Control feedback law:u =
More informationThe Hilbert Space of Random Variables
The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More information