Summary of factor model analysis by Fan at al.
|
|
- Florence Beasley
- 6 years ago
- Views:
Transcription
1 Summary of factor model analysis by Fan at al. W. Evan Durno July 11, 2015 W. Evan Durno Factoring Porfolios July 11, / 15
2 Brief summary Factor models for high-dimensional data are analysed [Fan et al., 2008]. Majority of work emphasizes O(n) O(p). 1 Convergence is studied from the perspective of matrix norms. 2 Convergence speed is theorized in the O( )-style, emphasizing how increasing samples, dimensions, and factors influence error. 3 Simulations demonstrate adequacy of a proposed norm. 4 Analysis is described as theoretical W. Evan Durno Factoring Porfolios July 11, / 15
3 Factor model review Our n data, Y, are in p dimensions with k factors. 1 Ψ R p p 0, diagonal, not random; L = [l i,j] R p k, not random 2 F = [f j ] N(0, I k ); E = [e i ] N(0, Ψ) 3 E does not depend on F 4 Y = µ + LF + E Y N(µ, Σ) s.t. Cov[Y ] = Σ = LL + Ψ Estimation for L can be done in a few ways, of which MLE and Eigenvectors are popular. The authors use least squares given Y & F. 1 ˆΣ = ˆB ˆ Cov[F ]ˆB + ˆΨ, ˆΨ = diag[e v E v /n], ˆB = YX (XX ) 1 2 such that E v = [E 1,..., E n ], X = [F 1,..., F n ] 3 Criticism: I don t get to observe F! So it is a theoretical exercise. W. Evan Durno Factoring Porfolios July 11, / 15
4 Matrix norms This paper primarily uses the following norms. 1 Frobenious [Horn & Johnson, 1990]: A = [tr(aa )] 1/2 = [ p j=1 λ(a)2 ] 1/2 2 Entropy loss: [James &) Stein, 1961] L 1 (ˆΣ, Σ) = tr (ˆΣΣ 1 log ˆΣΣ 1 p 3 Custom A Σ = p 1 Σ 1/2 AΣ 1/2 Their custom norm is a scaled quadratic norm L 2. Precisely, A Σ Σ = p 1 L 2 (ˆΣ, Σ) = p 1 tr [ AΣ 1 I ]. Convention: Eigenvalues of A are λ 1 (A),..., λ p (A) decreasing. W. Evan Durno Factoring Porfolios July 11, / 15
5 Assumptions The following assumptions are employed partially or entirely. Let b = E Y 2, c = max 1 i k E[f 4 j ], d = max 1 i k E[e 4 i ] 1 (A) (Y i, F i ) iid, E[E i F i ] = 0, Cov[E F ] = Ψ, distribution of F is continuous, k p. 2 (B) b = O(p), c & d bounded. σ 1 > 0 : λ k (LL ) σ 1 n 3 (C) σ 2 > 0 : λ p (Ψ) > σ 2 n 4 (D) The k factors F = [f j ] are fixed accross n. W. Evan Durno Factoring Porfolios July 11, / 15
6 Frobenious is a poor descriptor of convergence Theorem (1) Under assumptions (A) - (C), 1 ˆΣ Σ = O P (pk/ n) 2 ˆΣ s Σ = O P (pk/ n) λi 3 max 1 i p (ˆΣ) λ i (Σ) = O P (pk/ n) λi 4 max 1 i p (ˆΣ s ) λ i (Σ) = O P (pk/ n) Notation 1 Let ˆΣ denote the factor model estimate of Σ. 2 Let ˆΣ s denote the sample covariance estimate of Σ. W. Evan Durno Factoring Porfolios July 11, / 15
7 Quadratic norm better describes convergence Theorem (2) Under assumptions (A) - (C), if K = O(n α 1 ), p = O(n α ), then 1 ˆΣ Σ Σ = O P (n β/2 ) : β = min{1 2α 1, 2 α α 1 } 2 ˆΣ s Σ Σ = O P (n β1/2 ) : β 1 = 1 max{α, 3α 1 /2, 3α 1 α} Remark: α > 2α 1 & α 1 < 1 β > β 1 So the factor estimate converges faster when the number of dimensions grows faster than the number of factors. Also note that α 1 ˆΣ is root-n-consistent under Σ. W. Evan Durno Factoring Porfolios July 11, / 15
8 More theory Theorem (3) Under assumptions (A) - (C), 1 ˆΣ 1 Σ 1 = o P ([p 2 k 4 log n/n] 1/2 ) 2 ˆΣ 1 s Σ 1 = o P ([p 4 k 2 log n/n] 1/2 ) Remark: ˆΣ 1 & ˆΣ 1 s do refer matrix inverses of the estimates. Theorem (4) Under assumptions (A), (B), and (D), ˆΣ Σ converges in distribution to normal errors. Language is purposefully vague for brevity. W. Evan Durno Factoring Porfolios July 11, / 15
9 Markowitz portfolio optimization Define the following variables. 1 There are p risky assets with rate of return µ and covariance Σ. 2 ξ is the variance-minimum allocation vector with rate of return γ. 3 ξ g is the minimum variance porfolio (no desired rate of return). Theorem (Markowitz [Markowitz, H.M., 1952]) The solution to the problem: min ξ Σξ : ξ 1 = 1 & ξµ = γ ξ is this: ξ = φ γψ ρφ ψ 2 Σ γρ ψ ρφ ψ 2 Σ 1 µ such that ρ = 1 Σ 1 1, ψ = 1 Σ 1 µ, φ = µ Σ 1 µ with variance ξ Σξ = ργ2 2φγ+φ ρφ ψ 2. To get ˆξ g, replace γ with ψ/φ. W. Evan Durno Factoring Porfolios July 11, / 15
10 Factor estimates have less porfolio variance If we simply plug in estimates ˆΣ, ˆΣ n, ˆµ we get allocation estimates ˆξ, ˆξ g. Theorem (5, global minimum variance convergence) Under (A) - (C) and ρ > 0, 1 ˆξ g ˆΣˆξ g ξ g Σξ g = o P ([p 4 k 4 log n/n] 1/2 ) 2 ˆξ g ˆΣ s ˆξ g ξ g Σξ g = o P ([p 6 k 2 log n/n] 1/2 ) Theorem (5, optimal variance convergence) Under (A) - (C) and ρφ ψ 2 > max{0, ρ/(ρφ ψ 2 ), ψ/(ρφ ψ 2 ), φ/(ρφ ψ 2 )}, 1 ˆξ ˆΣˆξ ξ Σξ = o P ([p 4 k 4 log n/n] 1/2 ) 2 ˆξ ˆΣ s ˆξ ξ Σξ = o P ([p 6 k 2 log n/n] 1/2 ) W. Evan Durno Factoring Porfolios July 11, / 15
11 Weak convergence of variance Require ξ = O(1)1 to avoid extreme short positions. Theorem Under (A) and (B), 1 ξ ˆΣξ ξ Σξ = o P ([p 4 k 2 log n/n] 1/2 ) 2 ξ ˆΣ s ξ ξ Σξ = o P ([p 4 k 2 log n/n] 1/2 ) and requiring NO short positions results in 1 ξ ˆΣξ ξ Σξ = o P ([p 2 k 2 log n/n] 1/2 ) 2 ξ ˆΣ s ξ ξ Σξ = o P ([p 2 k 2 log n/n] 1/2 ) Notice the equivalence. W. Evan Durno Factoring Porfolios July 11, / 15
12 A simulation study Under fixed n and k, p was allowed to increase. The study revealed the following. 1 The Frobenious norm was insufficient to demonstrate covariance estimate inefficiency. 2 The scaled quadratic norm was sufficient to demonstrate covariance estimate efficiency. 3 When data follow a factor model, observing the factors and including them in the model greatly increases covariance modelling efficiency 4 Estimated portfolio variance MSEs greatly reduce with the factor model when data follow a factor model, even when p > n. W. Evan Durno Factoring Porfolios July 11, / 15
13 Recap 1 Estimation is easy when the model is known and latent variables are observable. 2 Results suggest that correct usage of a factor model may result in reduce variance in covariance estimation and porfolio variance. 3 Results suggest that a simple porfolio (ξ = O(1)1) without shorts does not enjoy a better variance reduction, so standard covariance estimation is adequate. 4 The Frobenious norm may is not always adequate for understanding covariance estimate inefficiency. W. Evan Durno Factoring Porfolios July 11, / 15
14 References Jianquing Fan, Yingying Fan, Jinchi Lv (2008) High dimensional covariance matrix estimation using a factor model Journal of Econometrics 147, Horn, R.A. & Johnson, C.R. (1990) Matrix Analysis Cambridge Press James, W. & Stein, C. (1961) Estimation iwth quadratic loss Proc. Fourth Berkeley Symp. Math. Stat. Prob. vol 1, California Press, Berkeley, pp Markowitz (1952) Portfolio selection Journal of Finance 7, W. Evan Durno Factoring Porfolios July 11, / 15
15 The End W. Evan Durno Factoring Porfolios July 11, / 15
arxiv:math/ v1 [math.st] 4 Jan 2007
High Dimensional Covariance Matrix Estimation arxiv:math/0701124v1 [math.st] 4 Jan 2007 Using a Factor Model By Jianqing Fan, Yingying Fan and Jinchi Lv Princeton University August 12, 2006 High dimensionality
More informationMLES & Multivariate Normal Theory
Merlise Clyde September 6, 2016 Outline Expectations of Quadratic Forms Distribution Linear Transformations Distribution of estimates under normality Properties of MLE s Recap Ŷ = ˆµ is an unbiased estimate
More informationA Method to Estimate the True Mahalanobis Distance from Eigenvectors of Sample Covariance Matrix
A Method to Estimate the True Mahalanobis Distance from Eigenvectors of Sample Covariance Matrix Masakazu Iwamura, Shinichiro Omachi, and Hirotomo Aso Graduate School of Engineering, Tohoku University
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More informationDelta Theorem in the Age of High Dimensions
Delta Theorem in the Age of High Dimensions Mehmet Caner Department of Economics Ohio State University December 15, 2016 Abstract We provide a new version of delta theorem, that takes into account of high
More informationPortfolio Allocation using High Frequency Data. Jianqing Fan
Portfolio Allocation using High Frequency Data Princeton University With Yingying Li and Ke Yu http://www.princeton.edu/ jqfan September 10, 2010 About this talk How to select sparsely optimal portfolio?
More informationFactor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA
Factor Analysis Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Factor Models The multivariate regression model Y = XB +U expresses each row Y i R p as a linear combination
More informationHomework 1. Yuan Yao. September 18, 2011
Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to
More informationarxiv: v2 [math.st] 2 Jul 2017
A Relaxed Approach to Estimating Large Portfolios Mehmet Caner Esra Ulasan Laurent Callot A.Özlem Önder July 4, 2017 arxiv:1611.07347v2 [math.st] 2 Jul 2017 Abstract This paper considers three aspects
More informationMarkowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations
Markowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations Sahand Rabbani Introduction Modern portfolio theory deals in part with the efficient allocation of investments
More informationTAMS39 Lecture 10 Principal Component Analysis Factor Analysis
TAMS39 Lecture 10 Principal Component Analysis Factor Analysis Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content - Lecture Principal component analysis
More information3.1. The probabilistic view of the principal component analysis.
301 Chapter 3 Principal Components and Statistical Factor Models This chapter of introduces the principal component analysis (PCA), briefly reviews statistical factor models PCA is among the most popular
More informationIntroduction to Computational Finance and Financial Econometrics Matrix Algebra Review
You can t see this text! Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Matrix Algebra Review 1 / 54 Outline 1
More information9.1 Orthogonal factor model.
36 Chapter 9 Factor Analysis Factor analysis may be viewed as a refinement of the principal component analysis The objective is, like the PC analysis, to describe the relevant variables in study in terms
More informationData Mining Stat 588
Data Mining Stat 588 Lecture 02: Linear Methods for Regression Department of Statistics & Biostatistics Rutgers University September 13 2011 Regression Problem Quantitative generic output variable Y. Generic
More informationPrior Information: Shrinkage and Black-Litterman. Prof. Daniel P. Palomar
Prior Information: Shrinkage and Black-Litterman Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter
More informationEstimation of large dimensional sparse covariance matrices
Estimation of large dimensional sparse covariance matrices Department of Statistics UC, Berkeley May 5, 2009 Sample covariance matrix and its eigenvalues Data: n p matrix X n (independent identically distributed)
More informationTheory and Applications of High Dimensional Covariance Matrix Estimation
1 / 44 Theory and Applications of High Dimensional Covariance Matrix Estimation Yuan Liao Princeton University Joint work with Jianqing Fan and Martina Mincheva December 14, 2011 2 / 44 Outline 1 Applications
More informationThreading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection. A Physics & Pizza on Wednesdays presentation
Threading Rotational Dynamics, Crystal Optics, and Quantum Mechanics to Risky Asset Selection A Physics & Pizza on Wednesdays presentation M. Hossein Partovi Department of Physics & Astronomy, California
More informationTHE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2014, Mr. Ruey S. Tsay. Solutions to Final Exam
THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2014, Mr. Ruey S. Tsay Solutions to Final Exam 1. City crime: The distance matrix is 694 915 1073 528 716 881 972 464
More informationAlternative Bayesian Estimators for Vector-Autoregressive Models
Alternative Bayesian Estimators for Vector-Autoregressive Models Shawn Ni, Department of Economics, University of Missouri, Columbia, MO 65211, USA Dongchu Sun, Department of Statistics, University of
More informationIntroduction Eigen Values and Eigen Vectors An Application Matrix Calculus Optimal Portfolio. Portfolios. Christopher Ting.
Portfolios Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 November 4, 2016 Christopher Ting QF 101 Week 12 November 4,
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationDependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.
MFM Practitioner Module: Risk & Asset Allocation September 11, 2013 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y
More informationStatistics 910, #5 1. Regression Methods
Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known
More informationCross Sectional Time Series: The Normal Model and Panel Corrected Standard Errors
Cross Sectional Time Series: The Normal Model and Panel Corrected Standard Errors Paul Johnson 5th April 2004 The Beck & Katz (APSR 1995) is extremely widely cited and in case you deal
More informatione-companion ONLY AVAILABLE IN ELECTRONIC FORM Electronic Companion Stochastic Kriging for Simulation Metamodeling
OPERATIONS RESEARCH doi 10.187/opre.1090.0754ec e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 009 INFORMS Electronic Companion Stochastic Kriging for Simulation Metamodeling by Bruce Ankenman,
More informationMSA220 Statistical Learning for Big Data
MSA220 Statistical Learning for Big Data Lecture 4 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology More on Discriminant analysis More on Discriminant
More informationEcon 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More informationEigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage?
Eigenvalue spectra of time-lagged covariance matrices: Possibilities for arbitrage? Stefan Thurner www.complex-systems.meduniwien.ac.at www.santafe.edu London July 28 Foundations of theory of financial
More informationEstimating Covariance Using Factorial Hidden Markov Models
Estimating Covariance Using Factorial Hidden Markov Models João Sedoc 1,2 with: Jordan Rodu 3, Lyle Ungar 1, Dean Foster 1 and Jean Gallier 1 1 University of Pennsylvania Philadelphia, PA joao@cis.upenn.edu
More informationJournal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error
Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA
More informationFixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility
American Economic Review: Papers & Proceedings 2016, 106(5): 400 404 http://dx.doi.org/10.1257/aer.p20161082 Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility By Gary Chamberlain*
More informationON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction
Acta Math. Univ. Comenianae Vol. LXV, 1(1996), pp. 129 139 129 ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES V. WITKOVSKÝ Abstract. Estimation of the autoregressive
More informationEstimation of Unique Variances Using G-inverse Matrix in Factor Analysis
International Mathematical Forum, 3, 2008, no. 14, 671-676 Estimation of Unique Variances Using G-inverse Matrix in Factor Analysis Seval Süzülmüş Osmaniye Korkut Ata University Vocational High School
More informationPrincipal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,
Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations
More informationCanonical Correlation Analysis of Longitudinal Data
Biometrics Section JSM 2008 Canonical Correlation Analysis of Longitudinal Data Jayesh Srivastava Dayanand N Naik Abstract Studying the relationship between two sets of variables is an important multivariate
More informationICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization
ICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization Xiaohui Xie University of California, Irvine xhx@uci.edu Xiaohui Xie (UCI) ICS 6N 1 / 21 Symmetric matrices An n n
More informationCh4. Distribution of Quadratic Forms in y
ST4233, Linear Models, Semester 1 2008-2009 Ch4. Distribution of Quadratic Forms in y 1 Definition Definition 1.1 If A is a symmetric matrix and y is a vector, the product y Ay = i a ii y 2 i + i j a ij
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationStatistics & Decisions 7, (1989) R. Oldenbourg Verlag, München /89 $
Statistics & Decisions 7, 377-382 (1989) R. Oldenbourg Verlag, München 1989-0721-2631/89 $3.00 + 0.00 A NOTE ON SIMULTANEOUS ESTIMATION OF PEARSON MODES L. R. Haff 1 ) and R. W. Johnson 2^ Received: Revised
More informationLecture 20 May 18, Empirical Bayes Interpretation [Efron & Morris 1973]
Stats 300C: Theory of Statistics Spring 2018 Lecture 20 May 18, 2018 Prof. Emmanuel Candes Scribe: Will Fithian and E. Candes 1 Outline 1. Stein s Phenomenon 2. Empirical Bayes Interpretation of James-Stein
More informationFactor Models for Asset Returns. Prof. Daniel P. Palomar
Factor Models for Asset Returns Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19, HKUST,
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationLinear Algebra in Actuarial Science: Slides to the lecture
Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations
More informationRobust Optimization: Applications in Portfolio Selection Problems
Robust Optimization: Applications in Portfolio Selection Problems Vris Cheung and Henry Wolkowicz WatRISQ University of Waterloo Vris Cheung (University of Waterloo) Robust optimization 2009 1 / 19 Outline
More informationDependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.
Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,
More informationEmpirical properties of large covariance matrices in finance
Empirical properties of large covariance matrices in finance Ex: RiskMetrics Group, Geneva Since 2010: Swissquote, Gland December 2009 Covariance and large random matrices Many problems in finance require
More informationPrincipal component analysis
Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and
More informationMS-E2112 Multivariate Statistical Analysis (5cr) Lecture 6: Bivariate Correspondence Analysis - part II
MS-E2112 Multivariate Statistical Analysis (5cr) Lecture 6: Bivariate Correspondence Analysis - part II the Contents the the the Independence The independence between variables x and y can be tested using.
More informationWLS and BLUE (prelude to BLUP) Prediction
WLS and BLUE (prelude to BLUP) Prediction Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 21, 2018 Suppose that Y has mean X β and known covariance matrix V (but Y need
More informationInferences about a Mean Vector
Inferences about a Mean Vector Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University
More informationLecture 9 SLR in Matrix Form
Lecture 9 SLR in Matrix Form STAT 51 Spring 011 Background Reading KNNL: Chapter 5 9-1 Topic Overview Matrix Equations for SLR Don t focus so much on the matrix arithmetic as on the form of the equations.
More informationStochastic Design Criteria in Linear Models
AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework
More informationMauricio A. Elzo, University of Florida, 1996, 2005, 2006, 2010, [8-1] ANIMAL BREEDING NOTES CHAPTER 8 BEST PREDICTION
Mauricio A. Elzo, University of Florida, 1996, 2005, 2006, 2010, 2014. [8-1] ANIMAL BREEDING NOTES CHAPTER 8 BEST PREDICTION Derivation of the Best Predictor (BP) Let g = [g 1 g 2... g p ] be a vector
More informationPhysics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester
Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability
More informationOutput entropy of tensor products of random quantum channels 1
Output entropy of tensor products of random quantum channels 1 Motohisa Fukuda 1 [Collins, Fukuda and Nechita], [Collins, Fukuda and Nechita] 1 / 13 1 Introduction Aim and background Existence of one large
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationDiagonalization. MATH 1502 Calculus II Notes. November 4, 2008
Diagonalization MATH 1502 Calculus II Notes November 4, 2008 We want to understand all linear transformations L : R n R m. The basic case is the one in which n = m. That is to say, the case in which the
More informationSiegel s formula via Stein s identities
Siegel s formula via Stein s identities Jun S. Liu Department of Statistics Harvard University Abstract Inspired by a surprising formula in Siegel (1993), we find it convenient to compute covariances,
More informationEstimation of the Global Minimum Variance Portfolio in High Dimensions
Estimation of the Global Minimum Variance Portfolio in High Dimensions Taras Bodnar, Nestor Parolya and Wolfgang Schmid 07.FEBRUARY 2014 1 / 25 Outline Introduction Random Matrix Theory: Preliminary Results
More informationStatistical Inference On the High-dimensional Gaussian Covarianc
Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional
More informationLinear Operators Preserving the Numerical Range (Radius) on Triangular Matrices
Linear Operators Preserving the Numerical Range (Radius) on Triangular Matrices Chi-Kwong Li Department of Mathematics, College of William & Mary, P.O. Box 8795, Williamsburg, VA 23187-8795, USA. E-mail:
More informationSTAT C206A / MATH C223A : Stein s method and applications 1. Lecture 31
STAT C26A / MATH C223A : Stein s method and applications Lecture 3 Lecture date: Nov. 7, 27 Scribe: Anand Sarwate Gaussian concentration recap If W, T ) is a pair of random variables such that for all
More informationMathematical Methods for Quantum Information Theory. Part I: Matrix Analysis. Koenraad Audenaert (RHUL, UK)
Mathematical Methods for Quantum Information Theory Part I: Matrix Analysis Koenraad Audenaert (RHUL, UK) September 14, 2008 Preface Books on Matrix Analysis: R. Bhatia, Matrix Analysis, Springer, 1997.
More informationFE670 Algorithmic Trading Strategies. Stevens Institute of Technology
FE670 Algorithmic Trading Strategies Lecture 8. Robust Portfolio Optimization Steve Yang Stevens Institute of Technology 10/17/2013 Outline 1 Robust Mean-Variance Formulations 2 Uncertain in Expected Return
More informationSpecification Test Based on the Hansen-Jagannathan Distance with Good Small Sample Properties
Specification Test Based on the Hansen-Jagannathan Distance with Good Small Sample Properties Yu Ren Department of Economics Queen s University Katsumi Shimotsu Department of Economics Queen s University
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationPrincipal Component Analysis
Principal Component Analysis Laurenz Wiskott Institute for Theoretical Biology Humboldt-University Berlin Invalidenstraße 43 D-10115 Berlin, Germany 11 March 2004 1 Intuition Problem Statement Experimental
More informationMS-E2112 Multivariate Statistical Analysis (5cr) Lecture 1: Introduction, Multivariate Location and Scatter
MS-E2112 Multivariate Statistical Analysis (5cr) Lecture 1:, Multivariate Location Contents , pauliina.ilmonen(a)aalto.fi Lectures on Mondays 12.15-14.00 (2.1. - 6.2., 20.2. - 27.3.), U147 (U5) Exercises
More informationThomas J. Fisher. Research Statement. Preliminary Results
Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these
More informationDATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD
DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary
More informationDynamic Asset Allocation - Identifying Regime Shifts in Financial Time Series to Build Robust Portfolios
Downloaded from orbit.dtu.dk on: Jan 22, 2019 Dynamic Asset Allocation - Identifying Regime Shifts in Financial Time Series to Build Robust Portfolios Nystrup, Peter Publication date: 2018 Document Version
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationPreliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38
Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Guy Lebanon February 19, 2011 Maximum likelihood estimation is the most popular general purpose method for obtaining estimating a distribution from a finite sample. It was
More informationThe exact bias of S 2 in linear panel regressions with spatial autocorrelation SFB 823. Discussion Paper. Christoph Hanck, Walter Krämer
SFB 83 The exact bias of S in linear panel regressions with spatial autocorrelation Discussion Paper Christoph Hanck, Walter Krämer Nr. 8/00 The exact bias of S in linear panel regressions with spatial
More informationMyths about the Analysis of Change
James H. Steiger Department of Psychology and Human Development Vanderbilt University, 2010 1 Introduction 2 Limitations of the Two-Observation Study 3 Myth 2: The Difference Score is Intrinsically Unreliable
More informationChapter 5. Eigenvalues and Eigenvectors
Chapter 5 Eigenvalues and Eigenvectors Section 5. Eigenvectors and Eigenvalues Motivation: Difference equations A Biology Question How to predict a population of rabbits with given dynamics:. half of the
More informationIntroduction to Computational Finance and Financial Econometrics Probability Review - Part 2
You can t see this text! Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Probability Review - Part 2 1 /
More informationNon white sample covariance matrices.
Non white sample covariance matrices. S. Péché, Université Grenoble 1, joint work with O. Ledoit, Uni. Zurich 17-21/05/2010, Université Marne la Vallée Workshop Probability and Geometry in High Dimensions
More informationTangency portfolio weights for singular covariance matrix in small and large dimensions: estimation and test theory
Mathematical Statistics Stockholm University Tangency portfolio weights for singular covariance matrix in small and large dimensions: estimation and test theory Taras Bodnar Stepan Mazur Krzysztof Podgórski
More informationEstimation for Mean and Standard Deviation of Normal Distribution under Type II Censoring
Communications for Statistical Applications and Methods 2014, Vol. 21, No. 6, 529 538 DOI: http://dx.doi.org/10.5351/csam.2014.21.6.529 Print ISSN 2287-7843 / Online ISSN 2383-4757 Estimation for Mean
More informationThe Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility
The Slow Convergence of OLS Estimators of α, β and Portfolio Weights under Long Memory Stochastic Volatility New York University Stern School of Business June 21, 2018 Introduction Bivariate long memory
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R
More informationL-momenty s rušivou regresí
L-momenty s rušivou regresí Jan Picek, Martin Schindler e-mail: jan.picek@tul.cz TECHNICKÁ UNIVERZITA V LIBERCI ROBUST 2016 J. Picek, M. Schindler, TUL L-momenty s rušivou regresí 1/26 Motivation 1 Development
More informationHigh Dimensional Covariance and Precision Matrix Estimation
High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance
More informationSolutions Problem Set 8 Math 240, Fall
Solutions Problem Set 8 Math 240, Fall 2012 5.6 T/F.2. True. If A is upper or lower diagonal, to make det(a λi) 0, we need product of the main diagonal elements of A λi to be 0, which means λ is one of
More informationTime Series Modeling of Financial Data. Prof. Daniel P. Palomar
Time Series Modeling of Financial Data Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19,
More informationAppendix to Portfolio Construction by Mitigating Error Amplification: The Bounded-Noise Portfolio
Appendix to Portfolio Construction by Mitigating Error Amplification: The Bounded-Noise Portfolio Long Zhao, Deepayan Chakrabarti, and Kumar Muthuraman McCombs School of Business, University of Texas,
More informationMIT Spring 2015
Regression Analysis MIT 18.472 Dr. Kempthorne Spring 2015 1 Outline Regression Analysis 1 Regression Analysis 2 Multiple Linear Regression: Setup Data Set n cases i = 1, 2,..., n 1 Response (dependent)
More informationLecture 15. Hypothesis testing in the linear model
14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma
More information15.S24 Sample Exam Solutions
5.S4 Sample Exam Solutions. In each case, determine whether V is a vector space. If it is not a vector space, explain why not. If it is, find basis vectors for V. (a) V is the subset of R 3 defined by
More informationAccurate directional inference for vector parameters
Accurate directional inference for vector parameters Nancy Reid February 26, 2016 with Don Fraser, Nicola Sartori, Anthony Davison Nancy Reid Accurate directional inference for vector parameters York University
More information