Testing a Normal Covariance Matrix for Small Samples with Monotone Missing Data
|
|
- Ira Garrett
- 6 years ago
- Views:
Transcription
1 Applied Mathematical Sciences, Vol 3, 009, no 54, Testing a Normal Covariance Matrix for Small Samples with Monotone Missing Data Evelina Veleva Rousse University A Kanchev Department of Numerical Methods and Statistics Bulgaria, 707 Rousse, 8 Studentska str, room 44 eveleva@ruacadbg Abstract We consider samples with monotone missing data, drawn from a normal population to test if the covariance matrix is equal to a given positive definite matrix We propose an imputation procedure for the missing data and give the exact distribution of the corresponding likelihood ratio test statistic from the classical complete case Mathematics Subject Classification: 6H0, 6H5 Keywords: monotone missing data, Bellman gamma distribution, covariance matrix, hypotheses testing Introduction The problem of missing data is an important applied problem, because missing values are encountered in many practical situations (see [4]) Two commonly used approaches to analyze incomplete data are the likelihood based approach and the multiple imputation The imputation method is to impute the missing data to form complete data and then use the standard methods available for the complete data analysis For a good exposition of the imputation procedures and validity of imputation inferences in practice, we refer to [6] In this paper we consider samples with monotone missing data pattern Let (X,,X n ) t be a random vector with multivariate normal distribution N n (μ, Σ),where the mean vector μ and the covariance matrix Σ are unknown Suppose that we have k + + k n independent observations, k of which are on (X,,X n ) t, k -on(x,,x n ) t and so on, k n on the random variable X n Assume that k j 0 and m j = k + + k j >j, j =,,n The data
2 696 E Veleva can be written in the following pattern, known as a monotone pattern x, x,m x n, x n,m x n,m + x n, x n, x n,m x n,m + x n, x n, () In the literature on inference for μ and Σ, it is noticeable that the exact distributions of ˆμ and ˆΣ, the maximum likelihood estimators of μ and Σ, have remained unknown This problem is basic to inference with incomplete data when large samples are infeasible or impractical (see []) In [] the authors initiate a program of research on inference for μ and Σ with the goal of deriving explicit results analogous to those existing in the classical complete case In this paper we consider hypotheses H 0 :Σ=Σ 0 against H a :Σ Σ 0, propose an imputation procedure for the missing data in () and give the exact distribution of the corresponding likelihood ratio test statistic from the classical complete case Preliminary Notes We shall use the following known propositions, which can be found in [5] Proposition Let the n random vector x be normally distributed according to x N n (μ, Σ), then the m random vector y, obtained by the linear transformation y =Ax +c, where A denotes an m n matrix of constants with full rank m and c an m vector of constants, has the normal distribution y N m (Aμ +c, AΣA t ) Proposition Let the n random vector x be normally distributed according to x N n (μ, Σ), then x t A x, where A is an n n matrix of constants has noncentral chi-square distribution χ (ranka,μ t Aμ), if and only if the matrix AΣ is idempotent Proposition 3 If A is idempotent, then rank A = tra Proposition 4 If the n n matrix A with ranka =r is idempotent, then I n A is also idempotent with rank(i n A) =n r Proposition 5 Let the n random vector x be normally distributed according to x N n (μ, Σ), then the linear form Ax and the quadratic form x t Bx with the positive definite or positive semidefinite matrix B are independent, if and only if AΣB = 0
3 Testing a normal covariance matrix 697 Let A be a real square matrix of order n Let α and β be nonempty subsets of the set N n = {,,n} By A[α, β] we denote the submatrix of A, composed of the rows with numbers from α and the colus with numbers from β When β α, A[α, α] is denoted simply by A[α] The Bellman gamma distribution is a matrix variate distribution, which is a generalization of the Wishart and the matrix gamma distributions The next definition is given in [3] By Γ n (a,,a n ) is denoted the generalized multivariate gamma function, Γ n(a,,a n )=π n(n )/4 n Γ ( a j (j )), for a j > (j ), j =,,n Definition 6 A random positive definite matrix U (n n) is said to follow Bellman gamma type I distribution, denoted by U BG I n (a,,a n ;C), if its probability density function is given by n (det C[{i,,n}]) a i a i i= (det U) an (n+)/ Γ n (a,,a n ) n etr( CU), (det U[{,,i }]) a i a i i= where C (n n) is a positive definite constant matrix, a 0 =0and a j > (j ), j =,,n 3 Main Results Let us replace the missing values in () by zero and denote the obtained matrix by X, x, x,m X = x n, x n,m x n, 0 0 () x n, x n,m x n, x n, + x n, Theorem 3 Let the matrix X, defined by () presents the observations () on a random vector (X,,X n ) t with multivariate standard normal distribution N n (0, I n ) Then the matrix W = XX t has Bellman gamma type I distribution BG I n ( m,, ; I n) Proof Let W =(w i,j ) be the matrix W = XX t and let w i =(w,i,, w i,i ) t, i =,,n Let us denote by z i the vector of available observations on X i, ie z i = (x i,,,x i,mi ) t, i =,,n Let X i be the matrix X i = X[{,,i}, {,,m i+ }], i =,,n The distribution of z i is N mi (0, I mi ), i =,,n Since w i = X i z i and X i X t i = j=
4 698 E Veleva W[{,,i }], according to Proposition the distribution of w i, i =,,n is N i (0, W[{,,i }]) under the condition that the matrix X i has fixed elements It is easy to see that ( ) W[{,,i }] wi W[{,,i}] = Let w t i w i,i w (i) = w i,i w t i (W[{,,i }]) w i, (3) ie w (i) = det W[{,,i}],i=,,n (4) det W[{,,i }] For i =,,n, w (i) can be written in the form w (i) = z t i z i z t i X t i (X i X t i ) X i z i = z t i[i mi X t i (X i X t i ) X i ]z i Since X i [I mi X t i (X i X t i ) X i ]=X i X i =0, according to Proposition 5, w i is independent of w (i) under the condition that the matrix X i has fixed elements Additionally, since the matrix X t i (X i X t i ) X i is idempotent, from Proposition 3 we have that rank(x t i (X i X t i ) X i )=tr(x t i (X i X t i ) X i )=tr(i i )=i Hence, according to Proposition 4 the matrix I mi X t i (X i X t i ) X i is also idempotent with the rank m i i+ Now, applying Proposition we get that w (i) has chi-square distribution χ (m i i + ), again under the condition that the matrix X i has fixed elements The conditional distributions of w (i) and w i depend only on the elements of the matrix W[{,,i }] Therefore the joint density of w,, w (), w,, w (n), w n will have the form w m, e w, Γ ( ) m m n w mi i+ (i) e w (i) Γ ( m i ) i+ i= m i i+ e wt i (W [{,,i }]) w i (π) i (det W [{,,i }]) By the transformation of the variables w (i) into w i,i by means of (3) with det J = we get the distribution of the elements w i,j of W, which using (4) can be written in the form m + + Γ n ( m ),, (det W ) [ (n+)]/ n (det W [{,,i }]) (m i m i )/ i= etr ( ) W
5 Testing a normal covariance matrix 699 ( Consequently, according to Definition W BG I m n,, ; I n) Let z i be the vector of available observations on X i in (), ie z i = (x i,,,x i,mi ) t, i =,,n Let us denote by z i the mean of the elements of z i, i =,,n Consider the data matrix x, x,m z z z X = x n, x n,m x n, z n z n, x n, x n,m x n, x n, + x n, (5) in which we substitute z,, z n for the missing values in,, n th row respectively of the data in () Theorem 3 Let the matrix X, defined by (5) presents the observations () on a random vector (X,,X n ) t with multivariate normal distribution N n (μ, I n ) Let x i, i =,,m n be the colu vectors of the matrix X, x be the vector x=( z,, z n ) t and S be the matrix m n S = (x i x)(x i x) t (6) i= Then x and S are independent, x ( N n (μ, diag(m,,m n )) and S has Bellman gamma distribution BG I m n,, ; I n) Proof Let Θ be an orthogonal m n m n matrix of the form: θ, θ,m θ,m + θ,m θ, θ, θ,m θ,m + θ,m θ, θ m, θ m,m θ,m + θ,m θ, Θ= 0 0 θ m +,m + θ m +,m θ, 0 0 θ m,m + θ m,m θ, θ, θ, The all elements in the first colu of Θ are equal to In colus from tom, the last m n m elements are equal to zero In colus from m j +
6 700 E Veleva to m j+, j =,,n the last m n m j+ elements are equal to zero and the first m j elements are equal to each other It can be easily checked that such orthogonal matrix exists and even is not unique Let Y =(y i,j ) be the matrix Y = X Θ Then YY t = X ΘΘ t X t = XX t Let us denote by y i, i =,,m n the colu vectors of Y It is easy to see that y = m n x Therefore the matrix S, defined by (6) can be written in the form m n m n S = x i x t i m n x x t = XX t y y t = YYt y y t = y i yi t (7) i= Since the matrix Θ is orthogonal, θ t θ i = 0 for i =,,m n, where θ i, i =,,m n denote the colu vectors of Θ Therefore, the sum of elements of θ i equals to zero, i =,,m n Hence it follows that E(y i )=0,i =,,m n It can be checked that in the matrix Y the all elements, lying on the places of the missing data in () are equal to zero Let y k,s, s be a nonzero element of Y Then <s, therefore the last m n elements of the vector θ s are equal to zero Let the vector θ s has m nonzero elements, then y k,s = x k, θ,s + + x k,m θ m,s (8) The distribution of y k,s as a linear combination of independent normal distributed random variables is also normal The variance of y k,s is Var(y k,s )=θ,s Var(x k,)+ + θ m,s Var(x k,m) =θ,s + + θ m,s = For k =,,n i= y k, = m n z k = (x k, + + x k,mk ) (9) Consequently y k, is also a linear combination of the observations in the k th row of () and hence is normally distributed, E(y k, )= [E(x k, )+ + E(x k,mk )] = m n μ k, where μ k is the k th coordinate of the mean vector μ and Var(y k, )= m n [Var(x m k, )+ + Var(x k,mk )] = m n k Since (X,,X n ) t N n (μ, I n ), the vectors z,, z n are independent Consequently, if k,, k p are different integers from the interval [,n] and y k,s,, y kp,s p are nonzero elements of the matrix Y, then they are independent Hence for i =,,m n the nonzero elements of the vector y i have joint
7 Testing a normal covariance matrix 70 multivariate normal distribution For i =,,m n it is N n j (0, I n j ), where j is the biggest integer, such that m j <i(m 0 = ) Since y = m n x, the distribution of x is N n (μ, diag(m,,m n )) From (8) and (9) it follows that if y k,s,, y k,sp are arbitrary nonzero elements of Y, then they have joint multivariate normal distribution Moreover, (8) and (9) can be written in the form y k,s =(x k, μ k )θ,s + +(x k,m μ k )θ m,s,s, y k, =(x k, μ k ) + +(x k,mk μ k ) + μ k Hence for s> Cov(y k,,y k,s )=E(y k, y k,s )=E(x k, μ k ) θ,s + and for <s<q + E(x k,m μ k ) θ m,s =(θ,s + + θ m,s ) =0 Cov(y k,s,y k,q )=E(y k,s y k,q )=E(x k, μ k ) θ,s θ,q + + E(x k,m μ k ) θ m,s θ m,q = θ,s θ,q + + θ m,s θ m,q =0 Therefore y k,s,, y k,sp are independent and hence the vectors y,, y are independent Consequently from (7) and Theorem 3 the Theorem follows For the data in (), let us consider the hypotheses H 0 :Σ=Σ 0 against H a : Σ Σ 0, where Σ 0 is an arbitrary positive definite matrix of size n It is shown in [4], that the testing problem is invariant under a suitable transformation of the data in () and without loss of generality we can assume that Σ 0 =I n, where I n is the identity matrix of size n It is easy to see that under H 0 :Σ=I n, the maximum likelihood estimations for the missing values in the j th row in () are equal to z j, j =,,n (m n ) The matrix the data matrix (5) Let us consider the modified likelihood ratio test statistic λ, S is actually the empirical covariance matrix, obtained from λ =(e/m n ) n/ (det S) ( )/ e (trs)/, which is unbiased in the classical case of fully observed data matrix (see []) Theorem 33 Under H 0 :Σ=I n, λ is distributed as the product K (ζ ζ n ) ( )/ η η n,
8 70 E Veleva where K is the constant K =(e/m n ) n/ n( )/, ζ,, ζ n, η,, η n are mutually independent random variables, ζ j has Beta distribution Beta ( (m j+ j )/,j/), j =,,n and η j ξ ( )/ j e ξ j, where ξ j is Gamma distributed G((m j )/, ), j =,,n The proof of Theorem 33 will appear in a subsequent paper Stochastic representations of the Bellman gamma distribution Since Theorem 33 give the exact distribution of the test statistic λ, the test procedure, suggested in this paper is proper for small samples References [] W Chang and D St P Richards, Finite - sample inference with monotone incomplete multivariate normal data I, J Multivariate Anal, In Press, Corrected Proof, Available online 4 May 009 [] NC Giri, Multivariate Statistical Analysis, Marcel Dekker Inc, New York, 004 [3] AK Gupta and DK Nagar, Matrix variate distributions, Chapman & Hall/CRC, 000 [4] J Hao and K Krishnamoorthy, Inferences on a normal covariance matrix and generalized variance with monotone missing data, J Multivariate Anal, 78 (00), 6 8 [5] K R Koch, Estimation and Hypothesis Testing in Linear Models, Springer - Verlag, Berlin, 999 [6] JA Little and DB Rubin, Statistical Analysis With Missing Data, nd edition, Wiley - Interscience, Hoboken, NJ, 00 Received: May, 009
Some New Properties of Wishart Distribution
Applied Mathematical Sciences, Vol., 008, no. 54, 673-68 Some New Properties of Wishart Distribution Evelina Veleva Rousse University A. Kanchev Department of Numerical Methods and Statistics 8 Studentska
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationTEST FOR INDEPENDENCE OF THE VARIABLES WITH MISSING ELEMENTS IN ONE AND THE SAME COLUMN OF THE EMPIRICAL CORRELATION MATRIX.
Serdica Math J 34 (008, 509 530 TEST FOR INDEPENDENCE OF THE VARIABLES WITH MISSING ELEMENTS IN ONE AND THE SAME COLUMN OF THE EMPIRICAL CORRELATION MATRIX Evelina Veleva Communicated by N Yanev Abstract
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationBIOS 2083 Linear Models Abdus S. Wahed. Chapter 2 84
Chapter 2 84 Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random variables. For instance, X = X 1 X 2.
More informationInferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data
Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone
More informationBayesian Inference. Chapter 9. Linear models and regression
Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering
More informationChapter 17: Undirected Graphical Models
Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)
More information5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.
88 Chapter 5 Distribution Theory In this chapter, we summarize the distributions related to the normal distribution that occur in linear models. Before turning to this general problem that assumes normal
More informationRandom Vectors and Multivariate Normal Distributions
Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random 75 variables. For instance, X = X 1 X 2., where each
More informationCh4. Distribution of Quadratic Forms in y
ST4233, Linear Models, Semester 1 2008-2009 Ch4. Distribution of Quadratic Forms in y 1 Definition Definition 1.1 If A is a symmetric matrix and y is a vector, the product y Ay = i a ii y 2 i + i j a ij
More informationLecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012
Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationPreliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38
Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters
More informationLinear Models in Statistics
Linear Models in Statistics ALVIN C. RENCHER Department of Statistics Brigham Young University Provo, Utah A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationTAMS39 Lecture 2 Multivariate normal distribution
TAMS39 Lecture 2 Multivariate normal distribution Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content Lecture Random vectors Multivariate normal distribution
More informationCOMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS
Communications in Statistics - Simulation and Computation 33 (2004) 431-446 COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS K. Krishnamoorthy and Yong Lu Department
More informationLecture 15. Hypothesis testing in the linear model
14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma
More informationAnalysis of variance, multivariate (MANOVA)
Analysis of variance, multivariate (MANOVA) Abstract: A designed experiment is set up in which the system studied is under the control of an investigator. The individuals, the treatments, the variables
More informationRandom Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30
Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices
More informationOn the Distribution of a Quadratic Form in Normal Variates
On the Distribution of a Quadratic Form in Normal Variates Jin Zhang School of Mathematics and Statistics, Yunnan University, Kunming, Yunnan, 650091, China E-mail: jinzhang@ynu.edu.cn Abstract It is a
More informationStatistical Inference with Monotone Incomplete Multivariate Normal Data
Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful
More informationMultivariate Gaussian Analysis
BS2 Statistical Inference, Lecture 7, Hilary Term 2009 February 13, 2009 Marginal and conditional distributions For a positive definite covariance matrix Σ, the multivariate Gaussian distribution has density
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationSOME THEOREMS ON QUADRATIC FORMS AND NORMAL VARIABLES. (2π) 1 2 (σ 2 ) πσ
SOME THEOREMS ON QUADRATIC FORMS AND NORMAL VARIABLES 1. THE MULTIVARIATE NORMAL DISTRIBUTION The n 1 vector of random variables, y, is said to be distributed as a multivariate normal with mean vector
More informationRestricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model
Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives
More informationChapter 1. Matrix Algebra
ST4233, Linear Models, Semester 1 2008-2009 Chapter 1. Matrix Algebra 1 Matrix and vector notation Definition 1.1 A matrix is a rectangular or square array of numbers of variables. We use uppercase boldface
More informationMean. Pranab K. Mitra and Bimal K. Sinha. Department of Mathematics and Statistics, University Of Maryland, Baltimore County
A Generalized p-value Approach to Inference on Common Mean Pranab K. Mitra and Bimal K. Sinha Department of Mathematics and Statistics, University Of Maryland, Baltimore County 1000 Hilltop Circle, Baltimore,
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More informationLecture 15: Multivariate normal distributions
Lecture 15: Multivariate normal distributions Normal distributions with singular covariance matrices Consider an n-dimensional X N(µ,Σ) with a positive definite Σ and a fixed k n matrix A that is not of
More informationLecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution
Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution Relevant
More informationLECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More informationMLES & Multivariate Normal Theory
Merlise Clyde September 6, 2016 Outline Expectations of Quadratic Forms Distribution Linear Transformations Distribution of estimates under normality Properties of MLE s Recap Ŷ = ˆµ is an unbiased estimate
More informationSTAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.
STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More informationApplied Statistics Preliminary Examination Theory of Linear Models August 2017
Applied Statistics Preliminary Examination Theory of Linear Models August 2017 Instructions: Do all 3 Problems. Neither calculators nor electronic devices of any kind are allowed. Show all your work, clearly
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationJournal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error
Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA
More information4. Distributions of Functions of Random Variables
4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n
More informationMatrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved.
Matrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved. 1 Converting Matrices Into (Long) Vectors Convention:
More informationy = 1 N y i = 1 N y. E(W )=a 1 µ 1 + a 2 µ a n µ n (1.2) and its variance is var(w )=a 2 1σ a 2 2σ a 2 nσ 2 n. (1.
The probability density function of a continuous random variable Y (or the probability mass function if Y is discrete) is referred to simply as a probability distribution and denoted by f(y; θ) where θ
More informationON COMBINING CORRELATED ESTIMATORS OF THE COMMON MEAN OF A MULTIVARIATE NORMAL DISTRIBUTION
ON COMBINING CORRELATED ESTIMATORS OF THE COMMON MEAN OF A MULTIVARIATE NORMAL DISTRIBUTION K. KRISHNAMOORTHY 1 and YONG LU Department of Mathematics, University of Louisiana at Lafayette Lafayette, LA
More informationModelling Dropouts by Conditional Distribution, a Copula-Based Approach
The 8th Tartu Conference on MULTIVARIATE STATISTICS, The 6th Conference on MULTIVARIATE DISTRIBUTIONS with Fixed Marginals Modelling Dropouts by Conditional Distribution, a Copula-Based Approach Ene Käärik
More informationLikelihood Ratio Criterion for Testing Sphericity from a Multivariate Normal Sample with 2-step Monotone Missing Data Pattern
The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 473-481 Likelihood Ratio Criterion for Testing Sphericity from a Multivariate Normal Sample with 2-step Monotone Missing Data Pattern Byungjin
More informationAn Introduction to Multivariate Statistical Analysis
An Introduction to Multivariate Statistical Analysis Third Edition T. W. ANDERSON Stanford University Department of Statistics Stanford, CA WILEY- INTERSCIENCE A JOHN WILEY & SONS, INC., PUBLICATION Contents
More informationCS 195-5: Machine Learning Problem Set 1
CS 95-5: Machine Learning Problem Set Douglas Lanman dlanman@brown.edu 7 September Regression Problem Show that the prediction errors y f(x; ŵ) are necessarily uncorrelated with any linear function of
More informationDistributions of Quadratic Forms. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 31
Distributions of Quadratic Forms Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 31 Under the Normal Theory GMM (NTGMM), y = Xβ + ε, where ε N(0, σ 2 I). By Result 5.3, the NTGMM
More informationAdvanced topics from statistics
Advanced topics from statistics Anders Ringgaard Kristensen Advanced Herd Management Slide 1 Outline Covariance and correlation Random vectors and multivariate distributions The multinomial distribution
More informationOrthogonal decompositions in growth curve models
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 4, Orthogonal decompositions in growth curve models Daniel Klein and Ivan Žežula Dedicated to Professor L. Kubáček on the occasion
More informationOn Expected Gaussian Random Determinants
On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationHypothesis testing when a nuisance parameter is present only under the alternative - linear model case
Hypothesis testing when a nuisance parameter is present only under the alternative - linear model case By ROBERT B. DAVIES Statistics Research Associates imited PO Box 12 649, Thorndon, Wellington, New
More information3 Multiple Linear Regression
3 Multiple Linear Regression 3.1 The Model Essentially, all models are wrong, but some are useful. Quote by George E.P. Box. Models are supposed to be exact descriptions of the population, but that is
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationSmall area estimation with missing data using a multivariate linear random effects model
Department of Mathematics Small area estimation with missing data using a multivariate linear random effects model Innocent Ngaruye, Dietrich von Rosen and Martin Singull LiTH-MAT-R--2017/07--SE Department
More informationComputational Physics
Interpolation, Extrapolation & Polynomial Approximation Lectures based on course notes by Pablo Laguna and Kostas Kokkotas revamped by Deirdre Shoemaker Spring 2014 Introduction In many cases, a function
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationA Comparative Study of Imputation Methods for Estimation of Missing Values of Per Capita Expenditure in Central Java
IOP Conference Series: Earth and Environmental Science PAPER OPEN ACCESS A Comparative Study of Imputation Methods for Estimation of Missing Values of Per Capita Expenditure in Central Java To cite this
More informationTopic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form
Topic 7 - Matrix Approach to Simple Linear Regression Review of Matrices Outline Regression model in matrix form - Fall 03 Calculations using matrices Topic 7 Matrix Collection of elements arranged in
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter
More informationSOME ASPECTS OF MULTIVARIATE BEHRENS-FISHER PROBLEM
SOME ASPECTS OF MULTIVARIATE BEHRENS-FISHER PROBLEM Junyong Park Bimal Sinha Department of Mathematics/Statistics University of Maryland, Baltimore Abstract In this paper we discuss the well known multivariate
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationEconomics 573 Problem Set 5 Fall 2002 Due: 4 October b. The sample mean converges in probability to the population mean.
Economics 573 Problem Set 5 Fall 00 Due: 4 October 00 1. In random sampling from any population with E(X) = and Var(X) =, show (using Chebyshev's inequality) that sample mean converges in probability to..
More informationModule 9: Stationary Processes
Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.
More informationLecture 20: Linear model, the LSE, and UMVUE
Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;
More informationANALYSIS OF VARIANCE AND QUADRATIC FORMS
4 ANALYSIS OF VARIANCE AND QUADRATIC FORMS The previous chapter developed the regression results involving linear functions of the dependent variable, β, Ŷ, and e. All were shown to be normally distributed
More informationANOVA: Analysis of Variance - Part I
ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.
More informationidentity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij
Notation R R n m R n m r R n s real numbers set of n m real matrices subset of R n m consisting of matrices with rank r subset of R n n consisting of symmetric matrices NND n subset of R n s consisting
More informationLinear Models 1. Isfahan University of Technology Fall Semester, 2014
Linear Models 1 Isfahan University of Technology Fall Semester, 2014 References: [1] G. A. F., Seber and A. J. Lee (2003). Linear Regression Analysis (2nd ed.). Hoboken, NJ: Wiley. [2] A. C. Rencher and
More informationOptimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.
Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationThe Multivariate Normal Distribution 1
The Multivariate Normal Distribution 1 STA 302 Fall 2014 1 See last slide for copyright information. 1 / 37 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2
More informationEstimating the parameters of hidden binomial trials by the EM algorithm
Hacettepe Journal of Mathematics and Statistics Volume 43 (5) (2014), 885 890 Estimating the parameters of hidden binomial trials by the EM algorithm Degang Zhu Received 02 : 09 : 2013 : Accepted 02 :
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationc 2005 Society for Industrial and Applied Mathematics
SIAM J. MATRIX ANAL. APPL. Vol. XX, No. X, pp. XX XX c 005 Society for Industrial and Applied Mathematics DISTRIBUTIONS OF THE EXTREME EIGENVALUES OF THE COMPLEX JACOBI RANDOM MATRIX ENSEMBLE PLAMEN KOEV
More informationLinear Models for Multivariate Repeated Measures Data
THE UNIVERSITY OF TEXAS AT SAN ANTONIO, COLLEGE OF BUSINESS Working Paper SERIES Date December 3, 200 WP # 007MSS-253-200 Linear Models for Multivariate Repeated Measures Data Anuradha Roy Management Science
More informationStatistical Inference of Moment Structures
1 Handbook of Computing and Statistics with Applications, Vol. 1 ISSN: 1871-0301 2007 Elsevier B.V. All rights reserved DOI: 10.1016/S1871-0301(06)01011-0 1 Statistical Inference of Moment Structures Alexander
More information5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors
EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we
More informationABOUT PRINCIPAL COMPONENTS UNDER SINGULARITY
ABOUT PRINCIPAL COMPONENTS UNDER SINGULARITY José A. Díaz-García and Raúl Alberto Pérez-Agamez Comunicación Técnica No I-05-11/08-09-005 (PE/CIMAT) About principal components under singularity José A.
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationMultivariate Regression Generalized Likelihood Ratio Tests for FMRI Activation
Multivariate Regression Generalized Likelihood Ratio Tests for FMRI Activation Daniel B Rowe Division of Biostatistics Medical College of Wisconsin Technical Report 40 November 00 Division of Biostatistics
More informationTopics in Probability and Statistics
Topics in Probability and tatistics A Fundamental Construction uppose {, P } is a sample space (with probability P), and suppose X : R is a random variable. The distribution of X is the probability P X
More informationBasic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = The errors are uncorrelated with common variance:
8. PROPERTIES OF LEAST SQUARES ESTIMATES 1 Basic Distributional Assumptions of the Linear Model: 1. The errors are unbiased: E[ε] = 0. 2. The errors are uncorrelated with common variance: These assumptions
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More informationTMA4267 Linear Statistical Models V2017 (L10)
TMA4267 Linear Statistical Models V2017 (L10) Part 2: Linear regression: Parameter estimation [F:3.2], Properties of residuals and distribution of estimator for error variance Confidence interval and hypothesis
More informationMULTIVARIATE POPULATIONS
CHAPTER 5 MULTIVARIATE POPULATIONS 5. INTRODUCTION In the following chapters we will be dealing with a variety of problems concerning multivariate populations. The purpose of this chapter is to provide
More informationSTAT 151A: Lab 1. 1 Logistics. 2 Reference. 3 Playing with R: graphics and lm() 4 Random vectors. Billy Fang. 2 September 2017
STAT 151A: Lab 1 Billy Fang 2 September 2017 1 Logistics Billy Fang (blfang@berkeley.edu) Office hours: Monday 9am-11am, Wednesday 10am-12pm, Evans 428 (room changes will be written on the chalkboard)
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More informationLinear Algebra Review
Linear Algebra Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Linear Algebra Review 1 / 45 Definition of Matrix Rectangular array of elements arranged in rows and
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationMeixner matrix ensembles
Meixner matrix ensembles W lodek Bryc 1 Cincinnati April 12, 2011 1 Based on joint work with Gerard Letac W lodek Bryc (Cincinnati) Meixner matrix ensembles April 12, 2011 1 / 29 Outline of talk Random
More informationDepartment of Large Animal Sciences. Outline. Slide 2. Department of Large Animal Sciences. Slide 4. Department of Large Animal Sciences
Outline Advanced topics from statistics Anders Ringgaard Kristensen Covariance and correlation Random vectors and multivariate distributions The multinomial distribution The multivariate normal distribution
More information8 - Continuous random vectors
8-1 Continuous random vectors S. Lall, Stanford 2011.01.25.01 8 - Continuous random vectors Mean-square deviation Mean-variance decomposition Gaussian random vectors The Gamma function The χ 2 distribution
More informationProblems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B
Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2
More informationMultiple Imputation for Missing Data in Repeated Measurements Using MCMC and Copulas
Multiple Imputation for Missing Data in epeated Measurements Using MCMC and Copulas Lily Ingsrisawang and Duangporn Potawee Abstract This paper presents two imputation methods: Marov Chain Monte Carlo
More informationDepartment of Statistics
Research Report Department of Statistics Research Report Department of Statistics No. 05: Testing in multivariate normal models with block circular covariance structures Yuli Liang Dietrich von Rosen Tatjana
More information