Consistent Bivariate Distribution

Similar documents
Next is material on matrix rank. Please see the handout

An Introduction to Multivariate Statistical Analysis

2. Matrix Algebra and Random Vectors

MAT 1332: CALCULUS FOR LIFE SCIENCES. Contents. 1. Review: Linear Algebra II Vectors and matrices Definition. 1.2.

. a m1 a mn. a 1 a 2 a = a n

Probability and Stochastic Processes

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

Copula Regression RAHUL A. PARSA DRAKE UNIVERSITY & STUART A. KLUGMAN SOCIETY OF ACTUARIES CASUALTY ACTUARIAL SOCIETY MAY 18,2011

Multivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013

Exam 2. Jeremy Morris. March 23, 2006

Multivariate Distributions

Advanced Introduction to Machine Learning CMU-10715

The Multivariate Gaussian Distribution [DRAFT]

STAT 151A: Lab 1. 1 Logistics. 2 Reference. 3 Playing with R: graphics and lm() 4 Random vectors. Billy Fang. 2 September 2017

Lecture 22: A Review of Linear Algebra and an Introduction to The Multivariate Normal Distribution

Switching Regime Estimation

Chapter 4: Factor Analysis

Introduction to Matrix Algebra and the Multivariate Normal Distribution

Multivariate Statistical Analysis

1 Appendix A: Matrix Algebra

Bayesian Inference for the Multivariate Normal

The Bayesian Approach to Multi-equation Econometric Model Estimation

IDENTIFIABILITY OF THE MULTIVARIATE NORMAL BY THE MAXIMUM AND THE MINIMUM

Factorization of Seperable and Patterned Covariance Matrices for Gibbs Sampling

9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006.

A User's Guide To Principal Components

Reduction of Random Variables in Structural Reliability Analysis

Department of Econometrics and Business Statistics

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

Introduction to Computational Finance and Financial Econometrics Probability Theory Review: Part 2

On Hadamard and Kronecker Products Over Matrix of Matrices

Classical Measurement Error with Several Regressors

Latent variable interactions

A Note on Identification Test Procedures. by Phoebus Dhrymes, Columbia University. October 1991, Revised November 1992

7.6 The Inverse of a Square Matrix

Gibbs Sampling in Linear Models #2

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Estimating Variances and Covariances in a Non-stationary Multivariate Time Series Using the K-matrix

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Large Sample Properties of Estimators in the Classical Linear Regression Model

TAMS39 Lecture 2 Multivariate normal distribution

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Distributions

STA 2201/442 Assignment 2

The Expectation-Maximization Algorithm

Lecture 3. Inference about multivariate normal distribution

Generalization of Schur s Lemma in Ring Representations on Modules over a Commutative Ring

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

CMPE 58K Bayesian Statistics and Machine Learning Lecture 5

The General Linear Model. Monday, Lecture 2 Jeanette Mumford University of Wisconsin - Madison

A New Class of Positively Quadrant Dependent Bivariate Distributions with Pareto

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Some Results on the Multivariate Truncated Normal Distribution

STAT 501 Assignment 1 Name Spring 2005

STT 843 Key to Homework 1 Spring 2018

Primary Decomposition of Ideals Arising from Hankel Matrices

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

Linear Models for Multivariate Repeated Measures Data

Stat 5101 Lecture Notes

To Estimate or Not to Estimate?

Chapter 7. Tridiagonal linear systems. Solving tridiagonal systems of equations. and subdiagonal. E.g. a 21 a 22 a A =

Regression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y.

CS8803: Statistical Techniques in Robotics Byron Boots. Hilbert Space Embeddings

Geographically weighted regression approach for origin-destination flows

Dipartimento di Matematica

A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common

Multivariate Normal & Wishart

Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments

STA 437: Applied Multivariate Statistics

MATH 2030: MATRICES ,, a m1 a m2 a mn If the columns of A are the vectors a 1, a 2,...,a n ; A is represented as A 1. .

Modelling Dropouts by Conditional Distribution, a Copula-Based Approach

Appendix 2. The Multivariate Normal. Thus surfaces of equal probability for MVN distributed vectors satisfy

Statistics Examples. Cathal Ormond

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

Control of Directional Errors in Fixed Sequence Multiple Testing

Analysis of variance, multivariate (MANOVA)

Bayesian Linear Regression

EXPLICIT EXPRESSIONS OF PROJECTORS ON CANONICAL VARIABLES AND DISTANCES BETWEEN CENTROIDS OF GROUPS. Haruo Yanai*

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

Random Vectors and Multivariate Normal Distributions

DART_LAB Tutorial Section 2: How should observations impact an unobserved state variable? Multivariate assimilation.

[y i α βx i ] 2 (2) Q = i=1

1 Data Arrays and Decompositions

A Introduction to Matrix Algebra and the Multivariate Normal Distribution

Syllabus for Applied Mathematics Graduate Student Qualifying Exams, Dartmouth Mathematics Department

This note derives marginal and conditional means and covariances when the joint distribution may be singular and discusses the resulting invariants.

COMPOSITE RELIABILITY MODELS FOR SYSTEMS WITH TWO DISTINCT KINDS OF STOCHASTIC DEPENDENCES BETWEEN THEIR COMPONENTS LIFE TIMES

ECON 3150/4150, Spring term Lecture 6

MAXIMUM LIKELIHOOD IN GENERALIZED FIXED SCORE FACTOR ANALYSIS 1. INTRODUCTION

Supermodular ordering of Poisson arrays

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014

Statistical Techniques II

x + 2y + 3z = 8 x + 3y = 7 x + 2z = 3

Regression models for multivariate ordered responses via the Plackett distribution

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

Lecture 6: Geometry of OLS Estimation of Linear Regession

BIOS 2083 Linear Models Abdus S. Wahed. Chapter 2 84

Small area estimation with missing data using a multivariate linear random effects model

Transcription:

A Characterization of the Normal Conditional Distributions MATSUNO 79 Therefore, the function ( ) = G( : a/(1 b2)) = N(0, a/(1 b2)) is a solu- tion for the integral equation (10). The constant times of ( ) as well as ( ) = 0 can also be its solutions, though they are not probability density functions. 4 4.1 Consistent Bivariate Distribution Derivation of the marginal distribution We obtain the following results from Lemma 3. Theorem 4 The marginal distribution ( ) consistent with a pair of conditional distributions (6) and (7), where 0 < 1 β 2 δ 2, is Proof. Assuming = 0, = 0 without loss of generality, we consider the case of (6) and (7). Replacing conditional distribution β2 22 of Lemma 1 by, we get a = N(bw, a), where b = β δ and a = > 0. Regarding this function 11 + as the kernel H(, w), we obtain the integral equation (10). Then, under the condition 0 <1 b2 = 1 β 2 δ 2, we get its solution ( ) = N(0, a/(1 b2)) from Lemma 3. The variance of this solution distribution is given as where If 0, 0, then we can consider the distributions of = and = which have zero means, and then we go back to the zero mean case mentioned above. A similar result follows for the marginal distribution h( ) of Y: Theorem 5 The marginal distribution h( ) consistent with a pair of conditional distributions (6) and (7), where 0 < 1 β 2 δ 2, is

80 4.2 Derivation of the joint distribution Theorem 6 The joint distribution consistent with a pair of conditional distributions (6) and (7), where 0 < 1-β 2 δ 2, is Proof. We can assume = 0, = 0 without loss of generality as in Proof of Theorem 4. If we consider the case of (6) and (7), then Theorem 4 shows that the consistent marginal distribution ( ) is (12) ( ) = N(, /θ). The joint distribution is given by the product of (12) and the conditional distribution (7). So we have After rearranging this, we get where φ = δ 2 11 + 22 and = δ( / ) 1/2. Under the assumption 0 < 1- β 2 δ 2, the function (15) is the density function of the joint distribution (14) with = 0, = 0. We also get the following result in a similar way. Theorem 7 The joint distribution consistent with a pair of conditional distributions (6) and (7), where 0 < 1-β 2 δ 2, is

A Characterization of the Normal Conditional Distributions(MATSUNO) 81 where the correlation coefficient is = β( / ) 1/2. So far, we have obtained two families of joint distributions consistent with (6) and (7), that is (15) and (16). The regression coefficients of (15) are And those of (16) are The coefficient (17) may be or may not be equal β. The coefficient (20) may be or may not be equal δ. In other words, the four parameters (β,, 11, 22) of (6) and (7) have much " degrees of freedom to determine uniquely the parameters of the joint distribution. So that, we have obtained two families of joint distributions. To make the two families of joint distributions of Theorems 6 and 7 coincide, we introduce another restriction in addition to the one 0 < 1 -β 2 δ 2 already imposed. Theorem 8 The joint distribution consistent with a pair of conditional distributions (6) and (7) under the assumptions

82 is N(, ), where Proof. Under the assumptions (21), Theorems 6 and 7 hold. The further assumption (22) makes the variance matrices of (14) and (16) of Theorems have the common values This can be rearranged to get It would be possible to show that provided the joint distribution is N(, ) with (23) and (24), the conditional distributions are (6) and (7). Theorem 8 gives the sufficient condition for that the distributions (6) and (7) are consistent with the bivariate normal distribution N(, ) defined by (23) and (24). Now, we summarize the discussion for the bivariate distributions: (i) If the conditional distributions (6) and (7) are derived from the

A Characterization of the Normal Conditional Distributions(MATSUNO) 83 joint distribution N(, ) with (23) and (24), then the conditions 1 > βδ 0 and δ 11 = β 22 hold. (ii) If the conditions 0 < 1-β 2 δ 2 and δ 11 = β 22 hold, then the conditional distributions (6) and (7) are derived from the joint distribution N(, ) with (23) and (24). (iii) The condition 1 > βδ 0 implies 0 < 1-β 2 δ 2, but not the converse. 5 Multivariate Normal Distribution We here turn to a general multivariate distribution case and consider an s-dimension random vector Z whose probability density function is That is, Z is assumed to be distributed as N(, ) with a mean vector every element of which lies in the interval (-, + ), and with a positive definite (symmetric) covariance matrix. According to the partition of Z as Z' = [ X', Y' ] = [(1 p), (1 q)], we partition the parameters, where 12 = ' 21. Letting be the conditional density function of X given Y =, and be the one of Y given X =, we have 2) 2) See, for instance, T. W. Anderson(1958) : An Introduction to Mulitivariate Analysis, New York: John Wiley and Sons.

84 Letting the regression coefficients B = and =, we have a constraint on the coefficients, 5.1 The problem Let us now consider random vectors X = ( p 1) and Y = (q 1). Without loss of generality, it is assumed that p q. Suppose in general that the normal conditional distributions of X and Y are, respectively, The problem now is to find the p + q dimension joint distribution or their joint density function and their marginal density functions ( ) and h( ). In particular, we try to find condition under which the joint distribution is multivariate normal. In a similar way to get (10), we have an integral equation where with d denoting a q-dimensional integration. Now, the problem is to find first a solution ( ) for the integral equation (27), and second a joint distribution, which is given as a product of this ( ) and the conditional distribution (26),. It will be seen that the problem in a general multivariate case will need cumbersome calculations. The preparation for that matter is the next section.

A Characterization of the Normal Conditional Distributions(MATSUNO) 85 6 Preparation 2 Some of the propositions given here may be fairly well known so that we don t provide proofs for them. 6.1 Matrix algebra Lemma 9 For two non-singular matrices A = (m m) and C = (n n), and two rectangular matrices B = (m n) and D = (n m), define then we have We will make use of the following propositions concerning characteristic roots of matrices. The characteristic roots in the propositions can be multiple roots and/or zeros. Lemma 10 (a) If the characteristic roots of a square matrix A= (n n) are λ 1,..., λ n, then the characteristic roots of A' are λ 1,..., λ n. (b) For two rectartgular matrices A = (m n) and B = (n m), where m n, let the characteristic roots of AB = (m m) be λ 1,..., λ n (including multiple roots), then the characteristic roots of BA = (n n) are λ 1,..., λ n and n-m zeros. Proof. (a) See Dhrymes (1978, p49, Proposition 42, (a)). (b) See Dhrymes (1978, p.51, Corollary 5). 3)

86 Lemma 11 For the regression coefficients B = (p q) and = (q p) of the conditional distributions (25) and (26), let the characteristic roots of the product B be λ 1,..., λ n (including multiple roots), then we have: (a) The characteristic roots of 'B' = (p p) are λ 1,..., λ p. (b) The characteristic roots of B = (q q) are λ 1,..., λ p and q-p zeros. (c) The characteristic roots of B' ' = (q q) are λ 1,..., λ p and q-p zeros. Proof. (a), (b) and (c) are applications of Lemma 10. 6.2 Matrix equation It will be shown in Section 7 that the problem turns out to be solving an integral equation, and to be solving a matrix equation. We here deal with the matrix equation in advance. The matrices appearing in this subsection are assumed to be transformed into a diagonalized form. Section 8 discusses more general cases where the matrices are transformed into a block diagonalized form. Now, let us consider two square matrices U = (m m) and V = (n n), and a rectangular matrix W = (m n). We assume that U and V can be diagonalized (real and symmetric, for example), and that the characteristic roots of U be λ 1,..., λ m (distinct roots) and those of V be v 1,..., v n (distinct roots). Then, we get the following proposition concerning a matrix equation where X = (m n) is an unknown matrix. Lemma 12 Concerning the equation (28), we haυe the propositions (a), (b), (c) and (d) below: (a) Under the assumption 3) P. J. Dhrymes(1978) : Mathematics for Econometrics, New York: Springer Verlag.