STAT 501 Assignment 1 Name Spring Written Assignment: Due Monday, January 22, in class. Please write your answers on this assignment

Similar documents
STAT 501 Assignment 1 Name Spring 2005

EXAM # 3 PLEASE SHOW ALL WORK!

T i t l e o f t h e w o r k : L a M a r e a Y o k o h a m a. A r t i s t : M a r i a n o P e n s o t t i ( P l a y w r i g h t, D i r e c t o r )

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Exam 2. Jeremy Morris. March 23, 2006

STAT 501 Assignment 2 NAME Spring Chapter 5, and Sections in Johnson & Wichern.

STAT 501 EXAM I NAME Spring 1999

Stat 206: Sampling theory, sample moments, mahalanobis

Random Vectors. 1 Joint distribution of a random vector. 1 Joint distribution of a random vector

MATH 1553 SAMPLE FINAL EXAM, SPRING 2018

Announcements (repeat) Principal Components Analysis

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Basic Concepts in Matrix Algebra

Def. The euclidian distance between two points x = (x 1,...,x p ) t and y = (y 1,...,y p ) t in the p-dimensional space R p is defined as

Whitening and Coloring Transformations for Multivariate Gaussian Data. A Slecture for ECE 662 by Maliha Hossain

Review (Probability & Linear Algebra)

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Next is material on matrix rank. Please see the handout

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Exercise Sheet 1.

Lecture Notes 1: Vector spaces

Multiple Random Variables

Stat 5101 Lecture Notes

Principal Components Theory Notes

Sample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson

Lecture 21: Convergence of transformations and generating a random variable

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

Math 180B, Winter Notes on covariance and the bivariate normal distribution

2. Matrix Algebra and Random Vectors

Multiple Integrals and Probability Notes for Math 2605

Math 164-1: Optimization Instructor: Alpár R. Mészáros

which has a check digit of 9. This is consistent with the first nine digits of the ISBN, since

Lecture 21 Theory of the Lasso II

Multivariate Random Variable

6-1. Canonical Correlation Analysis

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

8 Eigenvectors and the Anisotropic Multivariate Gaussian Distribution

Minimum Error Rate Classification

Chapter 5. The multivariate normal distribution. Probability Theory. Linear transformations. The mean vector and the covariance matrix

The Multivariate Normal Distribution

Inner product spaces. Layers of structure:

The Multivariate Gaussian Distribution [DRAFT]

Linear Analysis Lecture 5

Bivariate distributions

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

Math 180B Problem Set 3

STAT 7032 Probability Spring Wlodek Bryc

MAS223 Statistical Inference and Modelling Exercises

Linear Algebra Massoud Malek

Euclidean Space. This is a brief review of some basic concepts that I hope will already be familiar to you.

A Peak to the World of Multivariate Statistical Analysis

ECON 4117/5111 Mathematical Economics Fall 2005

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

Lecture Note 1: Probability Theory and Statistics

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

Functional Analysis Exercise Class

BREVET BLANC N 1 JANUARY 2012

Applied Linear Algebra in Geoscience Using MATLAB

Homework sheet 4: EIGENVALUES AND EIGENVECTORS. DIAGONALIZATION (with solutions) Year ? Why or why not? 6 9

ECON 5111 Mathematical Economics

Chapter 5 continued. Chapter 5 sections

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

STAT 414: Introduction to Probability Theory

Immerse Metric Space Homework

Definitions and Properties of R N

Concentration Ellipsoids

Notes on Linear Algebra and Matrix Theory

More than one variable

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

STAT 418: Probability and Stochastic Processes

Algebra 1 Correlation of the ALEKS course Algebra 1 to the Washington Algebra 1 Standards

Common-Knowledge / Cheat Sheet

01 Probability Theory and Statistics Review

Ch. 5 Joint Probability Distributions and Random Samples

Lecture 11. Multivariate Normal theory

Designing Information Devices and Systems I Fall 2018 Lecture Notes Note 21

Lecture 14: Multivariate mgf s and chf s

Lecture 2: Linear Algebra Review

Lecture 12: Diagonalization

Multivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013

TAMS39 Lecture 2 Multivariate normal distribution

Math 302 Outcome Statements Winter 2013

Geometry and Motion, MA 134 Week 1

STAT 7032 Probability. Wlodek Bryc

Archive of past papers, solutions and homeworks for. MATH 224, Linear Algebra 2, Spring 2013, Laurence Barker

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Math 164-1: Optimization Instructor: Alpár R. Mészáros

Vectors and Matrices Statistics with Vectors and Matrices

Introduction and Preliminaries

Probability Lecture III (August, 2006)

Linear Algebra Review

STT 843 Key to Homework 1 Spring 2018

MATH 54 - FINAL EXAM STUDY GUIDE

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

Lecture 3: Review of Linear Algebra

Business Statistics. Lecture 10: Correlation and Linear Regression

Lecture 3: Review of Linear Algebra

Transcription:

STAT 5 Assignment Name Spring Reading Assignment: Johnson and Wichern, Chapter, Sections.5 and.6, Chapter, and Chapter. Review matrix operations in Chapter and Supplement A. Examine the matrix properties presented in exercises.,.,.,. and. at the end of Chapter. Written Assignment: Due Monday, January, in class. Please write your answers on this assignment. Consider a random vector (,, ) with mean vector µ (,, ) and covariance matrix Ó. The eigenvalues of Ó are ë 5, ë 9, ë 5 and the corresponding eigenvectors are ( 5 ) / ( ) ( ) 5 / / 5 e e 5 / / 5 5/ / e / Evaluate the following quantities. Parts (a), (b), (c), (d) and (g) can be easily done without using either a computer or a calculator. (a) Ó (b) trace(ó) (c) V(e ) (d) (e)

(f) / ' ' (g) Define Y e, Y e, and Y e. Find the mean vector and covariance matrix for Y (Y Y Y )' à ', where the i-th column of à is the i-th eigenvector for Ó. Evaluate ' (Y) E (Y) V Γ ' Γ (h) The linear transformation, Y Γ, examined in part (g) is often called a rotation because it corresponds to simply rotating the coordinate axes. Use the definition of eigenvectors to show that lengths of vectors are not changed by this transformation, i.e., show Y ' Y ' when Y Γ. (i) Let à be defined as in part (g), and let Y Γ, µ E( ), V (), µ Y Γ ' µ and Y V(Y) Γ ' Γ. Determine which of the following measures of variability or distance are unaffected by rotations. Justify your answers. (a) Is Ó Ó Y? (b) Is trace (Ó ) trace (Ó Y )? (c) Is Ó Ó Y?

(d) Is ( µ )' ( µ ) (Y µ )' (Y )? Y Y µ Y (e) Is ( µ )'( µ ) (Y µ )' (Y )? Y µ Y. Consider a bivariate normal population with mean vector where µ and 9 µ and covariance matrix Ó, 5 (a) Write down a formula for the joint density function. (b) Find the eigenvalues and eigenvectors for Ó. λ λ e e (c) Find values for Ó trace (Ó) (d) Write down the formula for the boundary of the smallest region such that there is probability.5 that a randomly selected observation will be inside the boundary. (e) Sketch the boundary of the region in part (d). - - - - 5 6 -

- (f) Determine the area of the region described in parts (d) and (e). (g) Determine the area of the smallest region such that there is probability.95 that a randomly selected observation will be in the region.. Let be a normally distributed random vector with µ and (a) Indicate which of the following are pairs of independent random variables by drawing a circle around the appropriate pairs. (i) and (iv) and + (ii) and (v) + and - (iii) and - (vi) + + and - + (b) What is the distribution of Y (, )? (c) What is the conditional distribution of Y (, ) given x? (d) Find the correlation between and and a formula for the partial correlation between and given x. ρ ρ (e) What is the conditional distribution of given x and x?

5. Suppose, N µ where 9 µ (a) What is the distribution of Z +? (b) What is the joint distribution of Z in part (a) and Z +. (c) Find the conditional distribution of given x, x. (d) Find the partial correlation between and given x and x.. ρ 5. Let be Y and, N µ be, N µ where and Y are independent and 8 6 µ µ (a) Find + Y Y, Cov

6 (b) Are Y and + Y independent random vectors? Explain. (c) Show that the joint distribution for the four dimensional random vector + Y Y is a multivariate normal distribution. 6. Let d(p,q) denote a measure of distance between p and q. Johnson and Wichern indicate that any distance measure should possess the following four properties: (i) d(p, q) d (q, p) (ii) d(p, q) > if (iii) d(p, q) if p p q q (iv) d(p, q) d(p, r) + d(r,q), where r ( r,r )'. These properties are satisfied, for example, by Euclidean distance, i.e.,. d(p, q) / [ (p q)' (p q)] (a) In this class we will often consider distance measures of the form / d(p, q) [ (p q)' A (p q)]. Sometimes A will be the inverse of a covariance matrix which implies that A is symmetric and positive definite. Show that properties (i) through (iv) are satisfied when A is symmetric and positive definite.

7 (b) Does A have to be both symmetric and positive definite for / d(p, q) [ (p q)' A (p q)] to satisfy properties (i) through (iv)? Present your proofs or explanations. (c) It is obvious that the measure d(p,q) max{ p q, p q,..., p q } satisfies properties (i), (ii), (iii). Either prove or disprove that it satisfies property (iv), the triangle in equality. k k 7. One frequently used technique in multivariate analysis first reduces a multivariate problem to a more familiar univariate problem by considering linear combinations of the elements of a random vector, and then defines the "optimal" linear combination as a multivariate measure. For example, consider a p-dimensional random vector (,,..., p) ' with mean vector µ ( µ, µ,..., µ p) ' and covariance matrix Ó. Using a vector of constants, a (a, a,..., ap) ', define a random variable

Y a' a + a +... + a. Then the mean and variance of Y are p p a' µ and a' a, respectively, and the standardized distance of Y from its mean is 8 d (a) / Y a' µ / a' a. Define D² maximum {d²( a ): ' a a and p a ε R }. as the maximum squared "standardized distance" of from µ. (a) Use the extended Cauchy-Schwarz inequality in Section.7 of Johnson and Wichern to show a c ( µ ) provides the maximum of d²( a ) in the definition of D², where c is a normalizing constant selected to make a '. a (b) Do you recognize D² as a measure presented in the lectures for STAT 5? If so, what is its name? (Note that we did not assume that has a multivariate normal distribution.) For additional practice you could do problems.6,.8,.9,.,.,.6 at the end of Chapter and problems.,.,.,.5,.,.5,.6,.7 at the end of Chapter. Problem.8 gives a trivial example of a non-normal bivariate distribution with normal marginal distributions. Do not hand in these additional problems, but answers will be given on the answer sheet for this assignment.