Handout #5 Computation of estimates, ANOVA Table and Orthogonal Designs

Similar documents
From Handout #1, the randomization model for a design with a simple block structure can be written as

Examples of non-orthogonal designs

236 Chapter 4 Applications of Derivatives

EXST Regression Techniques Page 1 SIMPLE LINEAR REGRESSION WITH MATRIX ALGEBRA

Properties of the least squares estimates

CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum

3 - Vector Spaces Definition vector space linear space u, v,

Modèles stochastiques II

[y i α βx i ] 2 (2) Q = i=1

Here are proofs for some of the results about diagonalization that were presented without proof in class.

Linear Regression Linear Regression with Shrinkage

T i t l e o f t h e w o r k : L a M a r e a Y o k o h a m a. A r t i s t : M a r i a n o P e n s o t t i ( P l a y w r i g h t, D i r e c t o r )

Lecture Notes 1: Vector spaces

This property turns out to be a general property of eigenvectors of a symmetric A that correspond to distinct eigenvalues as we shall see later.

Lecture VIII Dim. Reduction (I)

Linear Regression Linear Regression with Shrinkage

Ð"Ñ + Ð"Ñ, Ð"Ñ +, +, + +, +,,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Product Measures and Fubini's Theorem

EXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"

y(x) = x w + ε(x), (1)

1 :: Mathematical notation

Large Sample Properties of Estimators in the Classical Linear Regression Model

ETIKA V PROFESII PSYCHOLÓGA

Weighted Least Squares

STAT 540: Data Analysis and Regression

EXCERPTS FROM ACTEX CALCULUS REVIEW MANUAL

Weighted Least Squares

Economics 620, Lecture 4: The K-Varable Linear Model I

These notes give a quick summary of the part of the theory of autonomous ordinary differential equations relevant to modeling zombie epidemics.

Economics 620, Lecture 4: The K-Variable Linear Model I. y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N

Théorie Analytique des Probabilités

Bias Correction in the Balanced-half-sample Method if the Number of Sampled Units in Some Strata Is Odd

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

4.7 Confidence and Prediction Intervals

Lecture 8 January 30, 2014

Chemometrics. Matti Hotokka Physical chemistry Åbo Akademi University

Suggestions - Problem Set (a) Show the discriminant condition (1) takes the form. ln ln, # # R R

Applied Econometrics (QEM)

SIMULTANEOUS CONFIDENCE BANDS FOR THE PTH PERCENTILE AND THE MEAN LIFETIME IN EXPONENTIAL AND WEIBULL REGRESSION MODELS. Ping Sa and S.J.

STAT 100C: Linear models

The general linear model (and PROC GLM)

Lecture 34: Properties of the LSE

One-way ANOVA. Experimental Design. One-way ANOVA

Lecture 6: Linear models and Gauss-Markov theorem

X n = c n + c n,k Y k, (4.2)

14 Multiple Linear Regression

MMSE Equalizer Design

Example: A Markov Process

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

It's Only Fitting. Fitting model to data parameterizing model estimating unknown parameters in the model

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

On Least Squares Linear Regression Without Second Moment

Appendix A: Matrices

STAT 350: Geometry of Least Squares

Applied Econometrics (QEM)

Math 249B. Geometric Bruhat decomposition

2 Two-Point Boundary Value Problems

Economics 620, Lecture 5: exp

Estimation of the Response Mean. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 27

Multivariate Regression

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Orthogonal Projection and Least Squares Prof. Philip Pennance 1 -Version: December 12, 2016

Homework Set 2 Solutions

Fitting Linear Statistical Models to Data by Least Squares II: Weighted

Next is material on matrix rank. Please see the handout

be a deterministic function that satisfies x( t) dt. Then its Fourier

QUEEN MARY, UNIVERSITY OF LONDON

Introduction to Machine Learning

Regression coefficients may even have a different sign from the expected.

INTRODUCTORY ECONOMETRICS

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

ECE 275A Homework 6 Solutions

Multiple Linear Regression

Math 131 Exam 4 (Final Exam) F04M

18.S096 Problem Set 3 Fall 2013 Regression Analysis Due Date: 10/8/2013

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Analysis of variance using orthogonal projections

Lecture 4. Random Effects in Completely Randomized Design

Simple Linear Regression

Improving AOR Method for a Class of Two-by-Two Linear Systems

Linear Methods for Regression. Lijun Zhang

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Part I consists of 14 multiple choice questions (worth 5 points each) and 5 true/false question (worth 1 point each), for a total of 75 points.

SUPPORTING INFORMATION. Line Roughness. in Lamellae-Forming Block Copolymer Films

13. The Cochran-Satterthwaite Approximation for Linear Combinations of Mean Squares

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Homework 6 Solutions

{η : η=linear combination of 1, Z 1,, Z n

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Possible numbers of ones in 0 1 matrices with a given rank

1 Cricket chirps: an example

Framework for functional tree simulation applied to 'golden delicious' apple trees

Lecture 11: Regression Methods I (Linear Regression)

CSC Metric Embeddings Lecture 8: Sparsest Cut and Embedding to

Regression With a Categorical Independent Variable

Xβ is a linear combination of the columns of X: Copyright c 2010 Dan Nettleton (Iowa State University) Statistics / 25 X =

Lecture 11: Regression Methods I (Linear Regression)

Regression Models - Introduction

STAT5044: Regression and Anova. Inyoung Kim

Transcription:

Hout 5 Computation of estimates, ANOVA Table Orthogonal Designs In this hout e ill derive, for an arbitrary design, estimates of treatment effects in each stratum. Then the results are applied to completely romized designs, romized complete block designs Latin squares. 1. Computation of estimates in a given stratum Project the romization model (4.1) (4.2) in Hout 4 onto f, 1, e obtain the folloing linear model: ith WCœ W\ α W%, (5.1) E ÐW % œ 0 cov ÐW% œ 0 W, (5.2) here (5.2) follos from cov ÐW% œ WZ W œ WÐ 4œ! 04W4W œ 0W (since Á 4 Ê WW 4 œ!. Our goal is to compute the best linear unbiased estimator of any estimable treatment function - α in f, i.e., the estimator of - α ith minimum variance among unbiased estimators of the form +WC. When restricted to the vectors in f, the covariance matrix 0W is the same as 0M: Ð0W @ œ 0@ for all @ f. Therefore the Gauss-Markov Theorem applies, the least squares estimators are the best linear unbiased estimators. From (.) (.4) in Hout, e have Lemma 5.1 A treatment function - α is estimable in stratum f (i.e., b an unbiased estimator of the form +WC ) if only if - eð\ W\. If -α is estimable in stratum f, then its best linear unbiased estimator in f is - α^, here α^ is any solution of the normal equation \ W \ α^ œ \ WC, (5.) var Ð ^ - α œ 0 - Ð\ W \ -. Proof. The design matrix in model (5.1) is \ œ W \ ; so e have Ð\ \ œ ÐW\ W\ œ \ W\, the right-h side of the normal equation is Ð\ WC œ ÐW\ WC œ \ WC. The matrix \W\ is called the information matrix for treatment effects in stratum f. It has the folloing properties: = Lemma 5.2. The information matrix \ W \ is symmetric, nonnegative definite has zero ro sums. Proof. Symmetry nonnegative-definiteness are straightforard. To sho that it has zero ro sums, e note that by the definition of \, each of its ros has exactly one 1, all the other entries are 0's; therefore \ 1 œ 1, \ W \ 1 œ \ W 1 œ 0. R R -1-

For convenience, let G œ \ W \ (5.4) U œ \ W C. (5.5) Then (5.) becomes G α^ œ U. (5.6) Lemma 5.2 shos that rankðg Ÿ 1, that if - is estimable, then it must be a contrast. If rankðg œ 1, then e say that the design is connected in stratum f. In this case, all the treatment contrasts are estimable in. f α 2. ANOVA in a given stratum Since E ÐC œ. 1 \ α eð\ œg, under (5.1), EÐWCœWE ÐC W( g ). Let TW Ðg WC be the orthogonal projection of the data vector WC onto W( g ). Then, since W( g ) is the range of W\, from Hout, TW Ðg WC œöðw\ ÒÐW\ ÐW\ Ó ÐW\ WC œ ÐW\ Ð\ W\ \ WC. So T W C W Ðg œòðw \ Ð\ W \ \ WCÓÐW \ Ð\ W \ \ WC The residual is T f W Ðg œ CW\ Ð\ W\ \ WC (5.7) ^α œ ( ) U. (5.8) œ ( ^ α ) G α^ (5.9) W C, e have lwcl œ T WC T WC. (5.10) W Ðg f W Ðg Formula (5.10) gives the ANOVA in stratum f. The first term is the treatment sum of squares the second term is the residual sum of squares. No, EÐT W C f W Ðg œ EÐCWT WC f W Ðg œòe ÐC) ÓW T W EÐC trðw T W Z f W Ðg f W Ðg œ tr( WT W WZ ) [Since WE ÐC W( g, hich is orthogonal to f WÐg ] f Ðg -2-

œ tr( 0 WT W) (Since Z œ 0 W) f WÐg 4 4 4œ! œ tr Ð0 T (Since f W ( g ) f, e have T W œ P ) f W Ðg f W Ðg f W Ðg œ 0 Ödim( f ) dim ÒW Ðg Ó. EÐT Similarly, W C W Ðg W Ðg W Ðg œðt EÐW C trðw T W Z œ α \ W\ Ð\ W\ \ W\ α tr Ð0TW Ðg [In (5.7), replace WC ith E ÐWC, since Ðg f ] œ α Ð\ W \ α 0 dim[ W Ðg ] œ α G α 0 dim[ W Ðg ]. S = Summarizing the above discussion, e have the folloing ANOVA ithin stratum f : Sources Sums of Squares d.f. MS E(MS) Treatments ( α^ ) α^ dim[ ( )] ( α^ ) α^ G W g G 0 α G α.7òw Ðg Ó.7ÒW Ðg Ó Rresidual By subtraction 0 Total lwcl dim( f ) Note that rankðg œ rank Ð\ W \ ) œ dim[ W ( g )]. If the design is connected in f then the treatment sum of squares has 1 degrees of freedom in f.. Orthogonal designs If g Z f for some, then e say that e have an orthogonal design. In this case, for all 4Á, W4( g ) œ{ 0}, dimòw4ðg Óœ0, the treatment contrasts can only be estimated in f. We therefore drop the superscript from α^. In the folloing, e shall assume that < 4 0 for all 4 œ 1, á,, here < 4 is the number of replications of the 4 th treatment. Then dim( g Z) œ, hence all the treatment contrasts are estimable in. Furthermore, f W ( g ) œ W [ Z Š ( g Z)] œ W ( g Z) œ g Z, (5.11) --

for each B, W B œ W ÐKB W ÐT B œ W ÐT B œ T B. (5.12) g Z Z Z The last equality holds because TZ B g Z g Z f. It follos from (5.12) that W\ œ T Z \, hence G œ \ W \ œ \ T Z \ (5.1) Z U œ \ WCœ ÐW\ Cœ ÐT \ C œ \ T C. (5.14) Z Consider the folloing one-ay layout model Cœ. \ α %, E Ð% œ 0 cov Ð% œ 5 M. ~ ~ ~ ~ By (.8) in Hout, estimates of α are solutions of \\ α^ œ \C ~, here \ œ T Z \, ~ ~ ~ ~. Then ~ C œ T Z C \ \ \ C are the same as G U in (5.1) (5.14), respectively. Therefore hen g Z f, the best linear unbiased estimator - α^ of a treatment contrast is the same as under the above one-ay layout model, i.e., _ - α^ œ -4 4, (5.&) _ here 4 is the average of all the observations on the 4th treatment. In the folloing, e sho that By Lemma 5.1, e have var( - α^ ) œ 0 4 <. (5.16) 4 - - œ \ W \? for some? (5.17) var( ^ - α) œ 0 - (\ W \ ) -. (5.18) By the same argument as in the proof of Lemma 5.2, W\ K? œ 0, because \ K? has constant entries. It follos that - œ \ W \ Ð? K?. So by replacing? in (5.17) ith? K?, e may assume? Z. Then \? 1 œ 0, i.e., \? Z. Since \? g, e have \? g Z. Recall that g Z f. Therefore \? f, hence W \? œ \?. Then - œ \ W \? œ \ \? œ??, here? is the diagonal matrix hose 4 th diagonal entry is < 4. Hence From (5.18), var( ^ - α) œ 0 - (\ W \ ) -? œ? -. (5.19) œ 0? (\ W \ )(\ W \ ) - [by (5.17)] -4-

œ 0?- [ CœE Dis a solution of ECœ Dif a solution exists] œ 0 (-? -), [by (5.19)] hich is (5.16). Also, the treatment sum of squares W Ðg W Ðg W Ðg W Ðg T W C œ T C [Since W ( g ) f, T W œ T ] œ lt g Z Cl [By (5.11)] œ lð KC l _ R œ < [ ( C )]. 4 4 R 6 6œ The above discussion shos that if estimators simple ANOVA: g Z f for some, then e have simple variation Sums of Squares degrees of freedom Mean square E(MS).7Ð f lwcl dim( f ) lwcl 0 ã ã ã ã ã f f _ R _ Treatments [ ( )] 1 [ < C ã 0 < ( α α) ] 4 4 R 6 4 4 6œ Residual lwcl [ _ < 4 4 R R Ð C6Ó dim( f) 1 ã 0 6œ ã ã ã ã ã = = =.7Ð = = f lwcl dim( f ) lwcl 0 Total lc KCl R 1 f = -5-

_ In the above table, α œ < α. R 4 4 No e specialize these results to three simple orthogonal designs: completely romized designs, romized complete block designs, Latin squares. In a completely romized design, the to strata are f! œ Z f œ Z. Obviously g Z Z ; so it is an orthogonal design. In a romized complete block design, each treatment appears in each block exactly once. The condition of proportional frequencies is satisfied by the treatment block factors. It follos from the Theorem in Hout 2 that ( g Z) ( U Z). Also g Z Z. Therefore g Z U ( œ f.) This establishes the orthogonality of a romized complete block design. In a Latin square (or more generally, romized ro-column designs in hich all the treatments appear equally often in each ro equally often in each column), the condition of proportional frequencies is satisfied by the treatment ro factors, also by the treatment column factors. Thus ( g Z) ( e Z) ( g Z) ( V Z). As a result, g Z ( e V) ( œ f$ ). Therefore for these designs, estimators of treatment contrasts their variances are given by (5.15) (5.16). Their ANOVA tables follo. ANOVA table for a completely romized design: variation Sums of Squares degrees of freedom Mean square E(MS) Treatments [ ] 1 [ < C. ã 0 < ( α α) ] 4 4 4 4 residual By subtraction 0 R. œ Total ( C C) R 1 ANOVA table for a romized complete block design: variation Sums of Squares d.f. Mean square E(MS),,...,... 0 œ œ Blocks C ( C ), 1 C ( C) _ Treatments [ ] 1 [, C ã 0,( α α) ] 4.. 4. Residual By subtraction 0, Total ( C C ), 1 œ 4.. -6-

ANOVA table for a Latin square design: variation Sums of Squares d.f. Mean square E(MS)...... 0 œ œ Ros C ( C ) 1 C ( C). 4... 4.. 0 Columns C ( C ) 1 C ( C) _ Treatments [ ] 1 [ C ã 0 ( α α) ] 6.. $ 6. 6œ 6œ Residual By subtraction 0 $ Total ( C C ) 1 œ 4.. -7-