COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure)

Size: px
Start display at page:

Download "COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure)"

Transcription

1 STAT 52 Completely Randomized Designs COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure) Completely randomized means no restrictions on the randomization of units to treatments (except for sample sizes ordinarily fixed in advance), so with N available units: Randomly select n units for assignment to trt Randomly select n 2 remaining units for assignment to trt 2... Last n t units are assigned to trt t Or, any other procedure s.t. all N!/(n!n 2!...n t!) possible assignment patterns have equal probability Ordinarily reflects model and assumptions for -way ANOVA

2 STAT 52 Completely Randomized Designs 2 Cell Means model: y i,j = µ i + ɛ i,j i =...t, j =...n i, t i= n i = N µ i represents E[y i,j ], an expected response corresponding to treatment i, in the laboratory in which the experiment was performed on the day the experiment was run... E[ɛ i,j ] = 0, V ar[ɛ i,j ] = σ 2, independent, perhaps normal ɛ represents (in part) unit-to-unit variation; indep is justified (in part) by complete randomization Matrix representation: y = Xµ + ɛ [N ] = [N t][t ] + [N ]

3 STAT 52 Completely Randomized Designs 3 y,... y,n y 2,... y 2,n2... y t,... y t,nt = µ µ 2... µ t + ɛ,... ɛ,n ɛ 2,... ɛ 2,n2... ɛ t,... ɛ t,nt

4 STAT 52 Completely Randomized Designs 4 Effects model: y i,j = α + τ i + ɛ i,j This says the same thing mathematically... Let α = anything, then τ i = µ i α α now conceptually represents everything in common to all experimental runs, τ i is the deviation from this due only to treatment i Matrix representation: y = Xθ + ɛ, θ = (α, τ, τ 2,..., τ t ) [N ] = [N (t + )][(t + ) ] + [N ] X is a different matrix now... will often use X generically Less-than-full-rank

5 STAT 52 Completely Randomized Designs 5 y,... y,n y 2,... y 2,n2... y t,... y t,nt = α τ τ 2... τ t + ɛ,... ɛ,n ɛ 2,... ɛ 2,n2... ɛ t,... ɛ t,nt

6 STAT 52 Completely Randomized Designs 6 Estimation: a review y = Xθ + ɛ... (general X, iid errors) (Xθ is the expectation of y... the noiseless model ) MLE for normal errors, LS in any case, is any sol n to: X Xˆθ = X y Normal equations, as in norm, not normal dist n X of full column rank: ˆθ = (X X) X y Otherwise: ˆθ = (X X) X y generalized inverse, non-unique AA A = A Need to review statistical properties of ˆθ

7 STAT 52 Completely Randomized Designs 7 Expectations: Recall that for random y with E[y] = m, V ar[y] = V vector of means m, matrix of variances and covariances V E[Ay] = Am V ar[ay] = AVA (generalization of constant-squared in the scalar case) So, E[ˆθ] = (X X) X (Xθ)... generally E[ˆθ] = θ... full rank

8 STAT 52 Completely Randomized Designs 8 In the less-than-full-rank case, ˆθ is not uniquely determined, but consider estimable functions: c θ, where c = l X for some N-vector l ĉ θ = c ˆθ = (l X)(X X) X y X(X X) X = H is invariant to choice of g-inverse projection matrix, into space spanned by columns of X symmetric N N, idempotent, H for hat because ŷ = Xˆθ = Hy c ˆθ = l Xˆθ = l Hy E[c ˆθ] = l HXθ = l Xθ = c θ Likewise for C = c... c m, E[Cˆθ] = Cθ if C = LX

9 STAT 52 Completely Randomized Designs 9 CRD Examples: τ i isn t estimable µ i = α + τ i is estimable, but is a function of experiment as well as treatment µ i µ j = τ i τ j is estimable... comparative (see X, slide 5)

10 STAT 52 Completely Randomized Designs 0 Variances: Recall for estimable Cθ, Cˆθ = (LX)(X X) X y = LHy V ar[cˆθ] = LH(σ 2 I)H L = σ 2 LHL = σ 2 (LX)(X X) (X L ) = σ 2 C(X X) C... generally V ar[cˆθ] = σ 2 C(X X) C... full rank in particular, V ar[ˆθ] = σ 2 (X X)... full rank

11 STAT 52 Completely Randomized Designs Optimal allocation: Interest in estimating Cθ well, i.e. with minimum variance Allocate N units to treatments to minimize the average contrast variance min ni,i=...t ave j σ 2 c j (X X) c j, subject to t i= n i = N or g-inv if X less than full rank and Cθ estimable c j, the jth row in C σ 2 is a positive constant, doesn t matter min ni,i=...t trace C(X X) C, subject to t i= n i = N min ni,i=...t trace C C(X X), subject to t i= n i = N linear optimality, generalization of A-optimality If X is not of full rank but Cθ is estimable, same argument applies to (X X)

12 STAT 52 Completely Randomized Designs 2 Example: Full-rank cell means model, θ = µ (X X) = diag(/n, /n 2,.../n t ) Interest in estimating differences between the control (treatment ) and each of the other treatments C = t t

13 STAT 52 Completely Randomized Designs 3 C C = t (dimension t t) trace C C(X X) = (t )/n + t i=2 /n i want to: minimize, w.r.t. n, n 2,...n t, subject to: (t )/n + t i=2 /n i t i= n i = N

14 STAT 52 Completely Randomized Designs 4 Constrained Minimization by Method of Lagrangian Multipliers: φ = [(t )/n + t i=2 /n i] + λ( t i= n i N) (first term is what you want to minimize, second is what is constrained to be zero) Set each to zero and solve: φ/ n = (t )/n 2 + λ φ/ n i = /n 2 i + λ i = 2...t φ/ λ = t i= n i N n = (t )/λ n i = /λ i = 2...t n = t N t +(t ) n i = N t +(t )

15 STAT 52 Completely Randomized Designs 5 Note: This is really a continuous optimization technique applied to a discrete problem. In practice, need to round solutions to integer values, or use as starting point in a discrete search Note also that Cov[ µ µ i, µ µ j ] = σ 2 /n relatively larger n reduces these... Postscript: intuition might say you need less info about control, since you have more experience with it... but this means you intend to USE that information... e.g. via a Bayesian prior

16 STAT 52 Completely Randomized Designs 6 Example: (less-than-full-rank version) Effects model, θ = (α τ... τ t ) N n n 2... n t n n X X = n 2 0 n n t n t (X X) = /n /n /n t

17 STAT 52 Completely Randomized Designs 7 General algorithm for g-inverse of symmetric matrices (e.g. Linear Models, Searle, Wiley) find a full-rank, principal minor of maximum dimension invert it fill in zeros everywhere else C = C C and this (X X) are as before, but with leading row/column of zero... so problem and solution are the same

18 STAT 52 Completely Randomized Designs 8 Hypothesis testing: a review Alternative (full) and Null (restricted) hypotheses: Hyp A : y = X A θ A + ɛ Hyp 0 : y = X 0 θ 0 + ɛ = X A Pθ 0 + ɛ Columns of X 0 are all linear combinations of columns of X A X A is N-by-p A, P is p A -by-p 0, p 0 < p A

19 STAT 52 Completely Randomized Designs 9 Test for treatment effects: cell means model X A =, P =, X 0 = Test for treatment effects: effects model X A =, P = or 0, X 0 =

20 STAT 52 Completely Randomized Designs 20 Test for additional terms : partitioned model Hyp 0 : y = X θ + ɛ Hyp A : y = X θ + X 2 θ 2 + ɛ This is a special case of the general form: X A = (X, X 2 ), P = I 0, X 0 = X A P = X But X 0 doesn t have to be a subset of columns from X A

21 STAT 52 Completely Randomized Designs 2 For either model: SSE = i j (y i,j ŷ i,j ) 2... sometimes resid = i j (y i,j x iˆθ) 2, where x i is row from X = i j (y i,j x i (X X) X y) 2 = squared length of y X(X X) X y = y (I H)(I H)y = y (I H H + H 2 )y = y (I H)y (key result for what s coming up) = y y (Hy) (Hy) = length 2 of y - length 2 of Hy

22 STAT 52 Completely Randomized Designs 22 Given truth of Hyp A (which means Hyp 0 MAY also be true), the validity of Hyp 0 is judged by how much worse the data fit the corresponding model: SST = SSE(Hyp 0 ) - SSE(Hyp A ) = y (I H 0 )y y (I H A )y = y H A y y H 0 y = y (H A H 0 )y Notation: Here and later, the subscript on H is the subscript on the corresponding X. Here it is 0 or A, later also or 2.

23 STAT 52 Completely Randomized Designs 23 Example: Full-rank (cell means) model Hyp 0 : µ = µ 2 =... = µ t, or y i,j = µ + ɛ i,j Hyp A : at least two differ, or y i,j = µ i + ɛ i,j X 0 = n n2..., H 0 = ( ) = N J N N X A = nt n... 0 n 0 n n , H A = n J n n... 0 n n t 0 n2 n... 0 n2 n t nt... nt 0 nt n... n t J nt n t

24 STAT 52 Completely Randomized Designs 24 Letting y i represent the n i -element vector of observations associated with treatment i, SST = y H A y y H 0 y = i y i ( n i J)y i y ( N J)y = i n i y 2 i. N y2.. = i n i(ȳ i. ) 2 N(ȳ.. ) 2 = i n i(ȳ i. ȳ.. ) 2 Example (cont.) Reduced-rank (effects) model, same problem Due to the invariance of the hat matrices, H 0 and H A are the same as before So, the resultant sums of squares are also the same

25 STAT 52 Completely Randomized Designs 25 Expectations: Recall that for random y with E[y] = m, V ar[y] = V E[y Ay] = m Am + trace(av) Suppose Hyp A really is true (perhaps also Hyp 0 ), so: E[y] = X A θ A E[SSE(Hyp A )] = θ AX A (I H A)X A θ A + trace((i H A )σ 2 I)) = θ AX A (X A X A )θ A + σ 2 (trace(i) trace(h A )) = 0 + σ 2 (N rank(x A )) (because for idempotent matrices, rank=trace, and rank(h) = rank(x))

26 STAT 52 Completely Randomized Designs 26 E[ SSE( Hyp 0 ) ] = θ AX A (I H 0)X A θ A + trace((i H 0 )σ 2 I)) = θ AX A (I H 0)X A θ A + σ 2 (N rank(x 0 )) = Q A 0 (θ A ) + σ 2 (N rank(x 0 )) (subscript A 0 comes from space(x A ) -space(x 0 )) For the partitioned model: = (θ X + θ 2X 2)(I H )(X θ + X 2 θ 2 ) + σ 2 (N rank(x )) = θ 2X 2(I H )X 2 θ 2 + σ 2 (N rank(x )) (X terms vanish... why?) = Q 2 (θ 2 ) + σ 2 (N rank(x )) (subscript 2 comes from space(x 2 ) -space(x )) E[ SST ] = Q 2 (θ 2 ) + (rank(x A ) rank(x 0 ))σ 2

27 STAT 52 Completely Randomized Designs 27 Example: all-treatments-equal hypothesis In either parameterization: rank(x A ) rank(x 0 ) = t H 0 = ( ) = N J Full rank: X A (I H 0)X A = X A X A N X A JX A Q A 0 (µ) Reduced rank: = diag(n) N nn, where n = (n, n 2,...n t ) = t i= n iµ 2 i N ( n i µ i ) 2 = t i= n i(µ i µ) 2, µ = N t i= n iµ i X 2(I H )X 2 = X A (I H 0)X A in full-rank case Q 2 (τ ) = t i= n i(τ i τ) 2, τ = N ni τ i

28 STAT 52 Completely Randomized Designs 28 Distribution theory: a review. y MV N(m, σ 2 I), A and B p.s.d. symmetric, Any two of: A and B idempotent A + B idempotent AB = 0 Implies all of: σ 2 y Ay χ 2 (rank(a), σ 2 m Am) σ 2 y By χ 2 (rank(b), σ 2 m Bm) the two quadratic forms are independent 2. For idempotent A, rank(a) = trace(a)

29 STAT 52 Completely Randomized Designs 29 Implications for CRD: σ 2 SSE(Hyp A ) = σ 2 y (I H A )y χ 2 (N t, 0) i.e. central σ 2 SST= σ 2 y (H A H 0 )y χ 2 (t, Q(τ )/σ 2 ) SST t / SSE (Hyp A ) N t F (t, N t, Q(τ )/σ 2 ) or, replace τ with µ in Q... same thing Other things being equal, larger Q greater power

30 STAT 52 Completely Randomized Designs 30 Example: CRD N = 20 t = 3 n = n 2 = 8, n 3 = 4 numerator df = 2 denominator df = 7 Suppose µ = µ 2 = 2 µ 3 = 3 σ =.5 Then µ =.8 Q( µ ) = 8(.8) 2 + 8(2.8) 2 + 4(3.8) 2 =.2 Q( µ )/σ 2 =

31 STAT 52 Completely Randomized Designs 3 R: qf(.95, 2, 7) th quantile of central F, i.e. critical value -pf(3.595, 2, 7, ) probability of value in critical region for non-central F SAS: proc iml; finv(.95, 2, 7); probf(3.595, 2, 7, );

32 STAT 52 Completely Randomized Designs 32 A practical point: May not know enough about problem to speculate about the (complete) true value of µ, but willing to guess: Bounds for Q (assume equal n i ) D = µ max µ min = τ max τ min Most favorable situation (Q greatest) greatest variance of µ s t/2 µ s = µ max t/2 µ s = µ min Q = n i t ( 2 D)2 = N D 2 /4 Least favorable situation (Q smallest) smallest variance of µ s µ = µ max µ = µ min others = (µ max + µ min )/2 Q = 2 n i ( 2 D)2 = n i D 2 /2

33 STAT 52 Completely Randomized Designs 33 Reduced normal equations for τ : y = Xθ + ɛ = X β + X 2 τ + ɛ β are nuisance parameters for our CRD effects model, X β = α Full normal equations: (X X)ˆθ = X y X X X X 2 ˆβ = X 2X X 2X 2 ˆτ X y X 2y X X ˆβ + X X 2 ˆτ = X y X 2X ˆβ + X 2 X 2 ˆτ = X 2y

34 STAT 52 Completely Randomized Designs 34 X X ˆβ + X X 2 ˆτ = X y X 2X ˆβ + X 2 X 2 ˆτ = X 2y Pre-multiply first equation by X 2X (X X ), to match leading st term of second equation will need g-inverse if X is of less than full rank X 2X ˆβ + X 2 H X 2 ˆτ = X 2H y Subtract this from the second equation to get reduced normal equations : X 2(I H )X 2 ˆτ = X 2(I H )y

35 STAT 52 Completely Randomized Designs 35 Pattern of reduced normal equations: Same form found in weighted regression... there I H is replaced with a weight matrix, any matrix proportional to the inverse of V ar(y) X 2 -model projected into null space of X Same as regression of resid(y : X ) on resid(x 2 : X ) Note reliance of form on X 2 and H X 2, or (I H )X 2...

36 STAT 52 Completely Randomized Designs 36 In fact, suppose we transform the independent variables in this problem, pretending the model isn t really partitioned, and that the single model matrix is: X 2 = (I H )X 2 then the usual normal equations are: X 2 X 2 ˆτ = X 2 y i.e. the correct result... we ve corrected for β by projecting X 2 out of the space spanned by X so, for example, c τ is estimable iff c = l X 2 V ar(ĉ τ ) = σ 2 c (X 2 X 2 ) c

37 STAT 52 Completely Randomized Designs 37 X 2 X 2 is called the design information matrix in the book, and sometimes denoted I 2. For CRD s: H = ( ) = N J N N X 2 = (I H )X 2 = ( n N ) n n 2 N n... n t N n n N n 2 ( n 2 N ) n 2... n t N n n N n t n 2 N n t... ( n t N ) n t I 2 = X 2 X 2 = diag(n) N nn, where n = (n, n 2,..., n t ) Note that X 2 and I 2 are each of rank t, reflecting the fact that τ isn t estimable

38 STAT 52 Completely Randomized Designs 38 So, for CRD s, given interest in τ and accounting for α, the precision of ĉ τ is determined by σ 2 c I 2 c one g-inverse of I 2 = diag(n) N nn is I 2 = diag (n)... check it the power of the test for equality of τ s is determined by τ I 2 τ /σ 2

39 STAT 52 Completely Randomized Designs 39 A brief note on nonlinear models: How do nonlinear models change things? Consider a very simple example: t = 2, and for the two treatments respectively: y = exp(α + τ ) + ɛ y = exp(α + τ 2 ) + ɛ i.e. the effects model we ve used, but wrapping the fixed part of the model in exp. As in our Optimal Allocation discussion, consider: Designs for which the contrast τ τ 2 is to be estimated as precisely as possible, and Constrained design for which n + n 2 = N (given) The linear algebra framework we ve been using isn t entirely adequate, so we need to talk about variances in more fundamental terms:

40 STAT 52 Completely Randomized Designs 40 For normal i.i.d. ɛ, the likelihood function for this problem is: L = n i= c σexp[ (y i exp(α + τ )) 2 /2σ 2 ] n2 j= c σexp[ (y j exp(α + τ 2 )) 2 /2σ 2 ] Given the necessary regularity conditions, the Fisher Information Matrix for efficient inference concerning θ = (α, τ, τ 2 ) is: E 2 θ 2 lnl = ( nexp 2 (α + τ ) + n 2 exp 2 (α + τ 2 ) n exp 2 (α + τ ) n 2 exp 2 (α + τ 2 ) σ 2 n exp 2 (α + τ ) n exp 2 (α + τ ) 0 n 2 exp 2 (α + τ 2 ) 0 n 2 exp 2 (α + τ 2 ) For estimable functions, asymptotic variances can be computed from a g-inverse of this, e.g V = σ 2 0 n exp 2 (α+τ ) n 2 exp 2 (α+τ 2 ) )

41 STAT 52 Completely Randomized Designs 4 As we would do in the linear models setting, we can compute a variance for the MLE of the estimable contrast τ τ 2 as c Vc, where c = (0, +, ) Using Lagrangian multipliers, you can show that the optimal allocation for any N is n n 2 = exp 2 3 (τ 2 τ ) The trouble is, this is a function of θ... i.e., the optimal design (or the performance of any design) depends on the unknown parameters you set out to investigate! The beauty of linear models, from the design perspective, is that the parameter values don t matter... the parameters disappear as a result of how differentiation treats the linear form in Fisher s Information. How to do Optimal Allocation in a nonlinear setting?

42 STAT 52 Completely Randomized Designs 42 Two popular practical approaches:. Bayesian Design (which, it can be argued, really isn t Bayesian, but is often helpful): Specify a prior distribution for the parameter vector, p(θ) Find the design that minimizes E θ c Vc This tends to work well when the there is some reasonable prior knowledge of the parameters, i.e. when p(θ) is not too diffuse. 2. Sequential Experimentation: Begin with a reasonable first-stage design and iterate: Compute ˆθ based on the data collected so far Compute the next-stage design to minimize c Vc as if θ = ˆθ For some problems, this can be shown to converge to the known-parameters optimal design, but this convergence may be slow unless the first-stage design is actually reasonable

Matrix Approach to Simple Linear Regression: An Overview

Matrix Approach to Simple Linear Regression: An Overview Matrix Approach to Simple Linear Regression: An Overview Aspects of matrices that you should know: Definition of a matrix Addition/subtraction/multiplication of matrices Symmetric/diagonal/identity matrix

More information

STAT5044: Regression and Anova. Inyoung Kim

STAT5044: Regression and Anova. Inyoung Kim STAT5044: Regression and Anova Inyoung Kim 2 / 51 Outline 1 Matrix Expression 2 Linear and quadratic forms 3 Properties of quadratic form 4 Properties of estimates 5 Distributional properties 3 / 51 Matrix

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Chapter 5 Matrix Approach to Simple Linear Regression

Chapter 5 Matrix Approach to Simple Linear Regression STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is

More information

STAT 540: Data Analysis and Regression

STAT 540: Data Analysis and Regression STAT 540: Data Analysis and Regression Wen Zhou http://www.stat.colostate.edu/~riczw/ Email: riczw@stat.colostate.edu Department of Statistics Colorado State University Fall 205 W. Zhou (Colorado State

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

STAT 730 Chapter 4: Estimation

STAT 730 Chapter 4: Estimation STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model Outline 1 Multiple Linear Regression (Estimation, Inference, Diagnostics and Remedial Measures) 2 Special Topics for Multiple Regression Extra Sums of Squares Standardized Version of the Multiple Regression

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Q = (Y i β 0 β 1 X i1 β 2 X i2 β p 1 X i.p 1 ) 2, which in matrix notation is Q = (Y Xβ) (Y

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

General Linear Model: Statistical Inference

General Linear Model: Statistical Inference Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter 4), least

More information

BIOS 2083 Linear Models c Abdus S. Wahed

BIOS 2083 Linear Models c Abdus S. Wahed Chapter 5 206 Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R

More information

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions Spatial inference I will start with a simple model, using species diversity data Strong spatial dependence, Î = 0.79 what is the mean diversity? How precise is our estimate? Sampling discussion: The 64

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

Linear Algebra Review

Linear Algebra Review Linear Algebra Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Linear Algebra Review 1 / 45 Definition of Matrix Rectangular array of elements arranged in rows and

More information

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38 Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters

More information

Linear Models and Estimation by Least Squares

Linear Models and Estimation by Least Squares Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:

More information

3 Multiple Linear Regression

3 Multiple Linear Regression 3 Multiple Linear Regression 3.1 The Model Essentially, all models are wrong, but some are useful. Quote by George E.P. Box. Models are supposed to be exact descriptions of the population, but that is

More information

3. The F Test for Comparing Reduced vs. Full Models. opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics / 43

3. The F Test for Comparing Reduced vs. Full Models. opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics / 43 3. The F Test for Comparing Reduced vs. Full Models opyright c 2018 Dan Nettleton (Iowa State University) 3. Statistics 510 1 / 43 Assume the Gauss-Markov Model with normal errors: y = Xβ + ɛ, ɛ N(0, σ

More information

Need for Several Predictor Variables

Need for Several Predictor Variables Multiple regression One of the most widely used tools in statistical analysis Matrix expressions for multiple regression are the same as for simple linear regression Need for Several Predictor Variables

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

STAT 8260 Theory of Linear Models Lecture Notes

STAT 8260 Theory of Linear Models Lecture Notes STAT 8260 Theory of Linear Models Lecture Notes Classical linear models are at the core of the field of statistics, and are probably the most commonly used set of statistical techniques in practice. For

More information

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form Topic 7 - Matrix Approach to Simple Linear Regression Review of Matrices Outline Regression model in matrix form - Fall 03 Calculations using matrices Topic 7 Matrix Collection of elements arranged in

More information

Stat 579: Generalized Linear Models and Extensions

Stat 579: Generalized Linear Models and Extensions Stat 579: Generalized Linear Models and Extensions Mixed models Yan Lu March, 2018, week 8 1 / 32 Restricted Maximum Likelihood (REML) REML: uses a likelihood function calculated from the transformed set

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y

More information

STAT 151A: Lab 1. 1 Logistics. 2 Reference. 3 Playing with R: graphics and lm() 4 Random vectors. Billy Fang. 2 September 2017

STAT 151A: Lab 1. 1 Logistics. 2 Reference. 3 Playing with R: graphics and lm() 4 Random vectors. Billy Fang. 2 September 2017 STAT 151A: Lab 1 Billy Fang 2 September 2017 1 Logistics Billy Fang (blfang@berkeley.edu) Office hours: Monday 9am-11am, Wednesday 10am-12pm, Evans 428 (room changes will be written on the chalkboard)

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Linear Models Review

Linear Models Review Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign

More information

Lecture 6: Linear models and Gauss-Markov theorem

Lecture 6: Linear models and Gauss-Markov theorem Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables

More information

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012 Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.

More information

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u Interval estimation and hypothesis tests So far our focus has been on estimation of the parameter vector β in the linear model y i = β 1 x 1i + β 2 x 2i +... + β K x Ki + u i = x iβ + u i for i = 1, 2,...,

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

Applied Statistics Preliminary Examination Theory of Linear Models August 2017

Applied Statistics Preliminary Examination Theory of Linear Models August 2017 Applied Statistics Preliminary Examination Theory of Linear Models August 2017 Instructions: Do all 3 Problems. Neither calculators nor electronic devices of any kind are allowed. Show all your work, clearly

More information

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Much of the material we will be covering for a while has to do with designing an experimental study that concerns some phenomenon of interest.

Much of the material we will be covering for a while has to do with designing an experimental study that concerns some phenomenon of interest. Experimental Design: Much of the material we will be covering for a while has to do with designing an experimental study that concerns some phenomenon of interest We wish to use our subjects in the best

More information

STAT 730 Chapter 5: Hypothesis Testing

STAT 730 Chapter 5: Hypothesis Testing STAT 730 Chapter 5: Hypothesis Testing Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 28 Likelihood ratio test def n: Data X depend on θ. The

More information

Multivariate Linear Regression Models

Multivariate Linear Regression Models Multivariate Linear Regression Models Regression analysis is used to predict the value of one or more responses from a set of predictors. It can also be used to estimate the linear association between

More information

Lecture 20: Linear model, the LSE, and UMVUE

Lecture 20: Linear model, the LSE, and UMVUE Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;

More information

Linear models and their mathematical foundations: Simple linear regression

Linear models and their mathematical foundations: Simple linear regression Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction

More information

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review

More information

BTRY 4830/6830: Quantitative Genomics and Genetics

BTRY 4830/6830: Quantitative Genomics and Genetics BTRY 4830/6830: Quantitative Genomics and Genetics Lecture 23: Alternative tests in GWAS / (Brief) Introduction to Bayesian Inference Jason Mezey jgm45@cornell.edu Nov. 13, 2014 (Th) 8:40-9:55 Announcements

More information

Lectures on Simple Linear Regression Stat 431, Summer 2012

Lectures on Simple Linear Regression Stat 431, Summer 2012 Lectures on Simple Linear Regression Stat 43, Summer 0 Hyunseung Kang July 6-8, 0 Last Updated: July 8, 0 :59PM Introduction Previously, we have been investigating various properties of the population

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra 1.1. Introduction SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that

More information

18.S096 Problem Set 3 Fall 2013 Regression Analysis Due Date: 10/8/2013

18.S096 Problem Set 3 Fall 2013 Regression Analysis Due Date: 10/8/2013 18.S096 Problem Set 3 Fall 013 Regression Analysis Due Date: 10/8/013 he Projection( Hat ) Matrix and Case Influence/Leverage Recall the setup for a linear regression model y = Xβ + ɛ where y and ɛ are

More information

DS-GA 1002 Lecture notes 10 November 23, Linear models

DS-GA 1002 Lecture notes 10 November 23, Linear models DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.

More information

10. Linear Models and Maximum Likelihood Estimation

10. Linear Models and Maximum Likelihood Estimation 10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that

More information

STA 2101/442 Assignment Four 1

STA 2101/442 Assignment Four 1 STA 2101/442 Assignment Four 1 One version of the general linear model with fixed effects is y = Xβ + ɛ, where X is an n p matrix of known constants with n > p and the columns of X linearly independent.

More information

Inference in Regression Analysis

Inference in Regression Analysis Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

Bayesian Inference. Chapter 9. Linear models and regression

Bayesian Inference. Chapter 9. Linear models and regression Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices

Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices Lecture 3: Simple Linear Regression in Matrix Format To move beyond simple regression we need to use matrix algebra We ll start by re-expressing simple linear regression in matrix form Linear algebra is

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

Lecture 1: Linear Models and Applications

Lecture 1: Linear Models and Applications Lecture 1: Linear Models and Applications Claudia Czado TU München c (Claudia Czado, TU Munich) ZFS/IMS Göttingen 2004 0 Overview Introduction to linear models Exploratory data analysis (EDA) Estimation

More information

IV. Matrix Approximation using Least-Squares

IV. Matrix Approximation using Least-Squares IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that

More information

Ch 3: Multiple Linear Regression

Ch 3: Multiple Linear Regression Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

3 (Maths) Linear Algebra

3 (Maths) Linear Algebra 3 (Maths) Linear Algebra References: Simon and Blume, chapters 6 to 11, 16 and 23; Pemberton and Rau, chapters 11 to 13 and 25; Sundaram, sections 1.3 and 1.5. The methods and concepts of linear algebra

More information

2.7 Estimation with linear Restriction

2.7 Estimation with linear Restriction Proof (Method 1: show that that a C(W T ), which implies that the GLSE is an estimable function under the old model is also an estimable function under the new model; secnd show that E[a T ˆβ G ] = a T

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

A Bayesian Treatment of Linear Gaussian Regression

A Bayesian Treatment of Linear Gaussian Regression A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,

More information

THE ANOVA APPROACH TO THE ANALYSIS OF LINEAR MIXED EFFECTS MODELS

THE ANOVA APPROACH TO THE ANALYSIS OF LINEAR MIXED EFFECTS MODELS THE ANOVA APPROACH TO THE ANALYSIS OF LINEAR MIXED EFFECTS MODELS We begin with a relatively simple special case. Suppose y ijk = µ + τ i + u ij + e ijk, (i = 1,..., t; j = 1,..., n; k = 1,..., m) β =

More information

Topic 17 - Single Factor Analysis of Variance. Outline. One-way ANOVA. The Data / Notation. One way ANOVA Cell means model Factor effects model

Topic 17 - Single Factor Analysis of Variance. Outline. One-way ANOVA. The Data / Notation. One way ANOVA Cell means model Factor effects model Topic 17 - Single Factor Analysis of Variance - Fall 2013 One way ANOVA Cell means model Factor effects model Outline Topic 17 2 One-way ANOVA Response variable Y is continuous Explanatory variable is

More information

ML and REML Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 58

ML and REML Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 58 ML and REML Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 58 Suppose y = Xβ + ε, where ε N(0, Σ) for some positive definite, symmetric matrix Σ.

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

14 Multiple Linear Regression

14 Multiple Linear Regression B.Sc./Cert./M.Sc. Qualif. - Statistics: Theory and Practice 14 Multiple Linear Regression 14.1 The multiple linear regression model In simple linear regression, the response variable y is expressed in

More information

Classical Estimation Topics

Classical Estimation Topics Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Lecture 9 SLR in Matrix Form

Lecture 9 SLR in Matrix Form Lecture 9 SLR in Matrix Form STAT 51 Spring 011 Background Reading KNNL: Chapter 5 9-1 Topic Overview Matrix Equations for SLR Don t focus so much on the matrix arithmetic as on the form of the equations.

More information

Basic Concepts in Matrix Algebra

Basic Concepts in Matrix Algebra Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1

More information

x = 1 n (x = 1 (x n 1 ι(ι ι) 1 ι x) (x ι(ι ι) 1 ι x) = 1

x = 1 n (x = 1 (x n 1 ι(ι ι) 1 ι x) (x ι(ι ι) 1 ι x) = 1 Estimation and Inference in Econometrics Exercises, January 24, 2003 Solutions 1. a) cov(wy ) = E [(WY E[WY ])(WY E[WY ]) ] = E [W(Y E[Y ])(Y E[Y ]) W ] = W [(Y E[Y ])(Y E[Y ]) ] W = WΣW b) Let Σ be a

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

Multiple Regression. Dr. Frank Wood. Frank Wood, Linear Regression Models Lecture 12, Slide 1

Multiple Regression. Dr. Frank Wood. Frank Wood, Linear Regression Models Lecture 12, Slide 1 Multiple Regression Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 12, Slide 1 Review: Matrix Regression Estimation We can solve this equation (if the inverse of X

More information

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to 1.1. Introduction Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Introduction to Numerical Linear Algebra II

Introduction to Numerical Linear Algebra II Introduction to Numerical Linear Algebra II Petros Drineas These slides were prepared by Ilse Ipsen for the 2015 Gene Golub SIAM Summer School on RandNLA 1 / 49 Overview We will cover this material in

More information

1. (Rao example 11.15) A study measures oxygen demand (y) (on a log scale) and five explanatory variables (see below). Data are available as

1. (Rao example 11.15) A study measures oxygen demand (y) (on a log scale) and five explanatory variables (see below). Data are available as ST 51, Summer, Dr. Jason A. Osborne Homework assignment # - Solutions 1. (Rao example 11.15) A study measures oxygen demand (y) (on a log scale) and five explanatory variables (see below). Data are available

More information

Lecture 1 Review: Linear models have the form (in matrix notation) Y = Xβ + ε,

Lecture 1 Review: Linear models have the form (in matrix notation) Y = Xβ + ε, 2. REVIEW OF LINEAR ALGEBRA 1 Lecture 1 Review: Linear models have the form (in matrix notation) Y = Xβ + ε, where Y n 1 response vector and X n p is the model matrix (or design matrix ) with one row for

More information

STA442/2101: Assignment 5

STA442/2101: Assignment 5 STA442/2101: Assignment 5 Craig Burkett Quiz on: Oct 23 rd, 2015 The questions are practice for the quiz next week, and are not to be handed in. I would like you to bring in all of the code you used to

More information

Answer Key for STAT 200B HW No. 8

Answer Key for STAT 200B HW No. 8 Answer Key for STAT 200B HW No. 8 May 8, 2007 Problem 3.42 p. 708 The values of Ȳ for x 00, 0, 20, 30 are 5/40, 0, 20/50, and, respectively. From Corollary 3.5 it follows that MLE exists i G is identiable

More information

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Final Review. Yang Feng.   Yang Feng (Columbia University) Final Review 1 / 58 Final Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Final Review 1 / 58 Outline 1 Multiple Linear Regression (Estimation, Inference) 2 Special Topics for Multiple

More information

Power & Sample Size Calculation

Power & Sample Size Calculation Chapter 7 Power & Sample Size Calculation Yibi Huang Chapter 7 Section 10.3 Power & Sample Size Calculation for CRDs Power & Sample Size for Factorial Designs Chapter 7-1 Power & Sample Size Calculation

More information

Ch. 12 Linear Bayesian Estimators

Ch. 12 Linear Bayesian Estimators Ch. 1 Linear Bayesian Estimators 1 In chapter 11 we saw: the MMSE estimator takes a simple form when and are jointly Gaussian it is linear and used only the 1 st and nd order moments (means and covariances).

More information

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix Definition: Let L : V 1 V 2 be a linear operator. The null space N (L) of L is the subspace of V 1 defined by N (L) = {x

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

ECO Class 6 Nonparametric Econometrics

ECO Class 6 Nonparametric Econometrics ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................

More information

identity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij

identity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij Notation R R n m R n m r R n s real numbers set of n m real matrices subset of R n m consisting of matrices with rank r subset of R n n consisting of symmetric matrices NND n subset of R n s consisting

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X). 4. Interval estimation The goal for interval estimation is to specify the accurary of an estimate. A 1 α confidence set for a parameter θ is a set C(X) in the parameter space Θ, depending only on X, such

More information

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That Statistics Lecture 2 August 7, 2000 Frank Porter Caltech The plan for these lectures: The Fundamentals; Point Estimation Maximum Likelihood, Least Squares and All That What is a Confidence Interval? Interval

More information