8. Hypothesis Testing

Size: px
Start display at page:

Download "8. Hypothesis Testing"

Transcription

1 FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1

2 Introduction elements of statistical tests null hypothesis, alternative hypothesis test statistics rejection region type of errors: type I and type II errors confidence intervals, p-values examples of hypothesis tests: hypothesis tests for the mean, and for comparing the means hypothesis tests for the variance, and for comparing variances Hypothesis Testing 8-2

3 Testing procedures a test consists of providing a statement of the hypotheses (H 0 (null) and H 1 (alternative)) giving a rule that dictates if H 0 should be rejected or not the decision rule involves a test statistic calculated on observed data the Neyman-Pearson methodology partitions the sample space into two regions the set of values of the test statistic for which: the null hypothesis is rejected we fail to reject the null hypothesis rejection region acceptance region Hypothesis Testing 8-3

4 Test errors since a test statistic is random, the same test can lead to different conclusions type I error: the test leads to reject H 0 when it is true type II error: the test fails to reject H 0 when it is false; sometimes called false alarm probabilities of the errors: let β be the probability of type II error the size of a test is the probability of a type I error and denoted by α the power of a test is the probability of rejecting a false H 0 or (1 β) α is known as significance level and typically controlled by an analyst for a given α, we would like β to be as small as possible Hypothesis Testing 8-4

5 Some common tests normal test t-test F -test Chi-square test e.g. a test is called a t-test if the test statistic follows t-distribution two approaches of hypothesis test critical value approach p-value approach Hypothesis Testing 8-5

6 Critical value approach Definition: the critical value (associated with a significance level α) is the value of the known distribution of the test statistic such that the probability of type I error is α steps involved this test 1. define the null and alternative hypotheses. 2. assume the null hypothesis is true and calculate the value of the test statistic 3. set a small significance level (typically α = 0.01, 0.05, or 0.10) and determine the corresponding critical value 4. compare the test statistic to the critical value condition decision the test statistic is more extreme than the critical value reject H 0 the test statistic is less extreme than the critical value accept H 0 Hypothesis Testing 8-6

7 example: hypothesis test on the population mean samples N = 15, α = 0.05 the test statistic is t = x µ s/ N and has t-distribution with N 1 df test H 0 H 1 critical value reject H 0 if right-tail µ = 3 µ > 3 t α,n 1 t t α,n 1 left-tail µ = 3 µ < 3 t α,n 1 t t α,n 1 two-tail µ = 3 µ 3 t α/2,n 1, t α/2,n 1 t t α/2,n 1 or t t α/2,n pdf of t-distribution pdf of t-distribution pdf of t-distribution Hypothesis Testing 8-7

8 p-value approach Definition: the p-value is the probability that we observe a more extreme test statistic in the direction of H 1 steps involved this test 1. define the null and alternative hypotheses. 2. assume the null hypothesis is true and calculate the value of the test statistic 3. calculate the p-value using the known distribution of the test statistic 4. set a significance level α (small value such as 0.01, 0.05) 5. compare the p-value to α condition decision p-value α reject H 0 p-value α accept H 0 Hypothesis Testing 8-8

9 example: hypothesis test on the population mean (same as on page 8-7) samples N = 15, α = 0.01 (have only a 1% chance of making a Type I error) suppose the test statistic (calculated from data) is t = test H 0 H 1 p-value expression p-value right-tail µ = 3 µ > 3 P (t 14 2) left-tail µ = 3 µ < 3 P (t 14 2) two-tail µ = 3 µ 3 P (t 14 2) + P (t 14 2) Right tail Left tail Two tails pdf of t-distribution pdf of t-distribution pdf of t-distribution area A area B = area A + area B right-tail/left-tail tests: reject H 0, two-tail test: accept H 0 Hypothesis Testing 8-9

10 the two approaches assume H 0 were true and determine p-value the probability of observing a more extreme test statistic in the direction of the alternative hypothesis than the one observed critical value whether or not the observed test statistic is more extreme than would be expected (called critical value) the null hypothesis is rejected if p-value p value α critical value test statistic critical value Hypothesis Testing 8-10

11 Hypothesis testing in this chapter, we discuss about the following tests Wald test likelihood ratio test Lagrange multiplier (or score) test Hypothesis Testing 8-11

12 Wald test requires estimation of the unrestricted model linear hypotheses in linear models some Wald test statistics examples Hypothesis Testing 8-12

13 Linear hypotheses in linear models a generalization of tests for linear strictions in the linear regression model null and alternative hypotheses for a two-sided test of linear restrictions on the regression parameters in the model: y = Xβ + u are H 0 : Rβ b = 0 H 1 : Rβ b 0 where R R m n of full rank m, β R n and m n for example, one can test β 1 = 1 and β 2 β 3 = 2 the Wald test of Rβ b = 0 is a test of closeness to zero of the sample analogue R ˆβ b where ˆβ is the unrestricted OLS estimator Hypothesis Testing 8-13

14 assumption: suppose u N (0, σ 2 I) then ˆβ N (β, σ 2 (X T X) 1 ) R ˆβ b N (0, σ 2 R(X T X) 1 R T ) under H 0 where Rβ b = 0 define û = y X ˆβ ls, RSS = N N û 2 i, s 2 = RSS/(N n) = (N n) 1 i=1 i=1 û 2 i Facts: s 2 is an unbiased estimate for σ 2 (N n)s 2 /σ 2 χ 2 (N n) Hypothesis Testing 8-14

15 Some Wald test statistics on linear models known variance σ 2 (cannot be calculated in practice) W 1 = (R ˆβ b) T (σ 2 R(X T X) 1 R T ) 1 (R ˆβ b) χ 2 (m) under H 0 replace σ 2 by any consistent estimate s 2 (not necessarily s 2 on page 8-14) W 2 = (R ˆβ b) T (s 2 R(X T X) 1 R T ) 1 (R ˆβ b) a χ 2 (m) under H 0 use s 2 = (N n) 1 i û2 i W 3 = (1/m)(R ˆβ b) T (s 2 R(X T X) 1 R T ) 1 (R ˆβ b) F (m, N n) under H 0 Hypothesis Testing 8-15

16 simple proof: W 1 is in the form of z T A 1 z where cov(z) = A, then W 1 = (A 1/2 z) T (A 1/2 z) quadratic form of standard Gaussian vector use the result on page 3-51: quadratic form of Gaussian is Chi-square W 2 = (σ 2 /s 2 )W 1 and plim(σ 2 /s 2 ) = 1, so W 2 converges to a Chi-square asymptotically (use Tranformation theorem on page 4-15) we can write W 3 as a ratio between two scaled Chi-square RVs W 3 = W 1/m s 2 /σ 2 = W 1 /m ((N n)s 2 /σ 2 )/(N n) Hypothesis Testing 8-16

17 Wald test of one restriction for a test of one restriction on linear regression model: y = Xβ + u, u N (0, σ 2 I) homoskedasticity and X is deterministic the hypotheses are H 0 : a T β b = 0, H 1 : a T β b 0 where a R n and b R for the LS estimate, it s easy to show that a T ˆβ b is Gaussian with E[a T ˆβ b] = 0, cov(a T ˆβ b) = a T σ 2 (X T X) 1 a under H 0 Hypothesis Testing 8-17

18 therefore, we can propose two Wald statistics Wald z-test statistic: W 4 = a T ˆβ b at σ 2 (X T X) 1 a N (0, 1) Wald t-test statistic: use s 2 = RSS/(N n) W 5 = a T ˆβ b at s 2 (X T X) 1 a t N n we can write W 5 a ratio of standard normal to sqrt of scaled Chi-square: W 5 = a T ˆβ b at s 2 (X T X) 1 a = T a ˆβ b a T σ 2 (X T X) 1 a (N n)s 2 σ 2 (N n) t N n Hypothesis Testing 8-18

19 Example on one exclusion restriction consider the exclusion restriction that β 1 is zero: a = (1, 0,..., 0), b = 0 suppose we use s 2 = RSS/(N n) so that Âvar( ˆβ) = s 2 (X T X) 1, Âvar( ˆβ 1 ) = ( s 2 (X T X) 1) 11 Wald test statistics for exclusion restriction are W 3 = ˆβ 2 1 Âvar( ˆβ 1 ) F (1, N n) W 5 = ˆβ 1 Âvar( ˆβ 1 ) t N n W 2 = ˆβ 2 1 Âvar( ˆβ 1 ) a χ 2 (1) if another consistent s 2 is used Hypothesis Testing 8-19

20 Nonlinear hypotheses consider hypothesis tests of m restriction that are nonlinear in θ let θ R n and r(θ) : R n R m be restriction function the null and alternative hypotheses for a two-sided tests are H 0 : r(θ ) = 0, H 1 : r(θ ) 0 examples: r(θ) = θ 2 = 0 or r(θ) = θ 1 θ 2 1 = 0 assumptions: the Jacobian matrix of r: R(θ) = Dr(θ) R m n is full rank m at θ parameters are not at the boundary of Θ under H 0, e.g., we rule out H 0 : θ 1 = 0 if the model requires θ 1 0 Hypothesis Testing 8-20

21 Wald test statistic for nonlinear restriction intuition: obtain ˆθ w/o imposing restrictions and see if r(ˆθ) 0 the Wald test statistic W = r(ˆθ) T [R(ˆθ)Âvar(ˆθ)R(ˆθ) T ] 1 r(ˆθ) is asymptotically χ 2 (m) distributed under H 0 two equivalent conditions in testing: H 0 is rejected against H 1 at significance level α if W > χ 2 α(m) H 0 is rejected at level α if the p-value: P (χ 2 (m) > W ) < α Hypothesis Testing 8-21

22 Example of nonlinear restriction let θ R n and consider a test of single nonlinear restriction H 0 : r(θ) = θ 1 /θ 2 1 = 0 then R(θ) R 1 n and given by R(θ) = [ 1/θ 2 θ 1 /θ ] let a ij be (i, j) entry of Âvar(ˆθ) W = ( ) 2 θ1 1 1 [ θ θ 2 2 θ 1 θ ] a 11 a 12 a 21 a θ 2 θ 1 θ = [θ 2 (θ 1 θ 2 )] 2 ( θ 2 2a 11 2θ 1 θ 2 a 12 + θ 2 1a 22 ) 1 a χ 2 (1) (θ is evaluated at ˆθ and the sample size N is hidden in a ij ) Hypothesis Testing 8-22

23 Derivation of the Wald statistic assumption: ˆθ has a normal limit distribution: N(ˆθ θ ) d N (0, P ) proof: starting from the first-order Taylor expansion of r under H 0 expand r around θ r(ˆθ) = r(θ ) + r(ζ)(ˆθ θ ), ζ is between ˆθ and θ write N(r(ˆθ) r(θ )) = R(ζ) N(ˆθ θ ) and note that R(ζ) p R(θ ), N(ˆθ θ ) d N (0, P ) (use that R is continuous, apply Slustky and sandwich theorems) Hypothesis Testing 8-23

24 by Product Limit Normal Rule on page 4-16 N(r(ˆθ) r(θ )) d N (0, R(θ )P R(θ ) T ) under H 0 : r(θ ) = 0 and use Avar(ˆθ) = P/N, we can write r(ˆθ) d N (0, R(θ ) Avar(ˆθ)R(θ ) T ) a quadratic form of standard Gaussian is a chi-square (on page 3-51) r(ˆθ) T [ R(θ ) Avar(ˆθ)R(θ ) T ] 1 r(ˆθ) a χ 2 (m) the Wald test statistic is obtained by using estimates of R(θ ) and Avar(ˆθ) R(ˆθ), Âvar(ˆθ) = ˆP /N Hypothesis Testing 8-24

25 Likelihood-based tests hypothesis testings when the likelihood function is known Wald test likelihood ratio (LR) test Lagrange multiplier (or score) test we denote L(θ) = f(y 1,..., y N x 1,..., x N, θ) likelihood function r(θ) : R n R m restriction function with H 0 : r(θ) = 0 ˆθ u : unrestricted MLE which maximizes L ˆθ r : restricted MLE which maximizes the Lagrangian log L(θ) λ T r(θ) Hypothesis Testing 8-25

26 Likelihood ratio test idea: if H 0 is true, the unconstrained and constrained maximization of log L should be the same it can be shown that the likelihood ratio test statistic: LR = 2[log L(ˆθ r ) log L(ˆθ u )] is asymptotically chi-square distributed under H 0 with degree of freedom m if H 0 is true, r(ˆθ u ) should be close to zero note that log L(ˆθ u ) is always greater than log L(ˆθ r ) LR test requires both ˆθ u and ˆθ r m is the number of restriction equations Hypothesis Testing 8-26

27 Wald test idea: if H 0 is true, ˆθ u should satisfy r(ˆθ u ) 0 specifically for MLE, the estimate covariance satisfies CR bound and IM equality: Avar(ˆθ u ) = I N (θ) 1 = ( E[ 2 log L(θ )] ) 1 P/N this leads to the Wald test statistic W = r(ˆθ u ) T [ R(ˆθ u )Âvar(ˆθ u )R(ˆθ u ) T ] 1 r(ˆθu ) a χ 2 (m) where Âvar(ˆθ u ) is an estimated asymptotic covariance of ˆθ u the advantage over LR test is that only ˆθ u is required Hypothesis Testing 8-27

28 Lagrange multiplier (or score) test ideas: we know that log L(ˆθ u ) = 0 (because it s an unconstrained maximization) if H 0 is true, then maximum should also occur at ˆθ r : log L(ˆθ r ) 0 LM test is called score test because log L(θ) is the score vector maximizing the Lagrangian: log L(θ) λ T r(θ) implies that log L(ˆθ r ) = r(ˆθ r ) T λ tests based on λ are equivalent to tests based on log L(ˆθ r ) because we assume r(θ) to be full rank Hypothesis Testing 8-28

29 the LM test requires the asymptotic distribution of log L(ˆθ r ) note that the asymptotic covariance of log L(ˆθ r ) is the information matrix this leads to the Lagrange multiplier test or score test statistic LM = ( log L(ˆθ r )) T [I N (ˆθ r )] 1 ( log L(ˆθ r )) which is asymptotically chi-square with m degree of freedoms Hypothesis Testing 8-29

30 Graphical interpretation of loglikelihood based tests Likelihood ratio Wald test checks if r(ˆθ u ) 0 0 Lagrange multiplier Wald LR test checks the difference between log L(ˆθ u ) and log L(ˆθ r ) LM test checks that the slope of loglikelihood at the restricted estimator should be near zero W.H. Greene, Econometric Analysis, Prentice Hall, 2008 Hypothesis Testing 8-30

31 Gaussian example consider i.i.d. example with y i N (µ, 1) with hypothesis H 0 : µ = µ 0, (µ 0 is just a constant value, and given) therefore, ˆµ u = ȳ and ˆµ r = µ 0 (restricted solution) log L(µ) = (N/2) log 2π (1/2) i (y i µ) 2 µ log L(µ) = i (y i µ) LR test: with some algebra, we can write that LR = 2[log L(ȳ) log L(µ 0 )] = N(ȳ µ 0 ) 2 Wald test: we check if ȳ µ 0 0 Hypothesis Testing 8-31

32 under H 0, the true mean of y i is µ 0, so E[ȳ] = µ 0 and var(ȳ) = 1/N hence, (ȳ µ 0 ) N (0, 1/N) W = (ȳ µ 0 )(1/N) 1 (ȳ µ 0 ) = N(ȳ µ 0 ) 2 LM test: check if log L(µ 0 ) = N(ȳ µ 0 ) 0 log L(µ 0 ) = i (y i µ 0 ) = N(ȳ µ 0 ) LM is just a rescaling of (ȳ µ 0 ), so LM = W in conclusion, the three statistics are equivalent aysmptotically but they differ in finite samples LM = W = LR Hypothesis Testing 8-32

33 MATLAB example the example is based on estimating the mean of a Gaussian model N = 50; mu = 2; y = mu + randn(n,1); ybar = mean(y); mu0 = mu; dof = 1; % number of restrictions % Wald rw = ybar-mu0; Rw = 1; EstCov = 1/N; [h,pvalue,stat,cvalue] = waldtest(rw,rw,estcov) % LR ull = -(1/2)*sum((y-ybar).^2);rLL = -(1/2)*sum((y-mu0).^2); [h,pvalue] = lratiotest(ull,rll,dof) % LM score = N*(ybar-mu0);EstCov = 1/N; % I_N(\theta) = N [h,pvalue] = lmtest(score,estcov,dof) Hypothesis Testing 8-33

34 the three tests are the same and the result is ybar = h = 0 pvalue = stat = cvalue = the p-value is greater than α = 0.05 (default value), so H 0 is accepted Hypothesis Testing 8-34

35 Poisson regression example consider the log-likelihood function in the Poisson regression model log L(β) = N [ e xt i β + y i x T i β log y i!] i=1 suppose β = (β 1, β 2 ) and H 0 : r(β) = β 2 = 0 the first and second derivatives of log L(β) are log L(β) = i (y i e xt i β )x i, 2 log L(β) = i e xt i β x i x T i unrestricted MLE, ˆβ u = ( ˆβ u1, ˆβ u2 ), satisfies log L(β) = 0 restricted MLE, ˆβ r = ( ˆβ r1, 0) where ˆβ r1 solves i (y i e xt i1 β 1)x i1 = 0 Hypothesis Testing 8-35

36 all the there statistics can be derived as LR test: calculate the fitted log-likelihood of ˆβ u and ˆβ r Wald test: compute the asympototic covariance of ˆβ u and its estimate Avar( ˆβ u ) = I N (β) 1 = E[ 2 log L(β)] 1 = ( E[ i e xt i β x i x T i ] ) 1 Âvar( ˆβ u ) = ( e xt i ˆβ u x i x T i i ) 1 from r(β) = ˆβ 2 and R(β) = [ 0 I ], Wald statistic is W = ˆβ T u2 (Âvar( ˆβu ) 22 ) 1 ˆβu2 Hypothesis Testing 8-36

37 where Âvar( ˆβ u ) 22 denotes the (2, 2) block of Âvar( ˆβ u ) LM test: it is based on log L( ˆβ r ) log L( ˆβ r ) = i (y i e xt i ˆβ r )x i = i x i û i where û i = y i e xt i1 ˆβ r1 the LM statistic is LM = [ ] T [ x i û i i e xt i1 ˆβ r1 x i x T i ] 1 [ ] x i û i i i some further simplification is possible since i x i1û i = 0 (from first-order condition) LM test here is based on the correlation between the omitted regressors and the residual, û Hypothesis Testing 8-37

38 MATLAB example data generation and solve for unrestricted and restricted MLE estimates % Data generation beta = [1 0] ; % The true value x = randn(n,2); y = zeros(n,1); lambda = zeros(n,1); for k=1:n, lambda(k) = exp(x(k,:)*beta); y(k) = poissrnd(lambda(k)); % generate samples of y end % minimization of -Loglikelihood function (change the sign of LogL) neglogfun -sum(-exp( sum(x.*repmat(beta,n,1),2) )... +y.*sum(x.*repmat(beta,n,1),2) ); beta0 = [2 2] ; % initial value [beta_u,ulogl] = fminunc(neglogfun,beta0); ulogl = -ulogl; % change back the sign of LogL Hypothesis Testing 8-38

39 % solving for restricted MLE neglogfun_r -sum(-exp( sum(x(:,1).*repmat(beta(1),n,1),2))... +y.*sum(x(:,1).*repmat(beta(1),n,1),2) ); % when beta2 = 0 [beta_r1,rlogl] = fminunc(neglogfun_r,beta0(1)) rlogl = -rlogl; beta_r = [beta_r1 0] ; The two estimates are beta_u = beta_r = Hypothesis Testing 8-39

40 Wald test TMP = 0; for ii=1:n, TMP = TMP + exp(x(ii,:)*beta_u)*x(ii,:) *x(ii,:); end rw = beta_u(2); Rw = [0 1]; EstCovw = TMP\eye(2) [h,pvalue,stat,cvalue] = waldtest(rw,rw,estcovw) h = 0 pvalue = stat = cvalue = accept H 0 since p-value is greater than α = 0.05 Hypothesis Testing 8-40

41 LR test LR = 2*(uLogL - rlogl) [h,pvalue,stat,cvalue] = lratiotest(ulogl,rlogl,dof) LR = h = 0 pvalue = stat = cvalue = Hypothesis Testing 8-41

42 LM test uhat = y - exp(sum(x.*repmat(beta_r,n,1),2)) ; scorei = x.*repmat(uhat,1,2) ; scorei = scorei ; score = sum(scorei,2); % expect of outer product of gradient of LogL EstCovlm1 = scorei*scorei ; EstCovlm2 = 0; % expectation of Hessian of LogL for ii=1:n, EstCovlm2 = EstCovlm2 + exp(x(ii,:)*beta_r)*x(ii,:) *x(ii,:); end % note that EstCovlm1 and EstCovlm2 should be close to each other % choose to use EstCovlm2 EstCovlm = EstCovlm2\eye(2); % covariance matrix of parameters [h,pvalue,stat,cvalue] = lmtest(score,estcovlm,dof) Hypothesis Testing 8-42

43 h = 0 pvalue = stat = cvalue = all the three tests agree what we should accept H 0 since p-value is greater than α Hypothesis Testing 8-43

44 if we change the true value to β = (1, 1) then ˆβ u = (1.0954, ), ˆβr = (1.3942, 0) we found that all the three tests reject H 0, i.e., β 2 is not close to zero Hypothesis Testing 8-44

45 Summary the three tests are asymptotically equivalent under H 0 but they can behave rather differently in a small sample LR test requires calculation of both restricted and unrestricted estimators Wald test requires only unrestricted estimator LM test requires only restricted estimator the choice among them typically depends on the ease of computation Hypothesis Testing 8-45

46 Hypothesis Testing introduction Wald test likelihood-based tests significance test for linear regression Hypothesis Testing 8-46

47 Recap of linear regression a linear regression model is y = Xβ + u homoskedasticity assumption: u i has the same variance for all i, given by σ 2 prediction (fitted) error: û := ŷ y = X ˆβ y residual sum of squares: RSS = û 2 2 a consistent estimate of σ 2 : s 2 = RSS/(N n) (N n)s 2 χ 2 (N n) square root of s 2 is called standard error of the regression Avar( ˆβ) = s 2 (X T X) 1 (can replace s 2 by any consistent ˆσ 2 ) Hypothesis Testing 8-47

48 Common tests for linear regression testing a hypothesis about a coefficient we can use both t and F statistics testing using the fit of the regression H 0 : β k = 0 VS H 1 : β k 0 H 0 : reduced model VS H 1 : full model if H 0 were true, the reduced model (β k = 0) would lead to smaller prediction error than that of the full model (β k 0) Hypothesis Testing 8-48

49 Testing a hypothesis about a coefficient statistics for testing hypotheses: H 0 : β k = 0 VS H 1 : β k 0 ˆβ k t N n s 2 ((X T X) 1 ) kk ( ˆβ k ) s 2 F 1,N n 2 ((X T X) 1 ) kk the above statistics are Wald statistics derived on page 8-17 through 8-19 the term s 2 ((X T X) 1 ) kk is referred to standard error of the coefficient the expression of SE can be simplified or derived in many ways (please check) e.g. R use t-statistic (two-tail test) Hypothesis Testing 8-49

50 Testing using the fit of the regression hypotheses are based on the fitting quality of reduced/full models H 0 : reduced model VS H 1 : full model reduced model: β k = 0 and full model: β k 0 the F -statistic used in this test (RSS R RSS F ) RSS F /(N n) F (1, N n) RSS R and RSS F are the residual sum squares of reduced and full models RSS R cannot be smaller than RSS F, so if H 0 were true, then the F statistic would be zero e.g. fitlm in MATLAB use this F statistic, or in ANOVA table Hypothesis Testing 8-50

51 MATLAB example perform t-test using α = 0.05 and the true parameter is β = (1, 0, 1, 0.5) realization 1: N = 100 >> [btrue b SE pvalue2side] = ˆβ is close to β it s not clear if ˆβ 2 is zero but the test decides ˆβ 2 = 0 note that all coefficients have pretty much the same SE Hypothesis Testing 8-51

52 realization 2: N = 10 >> [btrue b SE pvalue2side] = realization 3: N = 10 >> [btrue b SE pvalue2side] = some of ˆβ is close to the true value but some is not the test 2 decides ˆβ 2 and ˆβ 4 are zero while the test 3 decides all β are zero the sample size N affects type II error (fails to reject H 0 ) and we get different results from different data sets Hypothesis Testing 8-52

53 Summary common tests are available in many statistical softwares, e.g, minitab, lm in R, fitlm in MATLAB, one should use with care and interpret results correctly an estimator is random; one cannot trust its value calculated based on a data set examining statistical properties of an estimator is preferred Hypothesis Testing 8-53

54 Chapter 12 in References J.M. Wooldridge, Econometric Analysis of Cross Section and Panel Data, the MIT press, 2010 Chapter 5 in A.C. Cameron and P.K. Trivedi, Microeconometircs: Methods and Applications, Cambridge, 2005 Chapter 5, 11 and 16 in W.H. Greene, Econometric Analysis, Prentice Hall, 2008 Review of Basic Statistics (online course) Stat 501 (online course) Hypothesis Testing 8-54

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007 Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

Statistics and econometrics

Statistics and econometrics 1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample

More information

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ ) Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

LECTURE 5 HYPOTHESIS TESTING

LECTURE 5 HYPOTHESIS TESTING October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.

More information

The outline for Unit 3

The outline for Unit 3 The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.

More information

Greene, Econometric Analysis (6th ed, 2008)

Greene, Econometric Analysis (6th ed, 2008) EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived

More information

Some General Types of Tests

Some General Types of Tests Some General Types of Tests We may not be able to find a UMP or UMPU test in a given situation. In that case, we may use test of some general class of tests that often have good asymptotic properties.

More information

Practical Econometrics. for. Finance and Economics. (Econometrics 2)

Practical Econometrics. for. Finance and Economics. (Econometrics 2) Practical Econometrics for Finance and Economics (Econometrics 2) Seppo Pynnönen and Bernd Pape Department of Mathematics and Statistics, University of Vaasa 1. Introduction 1.1 Econometrics Econometrics

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Florian Pelgrin HEC September-December 2010 Florian Pelgrin (HEC) Constrained estimators September-December

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Economics 582 Random Effects Estimation

Economics 582 Random Effects Estimation Economics 582 Random Effects Estimation Eric Zivot May 29, 2013 Random Effects Model Hence, the model can be re-written as = x 0 β + + [x ] = 0 (no endogeneity) [ x ] = = + x 0 β + + [x ] = 0 [ x ] = 0

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is

More information

Lecture 15. Hypothesis testing in the linear model

Lecture 15. Hypothesis testing in the linear model 14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin March 23, 2011 1 Introduction 2 3 4 5 Outline Introduction 1 Introduction 2 3 4 5 Preliminaries Introduction Ordinary least squares

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Lecture 3: Multiple Regression

Lecture 3: Multiple Regression Lecture 3: Multiple Regression R.G. Pierse 1 The General Linear Model Suppose that we have k explanatory variables Y i = β 1 + β X i + β 3 X 3i + + β k X ki + u i, i = 1,, n (1.1) or Y i = β j X ji + u

More information

Ch 3: Multiple Linear Regression

Ch 3: Multiple Linear Regression Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Maximum Likelihood Tests and Quasi-Maximum-Likelihood Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

STAT763: Applied Regression Analysis. Multiple linear regression. 4.4 Hypothesis testing

STAT763: Applied Regression Analysis. Multiple linear regression. 4.4 Hypothesis testing STAT763: Applied Regression Analysis Multiple linear regression 4.4 Hypothesis testing Chunsheng Ma E-mail: cma@math.wichita.edu 4.4.1 Significance of regression Null hypothesis (Test whether all β j =

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 3 Jakub Mućk Econometrics of Panel Data Meeting # 3 1 / 21 Outline 1 Fixed or Random Hausman Test 2 Between Estimator 3 Coefficient of determination (R 2

More information

Exercises Chapter 4 Statistical Hypothesis Testing

Exercises Chapter 4 Statistical Hypothesis Testing Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

Spatial Regression. 9. Specification Tests (1) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Spatial Regression. 9. Specification Tests (1) Luc Anselin.   Copyright 2017 by Luc Anselin, All Rights Reserved Spatial Regression 9. Specification Tests (1) Luc Anselin http://spatial.uchicago.edu 1 basic concepts types of tests Moran s I classic ML-based tests LM tests 2 Basic Concepts 3 The Logic of Specification

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA March 6, 2017 KC Border Linear Regression II March 6, 2017 1 / 44 1 OLS estimator 2 Restricted regression 3 Errors in variables 4

More information

2014/2015 Smester II ST5224 Final Exam Solution

2014/2015 Smester II ST5224 Final Exam Solution 014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

2018 2019 1 9 sei@mistiu-tokyoacjp http://wwwstattu-tokyoacjp/~sei/lec-jhtml 11 552 3 0 1 2 3 4 5 6 7 13 14 33 4 1 4 4 2 1 1 2 2 1 1 12 13 R?boxplot boxplotstats which does the computation?boxplotstats

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests

Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests Throughout this chapter we consider a sample X taken from a population indexed by θ Θ R k. Instead of estimating the unknown parameter, we

More information

Maximum Likelihood (ML) Estimation

Maximum Likelihood (ML) Estimation Econometrics 2 Fall 2004 Maximum Likelihood (ML) Estimation Heino Bohn Nielsen 1of32 Outline of the Lecture (1) Introduction. (2) ML estimation defined. (3) ExampleI:Binomialtrials. (4) Example II: Linear

More information

Lecture 6: Hypothesis Testing

Lecture 6: Hypothesis Testing Lecture 6: Hypothesis Testing Mauricio Sarrias Universidad Católica del Norte November 6, 2017 1 Moran s I Statistic Mandatory Reading Moran s I based on Cliff and Ord (1972) Kelijan and Prucha (2001)

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

Motivation for multiple regression

Motivation for multiple regression Motivation for multiple regression 1. Simple regression puts all factors other than X in u, and treats them as unobserved. Effectively the simple regression does not account for other factors. 2. The slope

More information

Econometrics II - EXAM Answer each question in separate sheets in three hours

Econometrics II - EXAM Answer each question in separate sheets in three hours Econometrics II - EXAM Answer each question in separate sheets in three hours. Let u and u be jointly Gaussian and independent of z in all the equations. a Investigate the identification of the following

More information

Composite Hypotheses and Generalized Likelihood Ratio Tests

Composite Hypotheses and Generalized Likelihood Ratio Tests Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

BIOS 2083 Linear Models c Abdus S. Wahed

BIOS 2083 Linear Models c Abdus S. Wahed Chapter 5 206 Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter

More information

WISE International Masters

WISE International Masters WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are

More information

Review of Econometrics

Review of Econometrics Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,

More information

Binary choice 3.3 Maximum likelihood estimation

Binary choice 3.3 Maximum likelihood estimation Binary choice 3.3 Maximum likelihood estimation Michel Bierlaire Output of the estimation We explain here the various outputs from the maximum likelihood estimation procedure. Solution of the maximum likelihood

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

Lecture 4: Testing Stuff

Lecture 4: Testing Stuff Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed

More information

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u Interval estimation and hypothesis tests So far our focus has been on estimation of the parameter vector β in the linear model y i = β 1 x 1i + β 2 x 2i +... + β K x Ki + u i = x iβ + u i for i = 1, 2,...,

More information

INTERVAL ESTIMATION AND HYPOTHESES TESTING

INTERVAL ESTIMATION AND HYPOTHESES TESTING INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,

More information

Answer Key for STAT 200B HW No. 8

Answer Key for STAT 200B HW No. 8 Answer Key for STAT 200B HW No. 8 May 8, 2007 Problem 3.42 p. 708 The values of Ȳ for x 00, 0, 20, 30 are 5/40, 0, 20/50, and, respectively. From Corollary 3.5 it follows that MLE exists i G is identiable

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

This paper is not to be removed from the Examination Halls

This paper is not to be removed from the Examination Halls ~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,

More information

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2 Fall, 2013 Page 1 Random Variable and Probability Distribution Discrete random variable Y : Finite possible values {y

More information

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, 2016-17 Academic Year Exam Version: A INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Lecture 6 Multiple Linear Regression, cont.

Lecture 6 Multiple Linear Regression, cont. Lecture 6 Multiple Linear Regression, cont. BIOST 515 January 22, 2004 BIOST 515, Lecture 6 Testing general linear hypotheses Suppose we are interested in testing linear combinations of the regression

More information

Testing and Model Selection

Testing and Model Selection Testing and Model Selection This is another digression on general statistics: see PE App C.8.4. The EViews output for least squares, probit and logit includes some statistics relevant to testing hypotheses

More information

The Standard Linear Model: Hypothesis Testing

The Standard Linear Model: Hypothesis Testing Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 25: The Standard Linear Model: Hypothesis Testing Relevant textbook passages: Larsen Marx [4]:

More information

Probability and Statistics Notes

Probability and Statistics Notes Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline

More information

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments /4/008 Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University C. A Sample of Data C. An Econometric Model C.3 Estimating the Mean of a Population C.4 Estimating the Population

More information

Sampling distribution of GLM regression coefficients

Sampling distribution of GLM regression coefficients Sampling distribution of GLM regression coefficients Patrick Breheny February 5 Patrick Breheny BST 760: Advanced Regression 1/20 Introduction So far, we ve discussed the basic properties of the score,

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants 18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009

More information

Linear models and their mathematical foundations: Simple linear regression

Linear models and their mathematical foundations: Simple linear regression Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction

More information

MATH5745 Multivariate Methods Lecture 07

MATH5745 Multivariate Methods Lecture 07 MATH5745 Multivariate Methods Lecture 07 Tests of hypothesis on covariance matrix March 16, 2018 MATH5745 Multivariate Methods Lecture 07 March 16, 2018 1 / 39 Test on covariance matrices: Introduction

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

Likelihood Ratio tests

Likelihood Ratio tests Likelihood Ratio tests For general composite hypotheses optimality theory is not usually successful in producing an optimal test. instead we look for heuristics to guide our choices. The simplest approach

More information

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Eric Zivot Winter 013 1 Wald, LR and LM statistics based on generalized method of moments estimation Let 1 be an iid sample

More information

Graduate Econometrics I: Maximum Likelihood I

Graduate Econometrics I: Maximum Likelihood I Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood

More information

MEI Exam Review. June 7, 2002

MEI Exam Review. June 7, 2002 MEI Exam Review June 7, 2002 1 Final Exam Revision Notes 1.1 Random Rules and Formulas Linear transformations of random variables. f y (Y ) = f x (X) dx. dg Inverse Proof. (AB)(AB) 1 = I. (B 1 A 1 )(AB)(AB)

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = 0 + 1 x 1 + x +... k x k + u 6. Heteroskedasticity What is Heteroskedasticity?! Recall the assumption of homoskedasticity implied that conditional on the explanatory variables,

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

Association studies and regression

Association studies and regression Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Part 1.) We know that the probability of any specific x only given p ij = p i p j is just multinomial(n, p) where p k1 k 2

Part 1.) We know that the probability of any specific x only given p ij = p i p j is just multinomial(n, p) where p k1 k 2 Problem.) I will break this into two parts: () Proving w (m) = p( x (m) X i = x i, X j = x j, p ij = p i p j ). In other words, the probability of a specific table in T x given the row and column counts

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Interpreting Regression Results

Interpreting Regression Results Interpreting Regression Results Carlo Favero Favero () Interpreting Regression Results 1 / 42 Interpreting Regression Results Interpreting regression results is not a simple exercise. We propose to split

More information

Advanced Econometrics

Advanced Econometrics Advanced Econometrics Dr. Andrea Beccarini Center for Quantitative Economics Winter 2013/2014 Andrea Beccarini (CQE) Econometrics Winter 2013/2014 1 / 156 General information Aims and prerequisites Objective:

More information