2.5.3 Generalized likelihood ratio tests
|
|
- Dayna Lee
- 6 years ago
- Views:
Transcription
1 25 HYPOTHESIS TESTING Generalized likelihood ratio tests When a UMP test does not exist, we usually use a generalized likelihood ratio test to verify H 0 : ϑ Θ against H 1 : ϑ Θ\Θ It can be used when H 0 is composite, which none of the above methods can The generalized likelihood ratio test has rejection region R = {y : λ(y a}, where λ(y = max ϑ Θ L(ϑ y max ϑ Θ L(ϑ y is the generalized likelihood ratio and a is a constant chosen to give significance levelα, that is such that P(λ(Y a H 0 = α If we let ˆϑ denote the maximum likelihood estimate of ϑ and let ϑ 0 denote the value of ϑ which maximises the likelihood over all values of ϑ in Θ, then we may write λ(y = L( ϑ 0 y L( ϑ y The quantity ϑ 0 is called the restricted maximum likelihood estimate ofϑunder H 0 Example 237 Suppose that Y i iid N(µ,σ 2 and consider testing H 0 : µ = µ 0 againsth 1 : µ µ 0 Then the likelihood is L(µ,σ 2 y = (2πσ 2 n 2 exp { 1 2σ 2 } n (y i µ 2 The maximum likelihood estimate ofϑ=(µ,σ 2 T is ϑ = ( µ, σ 2 T, where µ = y and σ 2 = n (y i y 2 /n Similarly, the restricted maximum likelihood estimate of ϑ = (µ,σ 2 T under H 0 is ϑ 0 = ( µ 0, σ 2 0 T, where µ 0 = µ 0 and σ 2 0 = n (y i µ 0 2 /n
2 128 CHAPTER 2 ELEMENTS OF STATISTICAL INFERENCE Thus, the generalized likelihood ratio is λ(y = L(µ 0,ˆσ 2 0 y L(ˆµ,ˆσ 2 y = = { } (2πˆσ 0 2 n 2 exp 1 n 2ˆσ 0 2 (y i µ 0 2 (2πˆσ 2 n 2 exp { 1 n 2ˆσ 2 (y i y 2} n { 2 2 (ˆσ n n exp (y i y 2 ˆσ n (y i y n n (y } i µ n (y i µ 0 2 = { n (y i y 2 }n 2 n (y i µ 0 2 Since the rejection region isr = {y : λ(y a}, we reject H 0 if { n (y i y 2 }n 2 n n a (y i y (y 2 i µ 0 2 n (y i µ 0 b, 2 where a and b are constants chosen to give significance level α Now, we may write n n (y i µ 0 2 = {(y i y+(y µ 0 } 2 So we reject H 0 if = = n (y i y 2 +2(y µ 0 n (y i y+n(y µ 0 2 n (y i y 2 +n(y µ 0 2 Thus, rearranging, we reject H 0 if n (y i y 2 n (y i y 2 +n(y µ 0 2 b 1+ n(y µ 0 2 n (y i y c n(y µ n (y i y d, 2 1 n 1 where c and d are constants chosen to give significance level α, that is we can write ( n(y µ0 2 α = P(λ(Y a H 0 = P d H S 2 0, wheres 2 = 1 n 1 n (Y i Y 2
3 25 HYPOTHESIS TESTING 129 To getdwe need to work out the distribution of n(y µ 0 2 under the null hypothesis S 2 ForY i N(µ,σ 2 we have iid Y N (µ, σ2 n and so, underh 0, This gives Also Y N (µ 0, σ2 n and n ( Y µ0 σ 2 n(y µ 0 2 σ 2 χ 2 1 (n 1S 2 σ 2 χ 2 n 1 N(0,1 Now we may use the fact that if U and V are independent rvs such that U χ 2 ν 1 and V χ 2 ν 2, then U/ν 1 V/ν 2 F ν1,ν 2 Here, U = n(y µ 0 2 σ 2 and V = (n 1S2 Hence, ifh σ 2 0 is true, we have F = U/1 V/(n 1 = n(y µ 0 2 S 2 F 1,n 1 Therefore, we reject H 0 at a significance levelα if n(y µ 0 2 F s 2 1,n 1,α, wheref 1,n 1,α is such that P(F F 1,n 1,α = α Equivalently, we reject H 0 if n(y µ0 2 that is, if s 2 t n 1, α 2, y µ 0 s2 /n t n 1, α 2 Of course, this is the usual two-sidedttest In fact, many of the standard tests in situations with normal distributions are generalized likelihood ratio tests
4 130 CHAPTER 2 ELEMENTS OF STATISTICAL INFERENCE 254 Wilks theorem In more complex cases, we have to use the following approximation to find the rejection region The result is stated without proof Theorem 29 Wilks theorem Assume that the joint distribution ofy 1,,Y n depends onpunknown parameters and that, underh 0, the joint distribution depends onp 0 unknown parameters Let ν = p p 0 Then, under some regularity conditions, when the null hypothesis is true, the distribution of the statistic 2log{λ(Y } converges to a χ 2 ν distribution as the sample sizen, ie, when H 0 is true and n is large, 2 log{λ(y } approx χ2 ν Thus, for large n, the rejection region for a test with approximate significance levelαis R = {y : 2log{λ(y} χ 2 ν,α } Example 238 Suppose that Y i iid Poisson(λ and consider testing H 0 : λ = λ 0 againsth 1 : λ λ 0 Then we have seen that no UMP test exists in this case Now, the likelihood is y i e nλ L(λ y = λ n n y i! The maximum likelihood estimate of λ is ˆλ = y and the restricted maximum likelihood estimate of λ under H 0 is ˆλ 0 = λ 0 Thus, the generalized likelihood ratio is n λ(y = L(λ 0 y = λ y i 0 e nλ 0 n y i! n L(ˆλ y y i! y n y i e ny ( n λ0 y i = e n(y λ0 y It follows that { ( λ0 2log{λ(y} = 2 nylog y { ( y = 2n ylog +λ 0 y λ 0 } +n(y λ 0 }
5 25 HYPOTHESIS TESTING 131 Here, p = 1 andp 0 = 0, and soν = 1 Therefore, by Wilks theorem, whenh 0 is true and n is large, { ( } Y 2n Y log +λ 0 Y χ 2 1 λ 0 Hence, for a test with approximate significance level α, we reject H 0 if and only if { ( } y 2n ylog +λ 0 y χ 2 1,α λ 0 Example 239 Suppose that Y i, i = 1,,n, are iid random variables with the probability mass function given by { ϑj, if y = j, j = 1,2,3; P(Y = y = 0, otherwise, whereϑ j are unknown parameters such thatϑ 1 +ϑ 2 +ϑ 3 = 1 andϑ j 0 Consider testing H 0 : ϑ 1 = ϑ 2 = ϑ 3 against H 1 : H 0 is not true We will use the Wilks theorem to derive the critical region for testing this hypothesis at an approximate significance levelα Here the full parameter space Θ is two-dimensional because there are only two free parameters, ie,ϑ 3 = 1 ϑ 1 ϑ 2 and Hencep = 2 Θ = {ϑ = (ϑ 1,ϑ 2,ϑ 3 T : ϑ 3 = 1 ϑ 1 ϑ 2,ϑ j 0} The restricted parameter space is zero-dimensional, because under the null hypothesis all the parameters are equal and as they sum up to 1, they all must be equal to1/3 Hence { ( 1 Θ = ϑ = 3, 1 3, 1 } T 3 and so p 0 = 0 (zero unknown parameters That is the number of degrees of freedom of theχ 2 distribution is ν = p p 0 = 2 To calculateλ(y we need to find the MLE(ϑ inθand in the restricted spaceθ
6 132 CHAPTER 2 ELEMENTS OF STATISTICAL INFERENCE MLE(ϑ inθ: The likelihood function is L(ϑ;y = ϑ n 1 1 ϑn 2 2 (1 ϑ 1 ϑ 2 n 3, Then, the log- where n j is the number of responses equal to j, j = 1,2,3 likelihood is Forj = 1,2 we have, l(ϑ;y = n 1 log(ϑ 1 +n 2 log(ϑ 2 +n 3 log(1 ϑ 1 ϑ 2 When compared to zero, this gives l = n j n 3 + ( 1 ϑ j ϑ j 1 ϑ 1 ϑ 2 ϑ j n j = ϑ 3 n 3 = γ Now, since ϑ 1 + ϑ 2 + ϑ 3 = 1 we obtain γ = 1 n, where n = n 1 +n 2 +n 3 Then the estimates of the parameters are ϑ j = n j /n, j = 1,2,3 The second derivatives are 2 l ϑ 2 j = n j ϑ 2 j 2 l ϑ j ϑ i = n 3 ϑ 2 3 n 3 ϑ 2 3 The determinant of the second derivatives is positive for allϑand the element 2 l ϑ 2 1 is negative for allϑ, so also for ϑ Hence, there is a maximum at ϑ and MLE(ϑ inθ : L(ϑ;y = MLE(ϑ j = ϑ j = n j n ( 1 3 n1 ( 1 3 n2 ( n3 1 3 That is ϑ 0 = ( 1, 1, 1 T Now, we can calculateλ(y: λ(y = L( ϑ ( 0 ;y 1 n1 ( 1 n2 ( 1 n3 ( n1 ( n2 ( n3 L( ϑ;y = 3 3 ( 3 n n n n1 n1 ( n2 n2 ( n3 n3 = n n n 3n 1 3n 2 3n 3
7 25 HYPOTHESIS TESTING 133 That is 2log{λ(y} = 2 3 ( n n j log 3n j We reject H 0 at an approximate significance level α if the observed sample belongs to the critical regionr, where { 3 ( } 3nj R = y : 2 n j log χ 2 2;α n
8 134 CHAPTER 2 ELEMENTS OF STATISTICAL INFERENCE 255 Contingency tables We now show that the usual test for association in contingency tables is a generalized likelihood ratio test and that its asymptotic distribution is an example of Wilks theorem Suppose that we have anr c contingency table in which there arey ij individuals classified in row i and column j, when N individuals are classified independently Let ϑ ij be the corresponding probability that an individual is classified in row i and columnj, so thatϑ ij 0 and r c ϑ ij = 1 Then the variablesy ij have a multinomial distribution with parameters N and ϑ ij fori = 1,2,,r andj = 1,2,,c Example 240 In an experiment 150 patients were allocated to three groups of 45, 45 and 60 patients each Two groups were given a new drug at different dose levels and the third group received placebo The responses are as follows: Improved No difference Worse r i = c y ij Placebo Half dose Full dose c j = r y ij N = 150 We are interested in testing the hypothesis that the response to the drug does not depend on the dose level The null hypothesis is that the row and column classifications are independent, and the alternative is that they are dependent More precisely, the null hypothesis ish 0 : ϑ ij = a i b j for somea i > 0 andb j > 0, with r a i = 1 and c b j = 1, and the alternative is H 1 : ϑ ij a i b j for at least one pair(i,j The usual test statistic is given by X 2 = where E ij = R ic j N, (Y ij E ij 2 E ij,
9 25 HYPOTHESIS TESTING 135 R i = c Y ij and C j = r Y ij For large N, X 2 χ 2 (r 1(c 1 approximately when H 0 is true, and so we reject H 0 at the level of significanceαifx 2 χ 2 (r 1(c 1,α Now consider the generalized likelihood ratio test Then the likelihood is L(ϑ y = r N! c y ij! r c ϑ y ij ij = A r c ϑ y ij ij, where the coefficient A is the number of ways that N subjects can be divided in rc groups withy ij in theij-th group The log-likelihood is l(ϑ;y = log(a+ y ij log(ϑ ij We have to maximize this, subject to the constraint r c ϑ ij = 1 Let Σ represent the sum over all pairs(i,j except(r,c Then we may write l(ϑ y = log(a+σ y ij log(ϑ ij +y rc log(1 Σ ϑ ij Thus, for(i, j (r, c, solving the equation yields ˆϑ ij /y ij = ˆϑ rc /y rc = γ, say l = y ij y rc = y ij y rc = 0 ϑ ij ϑ ij 1 Σ ϑ ij ϑ ij ϑ rc Since r c ϑ ij = r c y ijγ = 1, we have γ = 1/N, so that the maximum likelihood estimate of ϑ ij is ˆϑ ij = y ij /N for i = 1,2,,r and j = 1,2,,c It follows that l( ϑ y = log(a+ y ij log ( yij N
10 136 CHAPTER 2 ELEMENTS OF STATISTICAL INFERENCE Now, underh 0, we haveϑ ij = a i b j and so l(ϑ y = log(a+ = log(a+ y ij {log(a i +log(b j } r i log(a i + c j log(b j Now, we maximize this subject to the constraints r a i = 1 and c b j = 1 That is, we maximize l(ϑ y = log(a+σ r i log(a i +r r log(a r +Σ c j log(b j +c c log(b c = log(a+σ r i log(a i +r r log(1 Σ a i +Σ c j log(b j +c c log(1 Σ b j Then, l = r i r r a i a i a r which, when compared to zero gives or However, Hence, γ = 1/N and so 1 = r i â i = r r â r â i r i = âr r r = γ â i = r i γ = Nγ â i = r i N Similarly, we obtain the ML estimates forb j as b j = c j /N Thus, the restricted maximum likelihood estimate ofϑ ij underh 0 is fori = 1,2,,r andj = 1,2,,c ϑ 0 ij = â i b j = r i C j N N = e ij/n It follows that l( ϑ 0 y = log(a+ y ij log ( eij N
11 25 HYPOTHESIS TESTING 137 Hence, we obtain { } L( ϑ 0 y 2log{λ(y} = 2log = 2{l( ϑ 0 y l( ϑ y} = 2 L( ϑ y y ij log ( yij e ij Here,p = rc 1 andp 0 = (r 1+(c 1 = r+c 2, and soν = (r 1(c 1 Therefore, by Wilks theorem, when H 0 is true andn is large, ( Yij 2 Y ij log E ij approx χ2 (r 1(c 1 The test statistics X 2 and 2log{λ(y} are asymptotically equivalent, that is, they differ by quantities that tend to zero as N To see this, first note thaty ij e ij is small relative toy ij and e ij when N is large Thus, since y ij = e ij +(y ij e ij, we may write (by Taylor series expansion of log(1+x around x, cut after second order term ( ( yij log = log 1+ y ij e ij (y ij e ij 1 (y ij e ij 2 e ij e ij e ij 2 e 2 ij Hence, we have y ij log ( yij e ij { (yij e ij {e ij +(y ij e ij } 1 e ij 2 (y ij e ij + 1 (y ij e ij 2 2 e ij } (y ij e ij 2 e 2 ij Since r c (y ij e ij = 0, it follows that 2 log{λ(y} (y ij e ij 2 e ij So we now see why we use thex 2 test statistic Example 241 continued Now we will test the hypothesis from the previous example The table ofe ij values is following
12 138 CHAPTER 2 ELEMENTS OF STATISTICAL INFERENCE Improved No difference Worse Placebo Half dose Full dose This gives X 2 obs = (y ij e ij 2 = 1417 e ij The critical value is χ 2 4;005 = 9488, hence there is no evidence to reject the null hypothesis saying that the different responses are independent of the levels of the new drug and placebo We obtain the same conclusion using the critical region derived from the Wilks theorem, which is { ( } yij R = y : 2 y ij log χ 2 ν;α e ij Here 2 y ij log ( yij e ij = 142
2.6.3 Generalized likelihood ratio tests
26 HYPOTHESIS TESTING 113 263 Generalized likelihood ratio tests When a UMP test does not exist, we usually use a generalized likelihood ratio test to verify H 0 : θ Θ against H 1 : θ Θ\Θ It can be used
More information2.5 Hypothesis Testing
118 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.5 Hypothesis Testing We assume that Y 1,...,Y n have a joint distribution which depends on the unknown parametersϑ = ϑ 1,...,ϑ p ) T. The set of all possible
More information2.3 Methods of Estimation
96 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.3 Methods of Estimation 2.3. Method of Moments The Method of Moments is a simple technique based on the idea that the sample moments are natural estimators
More information2.3 Analysis of Categorical Data
90 CHAPTER 2. ESTIMATION AND HYPOTHESIS TESTING 2.3 Analysis of Categorical Data 2.3.1 The Multinomial Probability Distribution A mulinomial random variable is a generalization of the binomial rv. It results
More information2.3.3 Method of Least Squares
.3. METHODS OF ESTIMATION 105.3.3 Method of Least Squares IfY 1,...,Y n are independent random variables, which have the same variance and higher-order moments, and, if each E(Y i ) is a linear function
More informationLecture 21: October 19
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 21: October 19 21.1 Likelihood Ratio Test (LRT) To test composite versus composite hypotheses the general method is to use
More informationNotes on the Multivariate Normal and Related Topics
Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions
More informationLecture 15. Hypothesis testing in the linear model
14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma
More informationMATH5745 Multivariate Methods Lecture 07
MATH5745 Multivariate Methods Lecture 07 Tests of hypothesis on covariance matrix March 16, 2018 MATH5745 Multivariate Methods Lecture 07 March 16, 2018 1 / 39 Test on covariance matrices: Introduction
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More information4.5.1 The use of 2 log Λ when θ is scalar
4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis
More informationHypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA
Hypothesis Testing Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA An Example Mardia et al. (979, p. ) reprint data from Frets (9) giving the length and breadth (in
More informationSummary of Chapters 7-9
Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationLikelihood-based inference with missing data under missing-at-random
Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014 Outline 1. Introduction. Parametric
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More information2014/2015 Smester II ST5224 Final Exam Solution
014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More informationAnswer Key for STAT 200B HW No. 8
Answer Key for STAT 200B HW No. 8 May 8, 2007 Problem 3.42 p. 708 The values of Ȳ for x 00, 0, 20, 30 are 5/40, 0, 20/50, and, respectively. From Corollary 3.5 it follows that MLE exists i G is identiable
More informationTopic 21 Goodness of Fit
Topic 21 Goodness of Fit Contingency Tables 1 / 11 Introduction Two-way Table Smoking Habits The Hypothesis The Test Statistic Degrees of Freedom Outline 2 / 11 Introduction Contingency tables, also known
More informationStat 579: Generalized Linear Models and Extensions
Stat 579: Generalized Linear Models and Extensions Mixed models Yan Lu March, 2018, week 8 1 / 32 Restricted Maximum Likelihood (REML) REML: uses a likelihood function calculated from the transformed set
More informationEstimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1
Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1
More informationStatistics 3858 : Contingency Tables
Statistics 3858 : Contingency Tables 1 Introduction Before proceeding with this topic the student should review generalized likelihood ratios ΛX) for multinomial distributions, its relation to Pearson
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationCategorical Variables and Contingency Tables: Description and Inference
Categorical Variables and Contingency Tables: Description and Inference STAT 526 Professor Olga Vitek March 3, 2011 Reading: Agresti Ch. 1, 2 and 3 Faraway Ch. 4 3 Univariate Binomial and Multinomial Measurements
More informationDiagnostics can identify two possible areas of failure of assumptions when fitting linear models.
1 Transformations 1.1 Introduction Diagnostics can identify two possible areas of failure of assumptions when fitting linear models. (i) lack of Normality (ii) heterogeneity of variances It is important
More informationLecture 26: Likelihood ratio tests
Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for
More informationLecture 23 Maximum Likelihood Estimation and Bayesian Inference
Lecture 23 Maximum Likelihood Estimation and Bayesian Inference Thais Paiva STA 111 - Summer 2013 Term II August 7, 2013 1 / 31 Thais Paiva STA 111 - Summer 2013 Term II Lecture 23, 08/07/2013 Lecture
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationMAS3301 Bayesian Statistics
MAS3301 Bayesian Statistics M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2008-9 1 11 Conjugate Priors IV: The Dirichlet distribution and multinomial observations 11.1
More informationAnswer Key for STAT 200B HW No. 7
Answer Key for STAT 200B HW No. 7 May 5, 2007 Problem 2.2 p. 649 Assuming binomial 2-sample model ˆπ =.75, ˆπ 2 =.6. a ˆτ = ˆπ 2 ˆπ =.5. From Ex. 2.5a on page 644: ˆπ ˆπ + ˆπ 2 ˆπ 2.75.25.6.4 = + =.087;
More informationNuisance parameters and their treatment
BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationHypothesis Testing: The Generalized Likelihood Ratio Test
Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 9 for Applied Multivariate Analysis Outline Addressing ourliers 1 Addressing ourliers 2 Outliers in Multivariate samples (1) For
More informationIntroductory Econometrics
Session 4 - Testing hypotheses Roland Sciences Po July 2011 Motivation After estimation, delivering information involves testing hypotheses Did this drug had any effect on the survival rate? Is this drug
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationAccurate directional inference for vector parameters
Accurate directional inference for vector parameters Nancy Reid October 28, 2016 with Don Fraser, Nicola Sartori, Anthony Davison Parametric models and likelihood model f (y; θ), θ R p data y = (y 1,...,
More informationMath 181B Homework 1 Solution
Math 181B Homework 1 Solution 1. Write down the likelihood: L(λ = n λ X i e λ X i! (a One-sided test: H 0 : λ = 1 vs H 1 : λ = 0.1 The likelihood ratio: where LR = L(1 L(0.1 = 1 X i e n 1 = λ n X i e nλ
More informationLikelihoods. P (Y = y) = f(y). For example, suppose Y has a geometric distribution on 1, 2,... with parameter p. Then the pmf is
Likelihoods The distribution of a random variable Y with a discrete sample space (e.g. a finite sample space or the integers) can be characterized by its probability mass function (pmf): P (Y = y) = f(y).
More informationClassical Inference for Gaussian Linear Models
Classical Inference for Gaussian Linear Models Surya Tokdar Gaussian linear model Data (z i, Y i ), i = 1,, n, dim(z i ) = p Model Y i = z T i Parameters: β R p, σ 2 > 0 Inference needed on η = a T β IID
More information2017 Financial Mathematics Orientation - Statistics
2017 Financial Mathematics Orientation - Statistics Written by Long Wang Edited by Joshua Agterberg August 21, 2018 Contents 1 Preliminaries 5 1.1 Samples and Population............................. 5
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationI i=1 1 I(J 1) j=1 (Y ij Ȳi ) 2. j=1 (Y j Ȳ )2 ] = 2n( is the two-sample t-test statistic.
Serik Sagitov, Chalmers and GU, February, 08 Solutions chapter Matlab commands: x = data matrix boxplot(x) anova(x) anova(x) Problem.3 Consider one-way ANOVA test statistic For I = and = n, put F = MS
More informationAccurate directional inference for vector parameters
Accurate directional inference for vector parameters Nancy Reid February 26, 2016 with Don Fraser, Nicola Sartori, Anthony Davison Nancy Reid Accurate directional inference for vector parameters York University
More informationComparing two independent samples
In many applications it is necessary to compare two competing methods (for example, to compare treatment effects of a standard drug and an experimental drug). To compare two methods from statistical point
More informationFoundations of Statistical Inference
Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2015 Julien Berestycki (University of Oxford) SB2a MT 2015 1 / 16 Lecture 16 : Bayesian analysis
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationLikelihood Ratio tests
Likelihood Ratio tests For general composite hypotheses optimality theory is not usually successful in producing an optimal test. instead we look for heuristics to guide our choices. The simplest approach
More informationMath 152. Rumbos Fall Solutions to Assignment #12
Math 52. umbos Fall 2009 Solutions to Assignment #2. Suppose that you observe n iid Bernoulli(p) random variables, denoted by X, X 2,..., X n. Find the LT rejection region for the test of H o : p p o versus
More informationSTAT 514 Solutions to Assignment #6
STAT 514 Solutions to Assignment #6 Question 1: Suppose that X 1,..., X n are a simple random sample from a Weibull distribution with density function f θ x) = θcx c 1 exp{ θx c }I{x > 0} for some fixed
More informationHomework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.
Stat 300A Theory of Statistics Homework 7: Solutions Nikos Ignatiadis Due on November 28, 208 Solutions should be complete and concisely written. Please, use a separate sheet or set of sheets for each
More informationMaximum likelihood estimation
Maximum likelihood estimation Guillaume Obozinski Ecole des Ponts - ParisTech Master MVA Maximum likelihood estimation 1/26 Outline 1 Statistical concepts 2 A short review of convex analysis and optimization
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationCherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants
18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009
More informationMathematics Qualifying Examination January 2015 STAT Mathematical Statistics
Mathematics Qualifying Examination January 2015 STAT 52800 - Mathematical Statistics NOTE: Answer all questions completely and justify your derivations and steps. A calculator and statistical tables (normal,
More informationSTT 843 Key to Homework 1 Spring 2018
STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationChapter 11. Hypothesis Testing (II)
Chapter 11. Hypothesis Testing (II) 11.1 Likelihood Ratio Tests one of the most popular ways of constructing tests when both null and alternative hypotheses are composite (i.e. not a single point). Let
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationChapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests
Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests Throughout this chapter we consider a sample X taken from a population indexed by θ Θ R k. Instead of estimating the unknown parameter, we
More informationPart 1.) We know that the probability of any specific x only given p ij = p i p j is just multinomial(n, p) where p k1 k 2
Problem.) I will break this into two parts: () Proving w (m) = p( x (m) X i = x i, X j = x j, p ij = p i p j ). In other words, the probability of a specific table in T x given the row and column counts
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationApplied Linear Statistical Methods
Applied Linear Statistical Methods (short lecturenotes) Prof. Rozenn Dahyot School of Computer Science and Statistics Trinity College Dublin Ireland www.scss.tcd.ie/rozenn.dahyot Hilary Term 2016 1. Introduction
More informationGARCH Models Estimation and Inference
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in
More informationSOLUTION FOR HOMEWORK 6, STAT 6331
SOLUTION FOR HOMEWORK 6, STAT 633. Exerc.7.. It is given that X,...,X n is a sample from N(θ, σ ), and the Bayesian approach is used with Θ N(µ, τ ). The parameters σ, µ and τ are given. (a) Find the joinf
More informationMaximum Likelihood Estimation
University of Pavia Maximum Likelihood Estimation Eduardo Rossi Likelihood function Choosing parameter values that make what one has observed more likely to occur than any other parameter values do. Assumption(Distribution)
More informationBIOS 2083 Linear Models c Abdus S. Wahed
Chapter 5 206 Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter
More informationFinal Examination Statistics 200C. T. Ferguson June 11, 2009
Final Examination Statistics 00C T. Ferguson June, 009. (a) Define: X n converges in probability to X. (b) Define: X m converges in quadratic mean to X. (c) Show that if X n converges in quadratic mean
More informationMaximum Likelihood Large Sample Theory
Maximum Likelihood Large Sample Theory MIT 18.443 Dr. Kempthorne Spring 2015 1 Outline 1 Large Sample Theory of Maximum Likelihood Estimates 2 Asymptotic Results: Overview Asymptotic Framework Data Model
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationLecture 10: Generalized likelihood ratio test
Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationStatistics. Statistics
The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,
More informationHypothesis Testing One Sample Tests
STATISTICS Lecture no. 13 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 12. 1. 2010 Tests on Mean of a Normal distribution Tests on Variance of a Normal
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationCategorical Data Analysis Chapter 3
Categorical Data Analysis Chapter 3 The actual coverage probability is usually a bit higher than the nominal level. Confidence intervals for association parameteres Consider the odds ratio in the 2x2 table,
More information10. Linear Models and Maximum Likelihood Estimation
10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationStatistics 135 Fall 2008 Final Exam
Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations
More informationStatistical Data Analysis Stat 3: p-values, parameter estimation
Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,
More informationDefinition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.
Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed
More informationInvestigation of an Automated Approach to Threshold Selection for Generalized Pareto
Investigation of an Automated Approach to Threshold Selection for Generalized Pareto Kate R. Saunders Supervisors: Peter Taylor & David Karoly University of Melbourne April 8, 2015 Outline 1 Extreme Value
More informationLecture 1 October 9, 2013
Probabilistic Graphical Models Fall 2013 Lecture 1 October 9, 2013 Lecturer: Guillaume Obozinski Scribe: Huu Dien Khue Le, Robin Bénesse The web page of the course: http://www.di.ens.fr/~fbach/courses/fall2013/
More informationMathematics Ph.D. Qualifying Examination Stat Probability, January 2018
Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationST3241 Categorical Data Analysis I Two-way Contingency Tables. 2 2 Tables, Relative Risks and Odds Ratios
ST3241 Categorical Data Analysis I Two-way Contingency Tables 2 2 Tables, Relative Risks and Odds Ratios 1 What Is A Contingency Table (p.16) Suppose X and Y are two categorical variables X has I categories
More informationDiscrete Multivariate Statistics
Discrete Multivariate Statistics Univariate Discrete Random variables Let X be a discrete random variable which, in this module, will be assumed to take a finite number of t different values which are
More informationLinear Models and Estimation by Least Squares
Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:
More informationTopic 19 Extensions on the Likelihood Ratio
Topic 19 Extensions on the Likelihood Ratio Two-Sided Tests 1 / 12 Outline Overview Normal Observations Power Analysis 2 / 12 Overview The likelihood ratio test is a popular choice for composite hypothesis
More informationGeneral Linear Model: Statistical Inference
Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter 4), least
More informationComposite Hypotheses and Generalized Likelihood Ratio Tests
Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve
More informationST495: Survival Analysis: Hypothesis testing and confidence intervals
ST495: Survival Analysis: Hypothesis testing and confidence intervals Eric B. Laber Department of Statistics, North Carolina State University April 3, 2014 I remember that one fateful day when Coach took
More informationHypothesis Testing Chap 10p460
Hypothesis Testing Chap 1p46 Elements of a statistical test p462 - Null hypothesis - Alternative hypothesis - Test Statistic - Rejection region Rejection Region p462 The rejection region (RR) specifies
More information