Chapter 4. Theory of Tests. 4.1 Introduction

Size: px
Start display at page:

Download "Chapter 4. Theory of Tests. 4.1 Introduction"

Transcription

1 Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule d, where d K, if x K d(x) = d A, if x A is called a non randomised test. One tries to choose K in such a way that the number of wrong decisions becomes as small as possible. We distinguish: Type I error: H 0 is correct, but is rejected (decision d K ). Type II error: H 1 is correct, but decision for H 0 (decision d A ). Decision for H 0 H 1 is correct is correct H 0 correct Type II error H 1 Type I error correct Given a boundary α (significance level) for the probability of committing an error of I. kind one tries to find a test which minimizes the probability of an error of II. kind. 37

2 38 CHAPTER 4. THEORY OF TESTS Def : (a) A measurable function ϕ : X [0, 1] is called a test function (a test). ϕ(x) is the probability for the decision d K, if x is the sample outcome. (b) ϕ is called an α-level test, if sup E θ [ϕ(x)] α. (4.1) θ H 0 (c) The probability of rejecting H 0 if P θ is the underlying distribution, β ϕ (θ) := P θ (d K ) = E θ [ϕ(x)] is called the power of the test. β ϕ : Θ [0, 1] is the power function of the test ϕ. The left hand side of (4.1) is called the size of the test ϕ. (d) If φ α is the set of all α level tests for the test problem (H 0, H 1 ), then ϕ 0 φ α is a most powerful test (MP-test) for an alternative θ H 1, if β ϕ0 (θ) β ϕ (θ) ϕ φ α and ϕ φ α is a uniformly most powerful test (UMP-test) for H 0 against H 1 of level α, if β ϕ (θ) = sup ϕ φ α β ϕ (θ) θ H 1. (4.2) (e) A test ϕ φ α is called unbiased, if β ϕ (θ) α θ H 1. (4.3) (f) A solution of (4.1), (4.2) and (4.3) is called a uniformly most powerful unbiased (UMPU-) α level test.

3 4.2. TEST OF A SIMPLE HYPOTHESIS AGAINST A SIMPLE ALTERNATIVE Test of a Simple Hypothesis against a Simple Alternative In this section Θ = {θ 0, θ 1 }, H 0 = {θ 0 }, H 1 = {θ 1 }. In this case there always exists a dominating measure, e.g. the measure µ = P θ0 + P θ1. The densities are denoted by f( ; θ 0 ) = f 0, f( ; θ 1 ) = f 1. Theorem (Foundamental Lemma of Neyman and Pearson): (a) Any test of the form 1, if f 1 (x) > kf 0 (x) ϕ(x) = γ(x), if f 1 (x) = kf 0 (x) 0, if f 1 (x) < kf 0 (x) (4.4) with k 0, 0 γ(x) 1, is a most powerful test of its size with 0 α 1 for H 0 : θ = θ 0 against H 1 : θ = θ 1. For k =, the test 1, if f 0 (x) = 0 ϕ(x) = 0, if f 0 (x) > 0 (4.5) is an M.P. test of its size α = 0 for H 0 against H 1. (b) For each level α, with 0 α 1, there exists a test of the form (4.4) or (4.5) with E θ0 [ϕ(x)] = α. Here γ(x) = γ (a constant). The constants k and γ, 0 γ 1, are determined by α = E θ0 [ϕ(x)] = P θ0 (f 1 (X) > kf 0 (X)) + γp θ0 (f 1 (X) = kf 0 (X))(4.6) Remark: A test of type (4.4) is called a Neyman-Pearson test with accompanying number k. Remarks:

4 40 CHAPTER 4. THEORY OF TESTS As the reasoning in the proof shows, the case µ{f 1 = kf 0 } = 0 leads to a non randomized test. Since the trivial α level test ϕ α with E θ0 [ϕ (X)] = E θ1 [ϕ (X)] = α does not have the form (4.4), it follows that E θ1 [ϕ(x)] α which means that a Neyman-Pearson test is unbiased. If there exists a sufficient statistic S for the family {f 0, f 1 }, then the NP test is a function of S. 4.3 Families With Monotone Likelihood Ratio In this section we consider the problem of testing one-sided hypotheses for Θ IR an interval. In the sequel let P = {P θ θ Θ} µ and assume, that for the µ densities f( ; θ) > 0 µ a.e. for all θ Θ holds. Def : We say that the family P has a monotone likelihood ratio (MLR) in the statistic T (X), if for θ 1 < θ 2, f( ; θ 1 ) f( ; θ 2 ), the ratio f(x; θ 2 )/f(x; θ 1 ) is a nondecreasing (nonincreasing) function of T (x) on {x X f(x; θ 1 ) > 0 f(x; θ 2 ) > 0}. Theorem 4.3.1: The familiy E 1 with density f(x; θ) = C(θ) exp{q(θ)t (x)}h(x) has for Q nondecreasing (nonincreasing) a monotone likelihood ratio in T (X). Remark: With the reparametrization λ := Q(θ) this property can always be achieved. Theorem 4.3.2: Let the family F = {f( ; θ) θ Θ} have a monotone likelihood ration in T (x). For testing H 0 : θ θ 0 against H 1 : θ > θ 0 any test of the form

5 4.4. UNBIASED TESTS 41 1, if T (x) > c ϕ(x) = γ, if T (x) = c 0, if T (x) < c (4.7) has a nondecreasing power function and is UMP of its size E θ0 [ϕ(x)] = α (provided the size α > 0). Remark: For symmetry reasons Theorem yields also a UMP test for the test problem H 0 : θ θ 0 against H 1 : θ < θ 0. In general, the results of the above Theorem cannot be extended to two-sided problems. One exception is the family E 1 : Theorem 4.3.3: For the family E 1 there exists a UMP test of the hypothesis H 0 : θ θ 1 or θ θ 2 (θ 1 < θ 2 against H 1 : θ 1 < θ < θ 2 that is of the form 1, if c 1 < T (X) < c 2 ϕ(x) = γ i, if T (X) = c i, i = 1, 2 0, if T (X) < c 1 or T (X) > c 2, where the c s and the γ s are given by E θ1 [ϕ(x)] = E θ2 [ϕ(x)] = α. Remark: UMP tests for H 0 : θ 1 θ θ 1 or H 0 : θ = θ 0 do not exist, even in the family E Unbiased Tests Unbiased tests we encountered already in Def They have the property that β ϕ (θ) α for θ Θ 0 and β ϕ (θ) α for θ Θ α Similar Tests Def : (1) Let U α φ α be the class of all unbiased sized tests of H 0.

6 42 CHAPTER 4. THEORY OF TESTS (2) A test ϕ is said to be α similar on a subset Θ Θ, if β ϕ (θ) = E θ [ϕ(x)] = α for θ Θ. (3) A test is said to be similar on a set Θ Θ, if its α similar for some α [0, 1]. Theorem 4.4.1: Let β ϕ (θ) be continuous in θ for any ϕ. If ϕ U α for H 0 against H 1, then it is α similar on the boundary Λ = Θ 0 Θ 1. Def : A test ϕ that is UMP among all α similar tests on the boundary Λ is said to ba a UMP α similar test. Theorem 4.4.2: Let the power function β of every test ϕ of H 0 against H 1 be continuous in θ. Then a UMP α similar test is UMP unbiased, provided its size is α. Remark: The continuity of β ϕ is not always easy to show Local MP Unbiased Tests To test the hypothesis H 0 : θ θ 0 we try to find a locally optimal unbiased test which, in a neighbourhood of θ 0 fulfils the following conditions: (0) β ϕ is twice continuously differentiable with respect to θ (1) β ϕ (θ 0 ) = α (2) β ϕ(θ 0 ) = 0 (3) β ϕ(θ 0 ) max. Theorem (Locally MP Unbiased Tests): Let f θ F = {f θ θ Θ} be twice continuously differentiable in θ. If the power function of a test ϕ, given by 1, f(x; θ0 ) > k 0 f(x; θ 0 ) + k 1 f(x; θ 0 ) ϕ(x; k 0, k 1, c) = γ, f(x; θ0 ) = k 0 f(x; θ 0 ) + k 1 f(x; θ 0 ) 0, f(x; θ0 ) < k 0 f(x; θ 0 ) + k 1 f(x; θ 0 )

7 4.4. UNBIASED TESTS 43 fulfils the conditions (0), (1) and (2), then also (3) is fulfilled. The question whether one can always find constants k 0, k 1 and γ such that (1) and (2) holds, remains open, excepting exponential families. Theorem 4.4.4: Let P = E 1 with µ density f(x; θ) = C(θ)e θt (x) h(x). If the power function of the test 1, T (X) / [T 1, T 2 ] ϕ(x; T 1, T 2, c) = c, T (X) = T 1 or T (X) = T 2 ] 0, T (X) [T 1, T 2 ] fulfils (1) and (2), then also (3). (4.8) UMP Unbiased Tests in One-Parameter Exponential Families Ref. Lehmann, Testing... (1997), pp. 134 ff. In 4.3 we have seen that UMP-test for hypotheses (i) H 0 : θ θ 0 against H 1 : θ > θ 0 or (ii) H 0 : θ θ 1 or θ θ 2 against H 1 : θ 1 < θ < θ 2 exist, but not for (iii) H 0 : θ 1 θ θ 2 against H 1 : θ < θ 1 or θ > θ 2. Theorem 4.4.5: Let P = E 1 with µ density f(x; θ) = C(θ)e θt (x) h(x). Then there exists a UMP Unbiased test, which is given by (4.8), where the constants T 1 and T 2 and γ are given by E θ1 [ϕ(x)] = E θ2 [ϕ(x)] = α. (4.9) Invariant Tests Def :

8 44 CHAPTER 4. THEORY OF TESTS (1) A group G of transformations on X leves the hypothesis testing problem invariant if it leaves both {P θ θ Θ 0 } and {P θ θ Θ 1 }invariant. (2) We say that ϕ is invariant under G if ϕ(g(x)) = ϕ(x) for all x X and g G. (3) A statistic T is (a) invariant under G, if T (g(x)) = T (x) x X and g G. (b) maximal invariant, if T (x 1 ) = T (x 2 ) x 1 g G. = g(x 2 ) for some Def : Let ΦI α donote the set of all invariant tests of size α with respect to G for H 0 : θ Θ 0 against H 1 : θ Θ 1. If there exists a UMP test in ΦI α, then we call it a UMP invariant test of H 0 against H 1. Theorem 4.4.6: Let T be maximal invariant with respect to G. ϕ is invariant under G if and only if ϕ is a function of T. Then Remark: If a hypothesis testing problem is invariant under a group G, it suffices to restrict attention to functions of maximal invariant statistics T. 4.5 Likelihood Ratio Tests Let P = {P θ θ Θ} µ and Θ = Θ 0 + Θ 1. In many cases UMP tests do not exist, and where they exist, the approach can only be applied to particular families of distributions. The Likelihood Ratio test (LR) is an intuitive and plausible procedure which often leads to UMPU tests. Def : For testing H 0 against H 1, a test of the form: reject H 0 if and only if λ(x) > c, where c is some constant and λ(x) = sup f(x 1,..., x n ; θ) θ Θ sup f(x 1,..., x n ; θ) = θ Θ 0 f(x 1,..., x n ; ˆθ ML ) f(x 1,..., x n ; θ)

9 4.6. ASYMPTOTIC TESTS 45 is called a likelihood ratio test. Here ˆθ ML is the unrestricted Maximum Likelihood estimator, θ is the MLestimator under the restriction θ Θ 0. The constant c is determined from the size restriction sup P θ (x λ(x) > c) = α. θ Θ 0 It can easily be seen that for testing a simple hypthesis against a simple alternative to a given size α(0 α 1) nonrandomized Neyman-Pearson tests and LR tests are equivalent, if they exist; the LR test for θ Θ 0 against θ Θ 1 is a function of every sufficient statistic S for θ (see Theorem resp ). Theorem 4.5.1: Let the regularity conditions of Theorem (Cramér- Rao inequality) hold. Then under H 0 the statistic2 ln λ(x) is asymptotically distributed as a χ 2 random variable with degrees of freedom equal to the difference between the number of independent parameters in Θ and the number in Θ Asymptotic Tests For see Buse, The American Statistician, 1982, 36, pp Let Θ IR k. H 0 : h(θ) = 0 where h : IR k IR r (r k) Wald-Test See: Transactions of the American Mathematical Society 1943, pp Let R θ := h(θ) with rank R θ T θ = r and W = h(ˆθ ML ) T [ RˆθML [I(ˆθ ML )] 1 RˆθML ] h(ˆθml ),

10 46 CHAPTER 4. THEORY OF TESTS where ˆθ ML is the unrestricted ML estimator. Under H 0 W is asymptotically χ 2 (r) distributed and the test is of the form 1, W > c ϕ(x) = for a certain constant c. 0, W < c Lagrange Multiplier Test It is based on the Lagrange multiplier approach: Φ(θ; η) = l(θ; η) + η T h(θ), where η is the Lagrange multiplier. Let ˆθ (r) = arg sup θ H 0 L(θ; X 1,..., X n ). The test statistic is where LM = (ˆθ (r) ) T [I(ˆθ (r) )] 1 (ˆθ (r) ) = Ψ(ˆθ (r) ) T I(ˆθ (r) ) 1 Ψ(ˆθ (r) ), (ˆθ (r) ) = Ψ(ˆθ (r) ) is the score function log f θ at θ = ˆθ (r). Under H 0 LM has an asymptotic χ 2 (r) distribution and the test statistic is distributed as in Likelihood Ratio Test It works as described in 4.5, where for determining the constant c the asymptotic χ 2 (r) distribution is being used. 4.7 Goodness of Fit Tests We consider the general testing problem H 0 : P P 0 against H 1 : P P 1, where P = P 0 +P 1. Def A sequence of tests (ϕ n (X)) n IN is called consistent for the testing problem H 0 : P P 0 against H 1 : P P 1, if lim β 1, P ϕ n n (x) = X P 1 0, P X P 0

11 4.7. GOODNESS OF FIT TESTS 47 For consistent tests the power function converges to the ideal power function, which for the onesided problem H 0 : θ θ 0 against H 1 : θ > θ 0 is given by the Heaviside function 1, θ > θ 0 H θ0 (θ) = 0, θ θ 0 At the first instant we look at the (two-sided) testing problem H 0 : P X = P 0 with c.d.f. F 0 against H 1 : P X P 0. Here the class P is the class of all distributions with densities with respect to Lebesque measure. In H 0 the density F 0 has to be specified completely. According to the Gliwenko-Cantinelli Lemma (Theorem 1.2) the empirical c.d.f. F n converges almost surely univormly to F 0. For the maximal difference n := sup F n (x) F 0 (x) x IR the following result holds: Theorem (Kolmogorov): Let the c.d.f. F 0 be continuous. Then lim P ( n n z) = H(z), n where ( 1) k e 2k2 z 2 z > 0 H(z) = k= 0 z 0. The limit distribution H does obviously not depend on F 0. Hence the asymptotic test 1 n n > k ϕ(x) = 0 n n < k with k equal to the (1 α) Quantil of H is distribution free. The Gliwenko- Cantelli lemma ensures consistency.

12 48 CHAPTER 4. THEORY OF TESTS A further asymptotic goodness of fit test is the so-called χ 2 goodness of fit test. It is based on a comparison between observed and under H 0 expected frequencies. Starting point is the following asymptotic result. Theorem 4.7.2: Let the random vector (X 1,..., X k ) have a polynomial M(n; p 1,..., p k ) distribution with 0 < p i < 1, p i = 1. Then the k statistic i=1 k X 2 (X i np i ) 2 =, X X k = n, np i i=1 is asymptotically χ 2 (k 1) distributed. The χ 2 -test is 1 X 2 > c ϕ(x) = 0 X 2 < c where c is the (1 α) Quantile of the χ 2 (k 1) distribution. Some rules of prudence are adequate when this test is applied in practice. For F 0 continuous an appropriate division of IR into k classes is necessary such that p i = F 0 (x i ) F 0 (x i 1). The p i should be approximately equal and as a rule of thumb np i 5 for all i = 1,..., k is recommended. The test is also applicable if the parameteres of some F θ are estimated by ML (with a corresponding reduction of degrees of freedom).

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Lecture 21. Hypothesis Testing II

Lecture 21. Hypothesis Testing II Lecture 21. Hypothesis Testing II December 7, 2011 In the previous lecture, we dened a few key concepts of hypothesis testing and introduced the framework for parametric hypothesis testing. In the parametric

More information

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007 Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two

More information

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Testing Statistical Hypotheses

Testing Statistical Hypotheses E.L. Lehmann Joseph P. Romano Testing Statistical Hypotheses Third Edition 4y Springer Preface vii I Small-Sample Theory 1 1 The General Decision Problem 3 1.1 Statistical Inference and Statistical Decisions

More information

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson

More information

Some General Types of Tests

Some General Types of Tests Some General Types of Tests We may not be able to find a UMP or UMPU test in a given situation. In that case, we may use test of some general class of tests that often have good asymptotic properties.

More information

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4 STA 73: Inference Notes. Neyman-Pearsonian Classical Hypothesis Testing B&D 4 1 Testing as a rule Fisher s quantification of extremeness of observed evidence clearly lacked rigorous mathematical interpretation.

More information

Lecture 17: Likelihood ratio and asymptotic tests

Lecture 17: Likelihood ratio and asymptotic tests Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Composite Hypotheses and Generalized Likelihood Ratio Tests

Composite Hypotheses and Generalized Likelihood Ratio Tests Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

Chapter 3. Point Estimation. 3.1 Introduction

Chapter 3. Point Estimation. 3.1 Introduction Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These

More information

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004 Hypothesis Testing BS2 Statistical Inference, Lecture 11 Michaelmas Term 2004 Steffen Lauritzen, University of Oxford; November 15, 2004 Hypothesis testing We consider a family of densities F = {f(x; θ),

More information

Hypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test.

Hypothesis Testing. A rule for making the required choice can be described in two ways: called the rejection or critical region of the test. Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two hypotheses:

More information

LECTURE NOTES 57. Lecture 9

LECTURE NOTES 57. Lecture 9 LECTURE NOTES 57 Lecture 9 17. Hypothesis testing A special type of decision problem is hypothesis testing. We partition the parameter space into H [ A with H \ A = ;. Wewrite H 2 H A 2 A. A decision problem

More information

λ(x + 1)f g (x) > θ 0

λ(x + 1)f g (x) > θ 0 Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where

More information

2014/2015 Smester II ST5224 Final Exam Solution

2014/2015 Smester II ST5224 Final Exam Solution 014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of

More information

Testing Statistical Hypotheses

Testing Statistical Hypotheses E.L. Lehmann Joseph P. Romano, 02LEu1 ttd ~Lt~S Testing Statistical Hypotheses Third Edition With 6 Illustrations ~Springer 2 The Probability Background 28 2.1 Probability and Measure 28 2.2 Integration.........

More information

Hypothesis testing: theory and methods

Hypothesis testing: theory and methods Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable

More information

Theory of Statistical Tests

Theory of Statistical Tests Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H

More information

Lecture 12 November 3

Lecture 12 November 3 STATS 300A: Theory of Statistics Fall 2015 Lecture 12 November 3 Lecturer: Lester Mackey Scribe: Jae Hyuck Park, Christian Fong Warning: These notes may contain factual and/or typographic errors. 12.1

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

10. Composite Hypothesis Testing. ECE 830, Spring 2014

10. Composite Hypothesis Testing. ECE 830, Spring 2014 10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters

More information

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES 557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES Example Suppose that X,..., X n N, ). To test H 0 : 0 H : the most powerful test at level α is based on the statistic λx) f π) X x ) n/ exp

More information

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma Theory of testing hypotheses X: a sample from a population P in P, a family of populations. Based on the observed X, we test a

More information

Charles Geyer University of Minnesota. joint work with. Glen Meeden University of Minnesota.

Charles Geyer University of Minnesota. joint work with. Glen Meeden University of Minnesota. Fuzzy Confidence Intervals and P -values Charles Geyer University of Minnesota joint work with Glen Meeden University of Minnesota http://www.stat.umn.edu/geyer/fuzz 1 Ordinary Confidence Intervals OK

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

8 Testing of Hypotheses and Confidence Regions

8 Testing of Hypotheses and Confidence Regions 8 Testing of Hypotheses and Confidence Regions There are some problems we meet in statistical practice in which estimation of a parameter is not the primary goal; rather, we wish to use our data to decide

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

Statistical hypothesis testing The parametric and nonparametric cases. Madalina Olteanu, Université Paris 1

Statistical hypothesis testing The parametric and nonparametric cases. Madalina Olteanu, Université Paris 1 Statistical hypothesis testing The parametric and nonparametric cases Madalina Olteanu, Université Paris 1 2016-2017 Contents 1 Parametric hypothesis testing 3 1.1 An introduction on statistical hypothesis

More information

Statistics and econometrics

Statistics and econometrics 1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample

More information

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple

More information

Exercises Chapter 4 Statistical Hypothesis Testing

Exercises Chapter 4 Statistical Hypothesis Testing Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Chapter 9: Hypothesis Testing Sections

Chapter 9: Hypothesis Testing Sections Chapter 9: Hypothesis Testing Sections 9.1 Problems of Testing Hypotheses 9.2 Testing Simple Hypotheses 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.6 Comparing the Means of Two

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

STAT 801: Mathematical Statistics. Hypothesis Testing

STAT 801: Mathematical Statistics. Hypothesis Testing STAT 801: Mathematical Statistics Hypothesis Testing Hypothesis testing: a statistical problem where you must choose, on the basis o data X, between two alternatives. We ormalize this as the problem o

More information

On the GLR and UMP tests in the family with support dependent on the parameter

On the GLR and UMP tests in the family with support dependent on the parameter STATISTICS, OPTIMIZATION AND INFORMATION COMPUTING Stat., Optim. Inf. Comput., Vol. 3, September 2015, pp 221 228. Published online in International Academic Press (www.iapress.org On the GLR and UMP tests

More information

f(y θ) = g(t (y) θ)h(y)

f(y θ) = g(t (y) θ)h(y) EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,

More information

Lecture 26: Likelihood ratio tests

Lecture 26: Likelihood ratio tests Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses. Stat 300A Theory of Statistics Homework 7: Solutions Nikos Ignatiadis Due on November 28, 208 Solutions should be complete and concisely written. Please, use a separate sheet or set of sheets for each

More information

To appear in The American Statistician vol. 61 (2007) pp

To appear in The American Statistician vol. 61 (2007) pp How Can the Score Test Be Inconsistent? David A Freedman ABSTRACT: The score test can be inconsistent because at the MLE under the null hypothesis the observed information matrix generates negative variance

More information

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Florian Pelgrin HEC September-December 2010 Florian Pelgrin (HEC) Constrained estimators September-December

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then

More information

2.1.3 The Testing Problem and Neave s Step Method

2.1.3 The Testing Problem and Neave s Step Method we can guarantee (1) that the (unknown) true parameter vector θ t Θ is an interior point of Θ, and (2) that ρ θt (R) > 0 for any R 2 Q. These are two of Birch s regularity conditions that were critical

More information

Lecture 23: UMPU tests in exponential families

Lecture 23: UMPU tests in exponential families Lecture 23: UMPU tests in exponential families Continuity of the power function For a given test T, the power function β T (P) is said to be continuous in θ if and only if for any {θ j : j = 0,1,2,...}

More information

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Maximum Likelihood Tests and Quasi-Maximum-Likelihood Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.

More information

Greene, Econometric Analysis (6th ed, 2008)

Greene, Econometric Analysis (6th ed, 2008) EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived

More information

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques Vol:7, No:0, 203 Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques D. A. Farinde International Science Index, Mathematical and Computational Sciences Vol:7,

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

ECE 275B Homework # 1 Solutions Winter 2018

ECE 275B Homework # 1 Solutions Winter 2018 ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,

More information

More Empirical Process Theory

More Empirical Process Theory More Empirical Process heory 4.384 ime Series Analysis, Fall 2008 Recitation by Paul Schrimpf Supplementary to lectures given by Anna Mikusheva October 24, 2008 Recitation 8 More Empirical Process heory

More information

Hypothesis Testing - Frequentist

Hypothesis Testing - Frequentist Frequentist Hypothesis Testing - Frequentist Compare two hypotheses to see which one better explains the data. Or, alternatively, what is the best way to separate events into two classes, those originating

More information

ECE 275B Homework # 1 Solutions Version Winter 2015

ECE 275B Homework # 1 Solutions Version Winter 2015 ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2

More information

1. Fisher Information

1. Fisher Information 1. Fisher Information Let f(x θ) be a density function with the property that log f(x θ) is differentiable in θ throughout the open p-dimensional parameter set Θ R p ; then the score statistic (or score

More information

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

ML Testing (Likelihood Ratio Testing) for non-gaussian models

ML Testing (Likelihood Ratio Testing) for non-gaussian models ML Testing (Likelihood Ratio Testing) for non-gaussian models Surya Tokdar ML test in a slightly different form Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) = {θ : l x (θ) max θ Θ l

More information

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Eric Zivot Winter 013 1 Wald, LR and LM statistics based on generalized method of moments estimation Let 1 be an iid sample

More information

Define characteristic function. State its properties. State and prove inversion theorem.

Define characteristic function. State its properties. State and prove inversion theorem. ASSIGNMENT - 1, MAY 013. Paper I PROBABILITY AND DISTRIBUTION THEORY (DMSTT 01) 1. (a) Give the Kolmogorov definition of probability. State and prove Borel cantelli lemma. Define : (i) distribution function

More information

Chapter 6 Testing. 1. Neyman Pearson Tests. 2. Unbiased Tests; Conditional Tests; Permutation Tests

Chapter 6 Testing. 1. Neyman Pearson Tests. 2. Unbiased Tests; Conditional Tests; Permutation Tests Chapter 6 Testing. eyman Pearson Tests 2. Unbiased Tests; Conditional Tests; Permutation Tests 2. Unbiased tests 2.2 Application to -parameter exponential families 2.3 UMPU tests for families with nuisance

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

Asymptotics for Nonlinear GMM

Asymptotics for Nonlinear GMM Asymptotics for Nonlinear GMM Eric Zivot February 13, 2013 Asymptotic Properties of Nonlinear GMM Under standard regularity conditions (to be discussed later), it can be shown that where ˆθ(Ŵ) θ 0 ³ˆθ(Ŵ)

More information

4 Invariant Statistical Decision Problems

4 Invariant Statistical Decision Problems 4 Invariant Statistical Decision Problems 4.1 Invariant decision problems Let G be a group of measurable transformations from the sample space X into itself. The group operation is composition. Note that

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

The outline for Unit 3

The outline for Unit 3 The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.

More information

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018 Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from

More information

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics MAS 713 Chapter 8 Previous lecture: 1 Bayesian Inference 2 Decision theory 3 Bayesian Vs. Frequentist 4 Loss functions 5 Conjugate priors Any questions? Mathematical Statistics

More information

Hypothesis Testing (May 30, 2016)

Hypothesis Testing (May 30, 2016) Ch. 5 Hypothesis Testing (May 30, 2016) 1 Introduction Inference, so far as we have seen, often take the form of numerical estimates, either as single points as confidence intervals. But not always. In

More information

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes

More information

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell, Recall The Neyman-Pearson Lemma Neyman-Pearson Lemma: Let Θ = {θ 0, θ }, and let F θ0 (x) be the cdf of the random vector X under hypothesis and F θ (x) be its cdf under hypothesis. Assume that the cdfs

More information

14.30 Introduction to Statistical Methods in Economics Spring 2009

14.30 Introduction to Statistical Methods in Economics Spring 2009 MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Lecture 2 Machine Learning Review

Lecture 2 Machine Learning Review Lecture 2 Machine Learning Review CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago March 29, 2017 Things we will look at today Formal Setup for Supervised Learning Things

More information

Optimal Tests of Hypotheses (Hogg Chapter Eight)

Optimal Tests of Hypotheses (Hogg Chapter Eight) Optimal Tests of Hypotheses Hogg hapter Eight STAT 406-0: Mathematical Statistics II Spring Semester 06 ontents Most Powerful Tests. Review of Hypothesis Testing............ The Neyman-Pearson Lemma............3

More information

Lecture 16 November Application of MoUM to our 2-sided testing problem

Lecture 16 November Application of MoUM to our 2-sided testing problem STATS 300A: Theory of Statistics Fall 2015 Lecture 16 November 17 Lecturer: Lester Mackey Scribe: Reginald Long, Colin Wei Warning: These notes may contain factual and/or typographic errors. 16.1 Recap

More information

5.1 Uniformly Most Accurate Families of Confidence

5.1 Uniformly Most Accurate Families of Confidence Chapter 5 Confidence Estimation Let (X, B X, P ), P {P θ θ Θ} be sample space and {Θ, B Θ, µ} be measure space with some σ finite measure µ. 5.1 Uniformly Most Accurate Families of Confidence Sets Def.

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Math 152. Rumbos Fall Solutions to Assignment #12

Math 152. Rumbos Fall Solutions to Assignment #12 Math 52. umbos Fall 2009 Solutions to Assignment #2. Suppose that you observe n iid Bernoulli(p) random variables, denoted by X, X 2,..., X n. Find the LT rejection region for the test of H o : p p o versus

More information

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That Statistics Lecture 2 August 7, 2000 Frank Porter Caltech The plan for these lectures: The Fundamentals; Point Estimation Maximum Likelihood, Least Squares and All That What is a Confidence Interval? Interval

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

Assumptions of classical multiple regression model

Assumptions of classical multiple regression model ESD: Recitation #7 Assumptions of classical multiple regression model Linearity Full rank Exogeneity of independent variables Homoscedasticity and non autocorrellation Exogenously generated data Normal

More information

Testing and Model Selection

Testing and Model Selection Testing and Model Selection This is another digression on general statistics: see PE App C.8.4. The EViews output for least squares, probit and logit includes some statistics relevant to testing hypotheses

More information