Introductory Econometrics. Review of statistics (Part II: Inference)

Similar documents
LECTURE 5 HYPOTHESIS TESTING

Preliminary Statistics. Lecture 5: Hypothesis Testing

Comparison of Bayesian and Frequentist Inference

Chapter 7: Hypothesis Testing

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

Chapter 5: HYPOTHESIS TESTING

6.4 Type I and Type II Errors

Tests about a population mean

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline)

Mathematical statistics

Lectures 5 & 6: Hypothesis Testing

Frequentist Statistics and Hypothesis Testing Spring

Visual interpretation with normal approximation

Hypothesis Testing The basic ingredients of a hypothesis test are

Introductory Econometrics

Mathematical Statistics

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

EC2001 Econometrics 1 Dr. Jose Olmo Room D309

Hypothesis Testing. ECE 3530 Spring Antonio Paiva

ECO220Y Review and Introduction to Hypothesis Testing Readings: Chapter 12

Probability and Statistics

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Topic 15: Simple Hypotheses

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Quantitative Methods for Economics, Finance and Management (A86050 F86050)

Hypotheses Test Procedures. Is the claim wrong?

Finansiell Statistik, GN, 15 hp, VT2008 Lecture 10-11: Statistical Inference: Hypothesis Testing

ME3620. Theory of Engineering Experimentation. Spring Chapter IV. Decision Making for a Single Sample. Chapter IV

Chapter Three. Hypothesis Testing

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Mathematical statistics

hypothesis a claim about the value of some parameter (like p)

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

PHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1

Statistical Inference. Hypothesis Testing

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Hypothesis Testing: One Sample

280 CHAPTER 9 TESTS OF HYPOTHESES FOR A SINGLE SAMPLE Tests of Statistical Hypotheses

Sampling Distributions

Topic 17: Simple Hypotheses

Hypothesis Tests Solutions COR1-GB.1305 Statistics and Data Analysis

LECTURE 5. Introduction to Econometrics. Hypothesis testing

STAT Chapter 8: Hypothesis Tests

Hypothesis testing. Data to decisions

Statistical Inference

10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing

MTMS Mathematical Statistics

BLAST: Target frequencies and information content Dannie Durand

Announcements. Unit 3: Foundations for inference Lecture 3: Decision errors, significance levels, sample size, and power.

F79SM STATISTICAL METHODS

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

Hypothesis Testing. ) the hypothesis that suggests no change from previous experience

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

Chapter 9. Hypothesis testing. 9.1 Introduction

The problem of base rates

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Statistical Preliminaries. Stony Brook University CSE545, Fall 2016

hypothesis testing 1

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Introduction to Statistical Inference

Exam 2 Practice Questions, 18.05, Spring 2014

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

Naive Bayes classification

Fin285a:Computer Simulations and Risk Assessment Section 2.3.2:Hypothesis testing, and Confidence Intervals

Computational Perception. Bayesian Inference

Information Retrieval

In any hypothesis testing problem, there are two contradictory hypotheses under consideration.

For use only in [the name of your school] 2014 S4 Note. S4 Notes (Edexcel)

Hypothesis tests

MA : Introductory Probability

Introduction to Statistical Inference

Maximum-Likelihood Estimation: Basic Ideas

Hypothesis Testing. Hypothesis: conjecture, proposition or statement based on published literature, data, or a theory that may or may not be true

The Logit Model: Estimation, Testing and Interpretation

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

(1) Introduction to Bayesian statistics

HYPOTHESIS TESTING. Hypothesis Testing

Problems ( ) 1 exp. 2. n! e λ and

Parameter Estimation, Sampling Distributions & Hypothesis Testing

Statistical testing. Samantha Kleinberg. October 20, 2009

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Introduction to Statistics

Statistical Inference. Why Use Statistical Inference. Point Estimates. Point Estimates. Greg C Elvers

Statistical Inference

Section 10.1 (Part 2 of 2) Significance Tests: Power of a Test

Test of Hypothesis for Small and Large Samples and Test of Goodness of Fit

CHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:

Hypothesis Testing One Sample Tests

Statistical Process Control (contd... )

Introduction 1. STA442/2101 Fall See last slide for copyright information. 1 / 33

Unit 19 Formulating Hypotheses and Making Decisions

Introduction to Statistical Hypothesis Testing

Confidence Intervals and Hypothesis Tests

MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL

Lecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya

Chapters 10. Hypothesis Testing

Two Sample Hypothesis Tests

Transcription:

Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16

Null and alternative hypotheses Usually, we have two competing hypotheses, and we want to draw a conclusion, based on the data, as to which of the hypotheses is true. Null hypothesis, denoted as H 0 : A hypothesis that is held to be true, unless the data provides a sufficient evidence against it. Alternative hypothesis, denoted as H 1 : A hypothesis against which the null is tested. It is held to be true if the null is found false. The two hypotheses are not treated symmetrically. H 0 is favored in the sense that it is only rejected if there is strong evidence against it. The two hypotheses must be disjoint: it should be the case that either H 0 is true or H 1 but never together simultaneously. 2/16

Heads or Tails? We believe that a coin is fair, unless we see enough evidence against it. In the coin tossing example, let p be the probability of Heads. Our null hypothesis could be H 0 : p = 1 2 and our alternative hypothesis could be H 1 : p = 1 2. We may believe that the coin is fair, but suspect that it is biased towards heads if it is biased. In this case, we may pick H 0 : p = 1 2 and H 1 : p > 1 2. 3/16

Hypothesis testing as proof by contradiction* Usually, the case we are interested in or want to prove is stated as H 1. The econometrician has to prove that his assertion (H 1 ) is true by showing that the data rejects H 0. In the hypothesis testing, we assume the null hypothesis being true and we want to prove that our data contradicts to the hypothesis. 4/16

Hypothesis testing procedure Suppose we are concerned with a parameter θ R (e.g. the population mean of some random variable for which we have repeated observations). We consider hypotheses like H 0 : θ Ξ 0 and H 1 : θ Ξ 1 with Ξ 0 Ξ 1 =. Ξ 0 Ξ 1 is the maintained hypothesis. We consider a test statistic T, which is a function of the data. T should be informative about the value of θ. T typically has a known distribution under H 0. Then we choose a significance level α, with 0 < α < 1. It is also called the size of the test. By convention, α is chosen to be a small number. 5/16

Hypothesis testing procedure Then one rejects H 0 if the test statistic falls into a critical region Γ α, which typically depends on α. Γ α is usually a set of possible values of the test statistic that contains the test statistic with probability α, under H 0, i.e. P [T Γ α ] = α under H 0. Note that this requires knowledge about the distribution of T under H 0. If the realized test statistic is in the critical region, T Γ α, we reject H 0 in favor of H 1. Otherwise, we fail to reject H 0. If α is small and T Γ α, it means that the sample is unlikely if H 0 is true and so we have evidence against it, there are two possibilities: we observed a very small-probability event or our assumption (H 0 ) is wrong. 6/16

Hypothesis testing as proof by contradiction* If we accept the principle that a very small-probability event cannot happen in the current state of the world, we accept the second reasoning: our assumption (H 0 ) is wrong. If T / Γ α, there is no significant contradiction between data and the null hypothesis. On the other hand, if T Γ α, there is a significant contradiction and we reject H 0 like what we conclude in proof by contradiction. Fail to reject H 0 is a better statement than accept H 0 since if we find T / Γ α, we did not prove anything and our test is non-conclusive, since we did not find a contradiction. The decision depends on the significance level α: larger values of α correspond to bigger critical regions Γ α. It is easier to reject the null for larger values of α. 7/16

A simple example Consider a normal population with mean µ and variance σ 2 and a random sample from the population. Suppose we know σ 2 for now. Consider a one-sided test: H 0 : µ = 0 against H 1 : µ > 0. Consider T = X σ/ n, which is a N (0, 1) random variable under H 0. H 1 is more likely to be true if T is large. Consider Γ α = [z 1 α, ), where z 1 α is the (1 α)-th quantile of N (0, 1). Under H 0, P [T Γ α ] = α. And notice that Γ α is a sub-interval of Γ α if α < α. 8/16

Summary of the steps The following are the steps of the hypothesis testing: 1. Specify H 0 and H 1. 2. Choose the significance level α. 3. Define a decision rule (test statistic and critical region). 4. Perform the test using the data: given the data compute the test statistic and see if it falls into the critical region. 9/16

P-value We can compute the lowest significance level consistent with rejecting H 0. This is called the p-value: p-value=min {0 < α < 1 : T Γ α }. The p-value could be viewed as a measure of contradiction. The smaller the p-value is, the larger the contradiction is. In a probabilistic model, we use the way of reasoning in proof by contradiction with a measure of contradiction based on data to carry out the similar reasoning. Now if the p-value is smaller than our tolerance(significance level), then we reject null hypothesis like what we conclude in proof by contradiction. 10/16

Errors There are two types of errors that the econometrician can make: Truth H 0 H 1 Decision H 0 Type II error H 1 Type I error Type I error is the error of rejecting H 0 when H 0 is true. P (Type I error) = P (reject H 0 H 0 is true) = α. Type II error is the error of not rejecting H 0 when H 1 is true. 11/16

Errors The real line is split into two regions: acceptance region and rejection region (critical region). When T / Γ α, we accept H 0 (and risk making a Type II error). When T Γ α, we reject H 0 (and risk making a Type I error). Unfortunately, the probabilities of Type I and II errors are inversely related. By decreasing the probability of Type I error α, one makes the critical region smaller, which increases the probability of the Type II error. Thus it is impossible to make both errors arbitrary small. We require that the test we carry out has to be valid: the probability of type-i error (P [T Γ α ] under H 0 ) must be α. 12/16

Power of tests* Power of a test: 1 P (Type II error) = 1 P (Accept H 0 H 0 is false). We want the power of a test to be as large as possible, for a given significance level. We may not know the distribution of the test statistic under H 1. The distribution typically depends on θ. The power function π (θ) of the test is the probability that H 0 is rejected as a function of the true parameter value θ. 13/16

The simple example* For any given value of µ, define T µ = T µ σ/ n = X µ σ/ n, which is always N (0, 1) if ( the true mean is µ. Under H 1 : µ > 0, T µ N σ/ )., 1 n We reject the test when T z 1 α, which holds if and only if T µ = T µ σ/ z 1 α µ n σ/. n The power function is π (µ) = 1 Φ ( z 1 α nµ σ ). Notice that π (0) = 1 Φ (z 1 α ) = α, which is simply the significance level. 14/16

The simple example* We notice that π (µ) is smaller for all µ if the significance level α is smaller (and hence z 1 α is larger). This reflects the trade-off between type-i error and type-ii error probabilities: we cannot reduce both simultaneously. π (µ) is increasing in µ. For µ s that are farther away from 0, the test can detect such deviation at a higher probability. As µ, the power converges to 1. The test is very likely to reject H 0 if the true mean is very large. π (µ) increases with the sample size n. The test can detect falseness of H 0 at a higher probability if our sample contains more information. 15/16

The simple example* We consider an alternative critical region: Γ α = (, z 1 α ]. Notice that under H 0, P [ T Γ α ] = α, so this is a valid test. We reject the test when T z 1 α, which holds if and only if T µ = T µ σ/ n z 1 α µ σ/ n. So the power function is π (µ) = Φ ( z 1 α nµ σ ). π (µ) is decreasing in both µ and n, which is undesirable. For all µ > 0, the power is smaller than α. At the same significance level, the power of the alternative critical region is much lower. 16/16