Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Size: px
Start display at page:

Download "Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing"

Transcription

1 Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas October 13, 2012

2 Outline Hypothesis Testing 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

3 Outline Hypothesis Testing 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

4 Hypothesis Testing: Basic Idea Null hypothesis: What we believe in the absence of further evidence, e.g. a two-sided coin is fair with equal likelihood. Think: Null hypothesis = default assumption. Two kinds of testing: There is only the null hypothesis, and we accept or reject it. There is a null as well as an alternate hypothesis, and we choose one or the other. The second kind of testing is easier: We choose whichever hypothesis is more likely under the data. The first kind of testing is harder.

5 Choosing Between Alternatives: Example We are given a coin. The null hypothesis is that the coin is fair with equal probabilities of heads and tails. Call it H 0. The alternative hypothesis is that the coin is biased with the probability of heads equal to 0.7. Call it H 1. Suppose we toss the coin 20 times and 12 heads result. Which hypothesis should we accept?

6 Choosing Between Alternatives: Example (Cont d) Let n = 20 (number of coin tosses), k = 12 (number of heads), p 0 = 0.5 (probability of heads under hypothesis H 0 ) and P 1 = 0.7 (probability of heads under hypothesis H 1 ). The likelihood of the observed outcome under each hypothesis is computed. ( ) 20 L 0 = (p 12 0 ) 12 (1 p 0 ) 8 = , L 1 = ( ) (p 1 ) 12 (1 p 1 ) 8 = So we accept hypothesis H 0, that the coin is fair, but only because the alternative hypothesis is even less likely!

7 Connection to MLE We choose the hypothesis that the coin is fair only because the alternate hypothesis is even more unlikely! So what is the value of p that maximizes ( ) 20 L = p 12 (1 p) 8? 12 Answer: p MLE = 12/20 = 0.6, the fraction of heads observed. With MLE (maximum likelihood estimation), we need not choose between two competing hypotheses MLE gives the most likely values for the parameters!

8 Outline Hypothesis Testing 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

9 Estimating Probabilities of Binary Outcomes Suppose an event has only two outcomes, e.g. coin toss. Let p equal the true but unknown probability of success, e.g. that the coin comes up heads. After n trials, suppose k successes result. Then ˆp := k/n is called the empirical probability of success. As we have seen, it is also the maximum likelihood estimate of p. Question: How close is the empirical probability ˆp to the true but unknown probability p? Hoeffding s inequalities answer this question.

10 : Statements Let ɛ > 0 be any specified accuracy. Then Pr{ˆp p ɛ} exp( 2nɛ 2 ). Pr{ˆp p ɛ} exp( 2nɛ 2 ). Pr{ ˆp p ɛ} 1 2 exp( 2nɛ 2 ).

11 : Interpretation Interpretations of Hoeffding s inequalities: With confidence 1 2 exp( 2nɛ 2 ), we can say that the true but unknown probability p lies in the interval (ˆp ɛ, ˆp + ɛ). As we increase ɛ, the term δ := 2 exp( 2nɛ 2 ) decreases, and we can be more sure of our interval. The widely used 95% confidence interval corresponds to δ = 0.5. The one-sided inequalities have similar interpretations.

12 An Example of Applying Hoeffding s Inequality Suppose we toss a coin 1000 times and it comes up heads 552 times. How sure can we be that the coin is biased? n = 1000, k = 552, ˆp = If p > 0.5 then we can say that the coin is biased. So let ɛ = ˆp p = Compute δ = exp( 2nɛ 2 ) = So with confidence 1 δ = , we can say that p > 0.5. In other words, we can be 99.55% sure that the coin is biased. Using the two-sided Hoeffding inequality, we can be 99.1% sure that ˆp (0.5, 0.614).

13 Another Example An opinion poll of 750 voters (ignoring don t know s) shows that 387 will vote for candidate A and 363 will vote for candidate B. How sure can we be that candidate A will win? Let p denote the true but unknown fraction of voters who will vote for A, and ˆp = 387/750 = denote the empirical estimate of p. If p < 0.5 then A will lose. So the accuracy ɛ = , and the number of samples n = 750. The one-sided confidence is δ = exp( 2nɛ 2 ) = So we can be only 1 δ 32% sure that A will win. In other words, the election cannot be called with any confidence based on such a small margin of preference.

14 Relating Confidence, Accuracy and Number of Samples For the two-sided Hoeffding inequality, the confidence δ associated with n samples and accuracy ɛ is given by δ = 2 exp( 2nɛ 2 ). We can turn this around and ask: Given an empirical estimate ˆp based on n samples, what is the accuracy corresponding to a given confidence level δ? Solving the above equation for ɛ in terms of δ and n gives ɛ(n, δ) = ( 1 2n log 2 ) 1/2. δ So with confidence δ we can say that the true but unknown probability p is in the interval [ˆp ɛ(n, δ), ˆp + ɛ(n, δ)].

15 for More Than Two Outcomes Suppose a random experiment has more than two possible outcomes (e.g. rolling six-sided die). Say there are k outcomes, and in n trials, the i-th outcome appears n i times (and of course k i=1 n i = n). We can define ˆp i = n i, i = 1,..., k, n and as we have seen, these are the maximum likelihood estimates for each probability. Question: How good are these estimates?

16 More Than Two Outcomes 2 Fact: For any sample size n and any accuracy ɛ, it is the case that Pr{max ˆp i p i > ɛ} 2k exp( 2nɛ 2 ). i So with confidence 1 2k exp( 2nɛ 2 ), we can assert that every empirical probability ˆp i is within ɛ of the correct value.

17 More Than Two Outcomes: Example Suppose we roll a six-sided die 1,000 times and get the outcomes 1 through 6 in the following order: ˆp 1 = 0.169, ˆp 2 = 0.165, ˆp 3 = 0.166, ˆp 4 = 0.165, ˆp 5 = 0.167, ˆp 6 = With what confidence can we say that the die is not fair, that is, that ˆp i 1/6 for all i?

18 More Than Two Outcomes: Example (Cont d) Suppose that indeed the true probability is p i = 1/6 for all i. Then max ˆp i p i = ˆp 1 1/ i Take ɛ = 0.233, n = 1000 and compute δ = 6 2 exp( 2nɛ 2 ) 11.87! How can a probability be greater than one? Note: This δ is just an upper bound for Pr{max i ˆp i p i > ɛ}; so it can be larger than one. So we cannot rule out the possibility that the die is fair (which is quite different from saying that it is fair).

19 Outline Hypothesis Testing K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

20 Outline Hypothesis Testing K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

21 K-S Tests: Problem Formulations K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements There are two widely used tests. They should be called the Kolmogorov test and the Smirnov test, respectively. Unfortunately the erroneous names one-sample K-S test and two-sample K-S test have become popular. Kolmogorov Test, or One-Sample K-S Test: We have a set of samples, and we have a candidate probability distribution. Question: How well does the distribution fit the set of samples? Smirnov Test, or Two-Sample K-S Test: We have two sets of samples, say x 1,..., x n and y 1,..., y m. Question: How sure are we that both sets of samples came from the same (but unknown) distribution?

22 Outline Hypothesis Testing K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

23 Empirical Distributions K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements Suppose X is a random variable for which we have generated n i.i.d. samples, call them x 1,..., x n. Then we define the empirical distribution of X, based on these observations, as follows: ˆΦ(a) = 1 n n i=1 I {xi a}, where I denotes the indicator function: I = 1 if the condition below is satisfied and I = 0 otherwise. So in this case ˆΦ(a) is just the fraction of the n samples that are a. The diagram on the next slide illustrates this.

24 Empirical Distribution Depicted K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements Note: The diagram shows the samples occurring in increasing order but they can be in any order. 1 1 Source: gm distribution function.htm

25 Glivenko-Cantelli Lemma K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements Theorem: As n, the empirical distribution ˆΦ( ) approaches the true distribution Φ( ). Specifically, if we define the Kolmogorov-Smirnov distance then d n 0 as n. d n = max ˆΦ(u) Φ(u), u At what rate does the convergence take place?

26 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements One-Sample Kolmogorov-Smirnov Statistic Fix a confidence level δ > 0 (usually δ is taken as 0.05 or 0.02). Define the threshold ( 1 θ(n, δ) = 2n log 2 ) 1/2. δ Then with probability 1 δ, we can say that max u ˆΦ(u) Φ(u) =: d n θ n.

27 One-Sample Kolmogorov-Smirnov Test K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements Given samples x 1,..., x n, fit it with some distribution F ( ) (e.g. Gaussian). Compute the K-S statistic d n = max ˆΦ(u) F (u). u Compare d n with the threshold θ(n, δ). If d n > θ(n, δ), we reject the null hypothesis at level δ. In other words, if d n > θ(n, δ), then we are 1 δ sure that the data was not generated by the distribution F ( ).

28 Outline Hypothesis Testing 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

29 : Motivation The student t test is used the null hypothesis that two sets of samples have the same mean, assuming that they have the same variance. The test has broad applicability even if the assumption of same variance is not satisfied. Problem: We are given two samples x 1,..., x m1 and x m1 +1,..., x m1 +m 2. Determine whether the two sets of samples arise from a distribution with the same mean. Application: Most commonly used in quality control.

30 : Theory Let x 1, x 2 denote the means of the two sample classes, that is, x 1 = 1 m 1 x i, x 2 = 1 m 2 m 1 m 2 i=1 i=1 x m1 +i. Let S 1, S 2 denote the unbiased estimates of the standard deviations of the two samples, that is, S 2 1 = S 2 2 = 1 m m 2 1 m 1 i=1 m 2 i=1 (x i x 1 ) 2, (x m1 +i x 2 ) 2.

31 : Theory 2 Now define the pooled standard deviation S 12 by Then the quantity S12 2 = (m 1 1)S1 2 + (m 2 1)S2 2. m 1 + m 2 2 d t = x 1 x 2 S 12 (1/m1 ) + (1/m 2 ) satisfies the t distribution with m 1 + m 2 2 degrees of freedom. As the number of d.o.f. becomes large, the t distribution approaches the normal distribution. The next slide shows the density of the t distribution for various d.o.f.

32 Density of the t Distribution

33 Outline Hypothesis Testing 1 Hypothesis Testing 2 3 K-S (Kolmogorov-Smirnov) Tests: Objectives Kolmogorov-Smirnov Tests: Statements 4 5

34 : Motivation The t test is to determine whether two samples have the same mean. The chi-squared test is to determine whether two samples have the same variance. The application is again to quality control.

35 : Theory Given two sets of samples, say x 1,..., x m1 and x m1 +1,..., x m1 +m 2 (where usually m 2 m 1 ), compute the unbiased variance estimate V 1 of the larger (first) sample V 1 = 1 m 1 1 m 1 i=1 (x i x 1 ) 2, and the sum of squares of the smaller (second) sample m 2 S 2 = (x m1 +i x 2 ) 2 = (m 2 1)V 2. i=1 Then the ratio S 2 /V 1 satisfies the chi-squared (or χ 2 ) distribution with m 2 1 degrees of freedom.

36 Distribution Function of the Chi-Squared Variable

37 Density Function of the Chi-Squared Variable

38 Application of the Note that the χ 2 r.v. is always nonnegative. So, given some confidence δ (usually δ = 0.05), we need to determine a confidence interval x l = Φ 1 χ 2,m 2 1 (δ), x u = Φ 1 χ 2,m 2 (1 δ). 1 If the test statistic S 2 /V 1 lies in the interval [x l, x u ], then we accept the null hypothesis that both samples have the same variance.

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu

More information

Fundamental Probability and Statistics

Fundamental Probability and Statistics Fundamental Probability and Statistics "There are known knowns. These are things we know that we know. There are known unknowns. That is to say, there are things that we know we don't know. But there are

More information

The Logit Model: Estimation, Testing and Interpretation

The Logit Model: Estimation, Testing and Interpretation The Logit Model: Estimation, Testing and Interpretation Herman J. Bierens October 25, 2008 1 Introduction to maximum likelihood estimation 1.1 The likelihood function Consider a random sample Y 1,...,

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

Introductory Econometrics. Review of statistics (Part II: Inference)

Introductory Econometrics. Review of statistics (Part II: Inference) Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University. Lecture 8 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University October 22, 2007 1 2 3 4 5 6 1 Define convergent series 2 Define the Law of Large Numbers

More information

Institute of Actuaries of India

Institute of Actuaries of India Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the

More information

Econ 325: Introduction to Empirical Economics

Econ 325: Introduction to Empirical Economics Econ 325: Introduction to Empirical Economics Chapter 9 Hypothesis Testing: Single Population Ch. 9-1 9.1 What is a Hypothesis? A hypothesis is a claim (assumption) about a population parameter: population

More information

Ling 289 Contingency Table Statistics

Ling 289 Contingency Table Statistics Ling 289 Contingency Table Statistics Roger Levy and Christopher Manning This is a summary of the material that we ve covered on contingency tables. Contingency tables: introduction Odds ratios Counting,

More information

6.4 Type I and Type II Errors

6.4 Type I and Type II Errors 6.4 Type I and Type II Errors Ulrich Hoensch Friday, March 22, 2013 Null and Alternative Hypothesis Neyman-Pearson Approach to Statistical Inference: A statistical test (also known as a hypothesis test)

More information

Inference and Regression

Inference and Regression Inference and Regression Assignment 4 - Solutions Department of IOMS Professor William Greene Phone: 212.998.0876 Office: KMC 7-90 Home page:www.stern.nyu.edu/~wgreene Email: wgreene@stern.nyu.edu Course

More information

. Find E(V ) and var(v ).

. Find E(V ) and var(v ). Math 6382/6383: Probability Models and Mathematical Statistics Sample Preliminary Exam Questions 1. A person tosses a fair coin until she obtains 2 heads in a row. She then tosses a fair die the same number

More information

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing Robert Vanderbei Fall 2014 Slides last edited on November 24, 2014 http://www.princeton.edu/ rvdb Coin Tossing Example Consider two coins.

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

Point Estimation. Vibhav Gogate The University of Texas at Dallas

Point Estimation. Vibhav Gogate The University of Texas at Dallas Point Estimation Vibhav Gogate The University of Texas at Dallas Some slides courtesy of Carlos Guestrin, Chris Bishop, Dan Weld and Luke Zettlemoyer. Basics: Expectation and Variance Binary Variables

More information

Topic 19 Extensions on the Likelihood Ratio

Topic 19 Extensions on the Likelihood Ratio Topic 19 Extensions on the Likelihood Ratio Two-Sided Tests 1 / 12 Outline Overview Normal Observations Power Analysis 2 / 12 Overview The likelihood ratio test is a popular choice for composite hypothesis

More information

Exam 2 Practice Questions, 18.05, Spring 2014

Exam 2 Practice Questions, 18.05, Spring 2014 Exam 2 Practice Questions, 18.05, Spring 2014 Note: This is a set of practice problems for exam 2. The actual exam will be much shorter. Within each section we ve arranged the problems roughly in order

More information

INTERVAL ESTIMATION AND HYPOTHESES TESTING

INTERVAL ESTIMATION AND HYPOTHESES TESTING INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,

More information

18.05 Practice Final Exam

18.05 Practice Final Exam No calculators. 18.05 Practice Final Exam Number of problems 16 concept questions, 16 problems. Simplifying expressions Unless asked to explicitly, you don t need to simplify complicated expressions. For

More information

Robustness and Distribution Assumptions

Robustness and Distribution Assumptions Chapter 1 Robustness and Distribution Assumptions 1.1 Introduction In statistics, one often works with model assumptions, i.e., one assumes that data follow a certain model. Then one makes use of methodology

More information

COMP2610/COMP Information Theory

COMP2610/COMP Information Theory COMP2610/COMP6261 - Information Theory Lecture 9: Probabilistic Inequalities Mark Reid and Aditya Menon Research School of Computer Science The Australian National University August 19th, 2014 Mark Reid

More information

Probability and Statistics Notes

Probability and Statistics Notes Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

F79SM STATISTICAL METHODS

F79SM STATISTICAL METHODS F79SM STATISTICAL METHODS SUMMARY NOTES 9 Hypothesis testing 9.1 Introduction As before we have a random sample x of size n of a population r.v. X with pdf/pf f(x;θ). The distribution we assign to X is

More information

Overview. Confidence Intervals Sampling and Opinion Polls Error Correcting Codes Number of Pet Unicorns in Ireland

Overview. Confidence Intervals Sampling and Opinion Polls Error Correcting Codes Number of Pet Unicorns in Ireland Overview Confidence Intervals Sampling and Opinion Polls Error Correcting Codes Number of Pet Unicorns in Ireland Confidence Intervals When a random variable lies in an interval a X b with a specified

More information

Visual interpretation with normal approximation

Visual interpretation with normal approximation Visual interpretation with normal approximation H 0 is true: H 1 is true: p =0.06 25 33 Reject H 0 α =0.05 (Type I error rate) Fail to reject H 0 β =0.6468 (Type II error rate) 30 Accept H 1 Visual interpretation

More information

Estimating the accuracy of a hypothesis Setting. Assume a binary classification setting

Estimating the accuracy of a hypothesis Setting. Assume a binary classification setting Estimating the accuracy of a hypothesis Setting Assume a binary classification setting Assume input/output pairs (x, y) are sampled from an unknown probability distribution D = p(x, y) Train a binary classifier

More information

Maximum-Likelihood Estimation: Basic Ideas

Maximum-Likelihood Estimation: Basic Ideas Sociology 740 John Fox Lecture Notes Maximum-Likelihood Estimation: Basic Ideas Copyright 2014 by John Fox Maximum-Likelihood Estimation: Basic Ideas 1 I The method of maximum likelihood provides estimators

More information

Chapters 10. Hypothesis Testing

Chapters 10. Hypothesis Testing Chapters 10. Hypothesis Testing Some examples of hypothesis testing 1. Toss a coin 100 times and get 62 heads. Is this coin a fair coin? 2. Is the new treatment on blood pressure more effective than the

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Statistical Methods for Astronomy

Statistical Methods for Astronomy Statistical Methods for Astronomy Probability (Lecture 1) Statistics (Lecture 2) Why do we need statistics? Useful Statistics Definitions Error Analysis Probability distributions Error Propagation Binomial

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

Inferences About Two Proportions

Inferences About Two Proportions Inferences About Two Proportions Quantitative Methods II Plan for Today Sampling two populations Confidence intervals for differences of two proportions Testing the difference of proportions Examples 1

More information

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages Name No calculators. 18.05 Final Exam Number of problems 16 concept questions, 16 problems, 21 pages Extra paper If you need more space we will provide some blank paper. Indicate clearly that your solution

More information

Rank-Based Methods. Lukas Meier

Rank-Based Methods. Lukas Meier Rank-Based Methods Lukas Meier 20.01.2014 Introduction Up to now we basically always used a parametric family, like the normal distribution N (µ, σ 2 ) for modeling random data. Based on observed data

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

Probability & Statistics - FALL 2008 FINAL EXAM

Probability & Statistics - FALL 2008 FINAL EXAM 550.3 Probability & Statistics - FALL 008 FINAL EXAM NAME. An urn contains white marbles and 8 red marbles. A marble is drawn at random from the urn 00 times with replacement. Which of the following is

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some

More information

The PAC Learning Framework -II

The PAC Learning Framework -II The PAC Learning Framework -II Prof. Dan A. Simovici UMB 1 / 1 Outline 1 Finite Hypothesis Space - The Inconsistent Case 2 Deterministic versus stochastic scenario 3 Bayes Error and Noise 2 / 1 Outline

More information

Bias Variance Trade-off

Bias Variance Trade-off Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]

More information

STAT 285 Fall Assignment 1 Solutions

STAT 285 Fall Assignment 1 Solutions STAT 285 Fall 2014 Assignment 1 Solutions 1. An environmental agency sets a standard of 200 ppb for the concentration of cadmium in a lake. The concentration of cadmium in one lake is measured 17 times.

More information

Slides for Data Mining by I. H. Witten and E. Frank

Slides for Data Mining by I. H. Witten and E. Frank Slides for Data Mining by I. H. Witten and E. Frank Predicting performance Assume the estimated error rate is 5%. How close is this to the true error rate? Depends on the amount of test data Prediction

More information

Probability Theory for Machine Learning. Chris Cremer September 2015

Probability Theory for Machine Learning. Chris Cremer September 2015 Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares

More information

Lecture 10: Generalized likelihood ratio test

Lecture 10: Generalized likelihood ratio test Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual

More information

Chapter 3: Probability 3.1: Basic Concepts of Probability

Chapter 3: Probability 3.1: Basic Concepts of Probability Chapter 3: Probability 3.1: Basic Concepts of Probability Objectives Identify the sample space of a probability experiment and a simple event Use the Fundamental Counting Principle Distinguish classical

More information

Confidence Intervals, Testing and ANOVA Summary

Confidence Intervals, Testing and ANOVA Summary Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0

More information

Sampling, Confidence Interval and Hypothesis Testing

Sampling, Confidence Interval and Hypothesis Testing Sampling, Confidence Interval and Hypothesis Testing Christopher Grigoriou Executive MBA HEC Lausanne 2007-2008 1 Sampling : Careful with convenience samples! World War II: A statistical study to decide

More information

Quantitative Analysis and Empirical Methods

Quantitative Analysis and Empirical Methods Hypothesis testing Sciences Po, Paris, CEE / LIEPP Introduction Hypotheses Procedure of hypothesis testing Two-tailed and one-tailed tests Statistical tests with categorical variables A hypothesis A testable

More information

Mathematical Notation Math Introduction to Applied Statistics

Mathematical Notation Math Introduction to Applied Statistics Mathematical Notation Math 113 - Introduction to Applied Statistics Name : Use Word or WordPerfect to recreate the following documents. Each article is worth 10 points and should be emailed to the instructor

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics MAS 713 Chapter 8 Previous lecture: 1 Bayesian Inference 2 Decision theory 3 Bayesian Vs. Frequentist 4 Loss functions 5 Conjugate priors Any questions? Mathematical Statistics

More information

ECO220Y Review and Introduction to Hypothesis Testing Readings: Chapter 12

ECO220Y Review and Introduction to Hypothesis Testing Readings: Chapter 12 ECO220Y Review and Introduction to Hypothesis Testing Readings: Chapter 12 Winter 2012 Lecture 13 (Winter 2011) Estimation Lecture 13 1 / 33 Review of Main Concepts Sampling Distribution of Sample Mean

More information

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting

More information

Problems ( ) 1 exp. 2. n! e λ and

Problems ( ) 1 exp. 2. n! e λ and Problems The expressions for the probability mass function of the Poisson(λ) distribution, and the density function of the Normal distribution with mean µ and variance σ 2, may be useful: ( ) 1 exp. 2πσ

More information

LECTURE 5 HYPOTHESIS TESTING

LECTURE 5 HYPOTHESIS TESTING October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.

More information

Just Enough Likelihood

Just Enough Likelihood Just Enough Likelihood Alan R. Rogers September 2, 2013 1. Introduction Statisticians have developed several methods for comparing hypotheses and for estimating parameters from data. Of these, the method

More information

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks Outline 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks Likelihood A common and fruitful approach to statistics is to assume

More information

Topic 15: Simple Hypotheses

Topic 15: Simple Hypotheses Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is

More information

Chapter 10. Hypothesis Testing (I)

Chapter 10. Hypothesis Testing (I) Chapter 10. Hypothesis Testing (I) Hypothesis Testing, together with statistical estimation, are the two most frequently used statistical inference methods. It addresses a different type of practical problems

More information

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis

More information

Composite Hypotheses and Generalized Likelihood Ratio Tests

Composite Hypotheses and Generalized Likelihood Ratio Tests Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve

More information

TUTORIAL 8 SOLUTIONS #

TUTORIAL 8 SOLUTIONS # TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level

More information

CSE 312 Final Review: Section AA

CSE 312 Final Review: Section AA CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material

More information

Bayesian Methods: Naïve Bayes

Bayesian Methods: Naïve Bayes Bayesian Methods: aïve Bayes icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Last Time Parameter learning Learning the parameter of a simple coin flipping model Prior

More information

Lecture 2 Sep 5, 2017

Lecture 2 Sep 5, 2017 CS 388R: Randomized Algorithms Fall 2017 Lecture 2 Sep 5, 2017 Prof. Eric Price Scribe: V. Orestis Papadigenopoulos and Patrick Rall NOTE: THESE NOTES HAVE NOT BEEN EDITED OR CHECKED FOR CORRECTNESS 1

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics Bayesian Parameter Estimation Introduction to Bayesian Statistics Harvey Thornburg Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University Stanford, California

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

POLI 443 Applied Political Research

POLI 443 Applied Political Research POLI 443 Applied Political Research Session 6: Tests of Hypotheses Contingency Analysis Lecturer: Prof. A. Essuman-Johnson, Dept. of Political Science Contact Information: aessuman-johnson@ug.edu.gh College

More information

Evaluating Classifiers. Lecture 2 Instructor: Max Welling

Evaluating Classifiers. Lecture 2 Instructor: Max Welling Evaluating Classifiers Lecture 2 Instructor: Max Welling Evaluation of Results How do you report classification error? How certain are you about the error you claim? How do you compare two algorithms?

More information

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag Decision Trees Nicholas Ruozzi University of Texas at Dallas Based on the slides of Vibhav Gogate and David Sontag Supervised Learning Input: labelled training data i.e., data plus desired output Assumption:

More information

EC2001 Econometrics 1 Dr. Jose Olmo Room D309

EC2001 Econometrics 1 Dr. Jose Olmo Room D309 EC2001 Econometrics 1 Dr. Jose Olmo Room D309 J.Olmo@City.ac.uk 1 Revision of Statistical Inference 1.1 Sample, observations, population A sample is a number of observations drawn from a population. Population:

More information

MAT 271E Probability and Statistics

MAT 271E Probability and Statistics MAT 71E Probability and Statistics Spring 013 Instructor : Class Meets : Office Hours : Textbook : Supp. Text : İlker Bayram EEB 1103 ibayram@itu.edu.tr 13.30 1.30, Wednesday EEB 5303 10.00 1.00, Wednesday

More information

Point Estimation. Maximum likelihood estimation for a binomial distribution. CSE 446: Machine Learning

Point Estimation. Maximum likelihood estimation for a binomial distribution. CSE 446: Machine Learning Point Estimation Emily Fox University of Washington January 6, 2017 Maximum likelihood estimation for a binomial distribution 1 Your first consulting job A bored Seattle billionaire asks you a question:

More information

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( ) Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr Pr = Pr Pr Pr() Pr Pr. We are given three coins and are told that two of the coins are fair and the

More information

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2 Probability Density Functions and the Normal Distribution Quantitative Understanding in Biology, 1.2 1. Discrete Probability Distributions 1.1. The Binomial Distribution Question: You ve decided to flip

More information

Introduction to Statistical Data Analysis Lecture 7: The Chi-Square Distribution

Introduction to Statistical Data Analysis Lecture 7: The Chi-Square Distribution Introduction to Statistical Data Analysis Lecture 7: The Chi-Square Distribution James V. Lambers Department of Mathematics The University of Southern Mississippi James V. Lambers Statistical Data Analysis

More information

Hypothesis Tests Solutions COR1-GB.1305 Statistics and Data Analysis

Hypothesis Tests Solutions COR1-GB.1305 Statistics and Data Analysis Hypothesis Tests Solutions COR1-GB.1305 Statistics and Data Analysis Introduction 1. An analyst claims to have a reliable model for Twitter s quarterly revenues. His model predicted that the most recent

More information

Advanced Herd Management Probabilities and distributions

Advanced Herd Management Probabilities and distributions Advanced Herd Management Probabilities and distributions Anders Ringgaard Kristensen Slide 1 Outline Probabilities Conditional probabilities Bayes theorem Distributions Discrete Continuous Distribution

More information

Example. χ 2 = Continued on the next page. All cells

Example. χ 2 = Continued on the next page. All cells Section 11.1 Chi Square Statistic k Categories 1 st 2 nd 3 rd k th Total Observed Frequencies O 1 O 2 O 3 O k n Expected Frequencies E 1 E 2 E 3 E k n O 1 + O 2 + O 3 + + O k = n E 1 + E 2 + E 3 + + E

More information

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0.

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0. () () a. X is a binomial distribution with n = 000, p = /6 b. The expected value, variance, and standard deviation of X is: E(X) = np = 000 = 000 6 var(x) = np( p) = 000 5 6 666 stdev(x) = np( p) = 000

More information

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Introduction: MLE, MAP, Bayesian reasoning (28/8/13) STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this

More information

14.30 Introduction to Statistical Methods in Economics Spring 2009

14.30 Introduction to Statistical Methods in Economics Spring 2009 MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Lecture 35: December The fundamental statistical distances

Lecture 35: December The fundamental statistical distances 36-705: Intermediate Statistics Fall 207 Lecturer: Siva Balakrishnan Lecture 35: December 4 Today we will discuss distances and metrics between distributions that are useful in statistics. I will be lose

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015 AMS7: WEEK 7. CLASS 1 More on Hypothesis Testing Monday May 11th, 2015 Testing a Claim about a Standard Deviation or a Variance We want to test claims about or 2 Example: Newborn babies from mothers taking

More information

Swarthmore Honors Exam 2012: Statistics

Swarthmore Honors Exam 2012: Statistics Swarthmore Honors Exam 2012: Statistics 1 Swarthmore Honors Exam 2012: Statistics John W. Emerson, Yale University NAME: Instructions: This is a closed-book three-hour exam having six questions. You may

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

Discrete Distributions

Discrete Distributions Discrete Distributions STA 281 Fall 2011 1 Introduction Previously we defined a random variable to be an experiment with numerical outcomes. Often different random variables are related in that they have

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Lecture 9 Two-Sample Test. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Lecture 9 Two-Sample Test. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech Lecture 9 Two-Sample Test Fall 2013 Prof. Yao Xie, yao.xie@isye.gatech.edu H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech Computer exam 1 18 Histogram 14 Frequency 9 5 0 75 83.33333333

More information

STA Module 10 Comparing Two Proportions

STA Module 10 Comparing Two Proportions STA 2023 Module 10 Comparing Two Proportions Learning Objectives Upon completing this module, you should be able to: 1. Perform large-sample inferences (hypothesis test and confidence intervals) to compare

More information

STAT 4385 Topic 01: Introduction & Review

STAT 4385 Topic 01: Introduction & Review STAT 4385 Topic 01: Introduction & Review Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso xsu@utep.edu Spring, 2016 Outline Welcome What is Regression Analysis? Basics

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc. Chapter 14 From Randomness to Probability Copyright 2012, 2008, 2005 Pearson Education, Inc. Dealing with Random Phenomena A random phenomenon is a situation in which we know what outcomes could happen,

More information

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1 Math 66/566 - Midterm Solutions NOTE: These solutions are for both the 66 and 566 exam. The problems are the same until questions and 5. 1. The moment generating function of a random variable X is M(t)

More information