Summary of Chapters 7-9
|
|
- Hugo Cook
- 5 years ago
- Views:
Transcription
1 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two normal distributions N(µ X, σx 2 ) and N(µ Y, σy 2 ), respectively. Case 1: σx 2 and σ2 Y are known. A 100(1 α)% confidence interval for µ X µ Y is [ x ȳ z α/2 σ W, x ȳ + z α/2 σ W ], where σ W = σx/n 2 + σy/m 2 is the standard deviation of the point estimator X Ȳ. Remark: If the sample sizes n and m are large (at least 30) and σ X and σ Y are unknown, we can replace σx 2 and σy 2 with s 2 x and s 2 y, respectively to find an approximate 100(1 α)% confidence interval: x ȳ ± z α/2 s 2 x/n + s 2 y/m. Case 2: σx 2 and σ2 Y are unknown but the sample sizes are small. A 100(1 α)% confidence interval for µ X µ Y is [ 1 x ȳ t 0 s p n m, x ȳ + t 0s p n + 1, m ] where x, ȳ and s p are the observed values of X, Ȳ and S p with (n 1)s 2 X s p = + (m 1)s2 Y n + m 2 and t 0 = t α/2 (n + m 2). For paired random sample (X 1, Y 1 ), (X 2, Y 2 ),, (X n, Y n ), let D i = X i Y i, i = 1, 2,..., n. Then, we could assume that D 1, D 2,..., D n is a random sample from N(µ D, σd 2 ), where µ D and σ D are the mean and standard deviation of each difference. A 100(1 α)% confidence interval for µ D = µ X µ Y is [ d s d t α/2 (n 1), d s + t α/2 (n 1) d ], n n where d and s d are the observed mean and standard deviation of the sample d 1, d 2,..., d n. 1
2 7.3. Confidence Intervals for Proportions Let Y b(n, p) and y be the observed value of Y. Then, an approximate 100(1 α)% confidence interval for p is [ y (y/n)(1 y/n) n z α/2, y ] (y/n)(1 y/n) n n + z α/2, n where y/n is a point estimate of p. Remark: one-sided 100(1 α)% confidence interval for p with (i) upper bound: [0, y/n + z α (y/n)(1 y/n)/n]; (ii) lower bound: [y/n z α (y/n)(1 y/n)/n, 1]. Let Y 1 b(n 1, p 1 ) and Y 1 b(n 1, p 1 ), and let y 1 and y 2 be the observed values of Y 1 and Y 2, respectively. Then, an approximate 100(1 α)% confidence interval for p 1 p 2 is (y 1 /n 1 y 2 /n 2 ) ± z α/2 (y1 /n 1 )(1 y 1 /n 1 )/n 1 + (y 2 /n 2 )(1 y 2 /n 2 )/n Sample Size (1) Sample Size of Estimating µ Let ε = z α/2 (σ/ n) be the maximum error of the estimate for µ. Then, the required sample size with a given maximum error of the estimation is which is rounded to the next integer. n = z2 α/2 σ2 ε 2, (2) Sample Size of Estimating Proportion p Let ε = z α/2 ˆp(1 ˆp)/n be the maximum error of the point estimate ˆp. Assume that we can obtain an estimate ˆp by using available historical data. Then, the required sample size within the given error bound ε is n = z2 α/2 ˆp(1 ˆp) ε 2, which is rounded to the next integer. 2
3 If ˆp is not available (i.e., no historical data), we can use the following formula to compute the required sample size: which is rounded to the next integer. n = z2 α/2 4ε 2, Chapter 8. Tests of Statistical Hypotheses 8.1. Tests about One Mean Definition. A statistical hypothesis is a statement about the parameters of one or more populations. Null hypothesis H 0 : specifies the distribution of the population (or the parameters will be equal to some specified values). Alternative hypothesis H 1 : a competing statement against H 0. Two-sided alternative hypothesis H 1 contains sign; the one-sided alternative hypothesis contains either > sign or < sign. Definition. (i) Type I error: rejecting H 0 when H 0 is true; (ii) Type II error: failing to reject H 0 when H 0 is false. Definition. α = P (Type I error) is called the significance level of the test. β = P (Type II error) is called the power of the test. Definition. p-value is the smallest level of significance that would lead to rejection of H 0 with the given data. Criterion: we reject H 0 if p-value α. Otherwise (i.e. p-value> α), we fail to reject H 0. (1) Test about One Mean Case 1: σ is known. Assume that X N(µ, σ 2 ), X 1,..., X n is a random sample from the distribution N(µ, σ 2 ). Null hypothesis H 0 : µ = µ 0. The sample mean is X = (1/n) n i=1 X i and the sample variance is S 2 = 1 n n 1 i=1 (X i X) 2. 3
4 Test statistics: Z = X µ 0 σ/ n. Table : Tests of Hypotheses about One Mean, Variance Known where z = x µ0 σ/ n µ = µ 0 µ > µ 0 z z α µ = µ 0 µ < µ 0 z z α µ = µ 0 µ µ 0 z z α/2 is the observed value of the test statistics Z. Let z be the observed value of the test statistics Z. We can compute the p-value via the formula: 2P (Z z ), if H 1 : µ µ 0 ; p value = 1 P (Z < z), if H 1 : µ > µ 0 ; P (Z z), if H 1 : µ < µ 0. Case 2: σ is unknown. Test statistics: T = X µ 0 S/ n. T has a t distribution with r = n 1 degrees of freedom. Table : Tests of Hypotheses about One Mean, Variance Unknown where t = x µ0 s/ n µ = µ 0 µ > µ 0 t t α (n 1) µ = µ 0 µ < µ 0 t t α (n 1) µ = µ 0 µ µ 0 t t α/2 (n 1) is the observed value of the test statistics T. Let t be the observed value of the test statistics T. We can compute the p-value via the formula: 2P (T t ), if H 1 : µ µ 0 ; p value = 1 P (T < t), if H 1 : µ > µ 0 ; P (T t), if H 1 : µ < µ 0. 4
5 (2) Comparison of Two Means (Paired t-test) Assume that X and Y are dependent. Let W = X Y. From the original data {(x 1, y 1 ), (x 2, y 2 ),..., (x n, y n )}, one can get the difference data {w 1, w 2,..., w n }. Null hypothesis H 0 : µ X = µ Y is equivalent to H 0 : µ W = 0. Test statistics: T = W 0 S W / n. Test procedures are the same as previous case Tests of the Equality of Two Means Let X N(µ X, σx 2 ) and Y N(µ Y, σy 2 ). Assume that X and Y are independent. We have two samples: {X 1, X 2,..., X n } and {Y 1, Y 2,..., Y m }. Null hypothesis H 0 : µ X µ Y = 0. Test statistics: where S P = T = X Ȳ S P 1/n + 1/m, (n 1)SX 2 + (m 1)S2 Y. n + m 2 T has a t distribution with r = n + m 2 degrees of freedom. Table : Tests of Hypotheses for Equality of Two Means when σ 2 X = σ2 Y where t = µ X = µ Y µ X > µ Y t t α (n + m 2) µ X = µ Y µ X < µ Y t t α (n + m 2) µ X = µ Y µ X µ Y t t α/2 (n + m 2) x ȳ s P 1/n+1/m is the observed value of the test statistics T. p-value can be computed by using formulas mentioned before. 5
6 8.3. Tests about Proportions Let Y be the number of successes in n independent trials with probability of success p and y be the observed value of Y. Table : Tests of Hypotheses for One Proportion p = p 0 p > p 0 z = y/n p0 α p0(1 p 0)/n p = p 0 p < p 0 z = y/n p0 z α p0(1 p 0)/n p = p 0 p p 0 z = y/n p0 z α/2 p0(1 p 0)/n Let Y 1 and Y 2 represent, respectively, the numbers of observed successes in n 1 and n 2 independent trials with probabilities of success p 1 and p 2. Table : Tests of Hypotheses for Two Proportions where ˆp = (y 1 + y 2 )/(n 1 + n 2 ) The Wilcoxon Tests y p 1 = p 2 p 1 > p 2 z = 1/n 1 y 2/n 2 ˆp(1 ˆp)(1/n1+1/n z α 2) p 1 = p 2 p 1 < p 2 z = p 1 = p 2 p 1 p 2 z = y 1/n 1 y 2/n 2 ˆp(1 ˆp)(1/n1+1/n 2) z α y 1/n 1 y 2/n 2 ˆp(1 ˆp)(1/n1+1/n 2) z α/2 Let m be the unknown median of a continuous-type random variable X. Let X 1, X 2,, X n denote the observations of a random sample from the distribution of X. We would like to test H 0 : m = m 0 against H 1 : m > m 0. We rank the absolute values X 1 m 0, X 2 m 0,, X n m 0 in non-decreasing order according to magnitude. Let R k denote the rank of X k m 0 among X 1 m 0, X 2 m 0,, X n m 0. With each R k, we associate the sign of the difference X k m 0. Namely, if X k m 0 > 0, then we use R k, but if X k m 0 < 0, we use R k. If the absolute values of the differences from m 0 of two or more observations are equal, each observation is assigned the average of the corresponding ranks. The Wilcoxon statistic W is the sum of these n signed ranks. For an approximate significance level α, the critical region is z z α or w z α n(n + 1)(2n + 1)/6. The p-value is computed by ( ) w 1 p value = P (W w) P Z n(n + 1)(2n + 1)/6 = 1 P ( Z ) w 1. n(n + 1)(2n + 1)/ Best Critical Regions 6
7 (Neyman-Pearson Lemma) Let X 1, X 2,..., X n be a random sample of size n from a distribution with pdf or pmf f(x; θ), where θ 0 and θ 1 are two possible values of θ. Denote the joint pdf or pmf of X 1, X 2,..., X n by the likelihood function L(θ) = f(x 1 ; θ)f(x 2 ; θ) f(x n ; θ). If there exists a positive constant k and a subset C of the sample space such that (a) P [(X 1, X 2,..., X n ) C; θ 0 ] = α, (b) L(θ0) L(θ k for (x 1) 1,..., x n C, and (c) L(θ0) L(θ k for (x 1) 1,..., x n C, then C is a best critical region of size α for testing the simple null hypothesis H 0 : θ = θ 0 against the simple alternative hypothesis H 1 : θ = θ 1. A test defined by a critical region C of size α is a uniformly most powerful test if it is a most powerful test against each simple alternative in H 1. The critical region C is called a uniformly most powerful critical region of size α Chi-square Goodness-of-Fit Tests Let an experiment have k mutually exclusive and exhaustive outcomes A 1,, A k. Denote p i = P (A i ), i = 1,, k. We would like to test the hypothesis H 0 : p i = p i0, i = 1,, k against all other alternative hypotheses H 1. Case 1: discrete distributions. Let the experiment be repeated n independent times. Let Y i be the observed number of times (frequency) that A i occurred. Then the expected frequency is np i0 (which should be at least 5). When H 0 is true, the test statistic is Q k 1 = k (Y i np i0 ) 2 χ 2 (k 1). np i0 i=1 The critical region is q k 1 χ 2 α(k 1), where α is the significance level. If there are d unknown parameters in the hypothesized distribution that need to be estimated from the given sample data, then we must calculate p i0 by using the estimates of the parameters. The test statistic is Q k 1 χ 2 (k 1 d) and the critical region would be q k 1 χ 2 α(k 1 d). Case 2: Continuous distributions. Let W be a continuous random variable with distribution function F (w). We would like to test H 0 : F (w) = F 0 (w) against all other alternatives H 1, where F 0 (w) is a known continuous distribution function. We partition the space of W into k class intervals: A 1 = (, a 1 ], A 2 = (a 1, a 2 ],, A k = (a k 1, ). Let p i = P (W A i ). Let Y i be the number of times that the observed values of W belong to A i, i = 1,, k in n independent repetitions of the experiment. Then, Y 1,, Y k have a multinomial distribution with parameters n, p 1,, p k 1. Let p i0 = P (W A i ) when the distribution function of W is F 0 (w). Then, H 0 is modified to be H 0 : p i = p i0, i = 1,, k. 7
8 H 0 is rejected if k (y i np i0 ) 2 q k 1 = χ 2 α(k 1 d), i=1 np i0 where d is the number of unknown parameters in F 0 (w) Contingency Tables Suppose that each of h independent experiments can result in one of the k mutually exclusive and exhaustive events A 1, A 2,..., A k. Let p ij = P (A i ), i = 1, 2,..., k, j = 1, 2,..., h. We want to test H 0 : p i1 = p i2 = = p ih = p i, i = 1, 2,..., k. We repeat the jth experiment n j independent times, and let Y 1j, Y 2j,.., Y kj denote the frequencies of the respective events A 1, A 2,..., A k. Under H 0, we estimate probabilities using h j=1 ˆp i = Y ij h j=1 n, i = 1, 2,..., k. j The chi-square test statistics is Q = h k j=1 i=1 (Y ij n j ˆp i ) 2 n j ˆp i. If the observed value q χ 2 α((h 1)(k 1)), then we reject H 0. Otherwise, we don t reject H 0. Test for Independence of Attributes of Classification: Suppose that a random experiment results in an outcome that can be classified by two different attributes. Assume that the first attribute is assigned to one and only one of k mutually exclusive and exhaustive events-say A 1, A 2,..., A k, and the second attribute is assigned to one and only one of h mutually exclusive and exhaustive events-say B 1, B 2,..., B h. Let the probability of A i B j be defined by p ij = P (A i B j ), i = 1, 2,..., k, j = 1, 2,..., h. The random experiment is to be repeated n independent times, and Y ij will denote the frequency of the event A i B j. Let p i = P (A i ), i = 1,..., k and p j = P (B j ), j = 1,..., h. We wish to test the independence of the A and B attibutes, namely H 0 : p ij = p i p j, i = 1,..., k, j = 1,..., h. Let Y i = h j=1 Y ij, i = 1,..., k (frequency of A i ) and Y j = k i=1 Y ij, j = 1,..., h (frequency of B j ). The chi-square test statistic is Q = h k j=1 i=1 [Y ij n(y i /n)(y j /n)] 2. n(y i /n)(y j /n) If the computed value q χ 2 α[(k 1)(h 1)], then we reject H 0 at the significance level α. Otherwsie, we don t reject H 0. 8
Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)
Summary of Chapter 7 (Sections 7.2-7.5) and Chapter 8 (Section 8.1) Chapter 7. Tests of Statistical Hypotheses 7.2. Tests about One Mean (1) Test about One Mean Case 1: σ is known. Assume that X N(µ, σ
More informationTUTORIAL 8 SOLUTIONS #
TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level
More informationStatistics. Statistics
The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,
More information1 Statistical inference for a population mean
1 Statistical inference for a population mean 1. Inference for a large sample, known variance Suppose X 1,..., X n represents a large random sample of data from a population with unknown mean µ and known
More informationStatistical Inference: Estimation and Confidence Intervals Hypothesis Testing
Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationProbability and Statistics Notes
Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationInstitute of Actuaries of India
Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the
More informationTHE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE
THE ROYAL STATISTICAL SOCIETY 004 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER II STATISTICAL METHODS The Society provides these solutions to assist candidates preparing for the examinations in future
More informationTopic 15: Simple Hypotheses
Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More information4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49
4 HYPOTHESIS TESTING 49 4 Hypothesis testing In sections 2 and 3 we considered the problem of estimating a single parameter of interest, θ. In this section we consider the related problem of testing whether
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationIntroductory Econometrics
Session 4 - Testing hypotheses Roland Sciences Po July 2011 Motivation After estimation, delivering information involves testing hypotheses Did this drug had any effect on the survival rate? Is this drug
More informationEconomics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,
Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem
More informationPartitioning the Parameter Space. Topic 18 Composite Hypotheses
Topic 18 Composite Hypotheses Partitioning the Parameter Space 1 / 10 Outline Partitioning the Parameter Space 2 / 10 Partitioning the Parameter Space Simple hypotheses limit us to a decision between one
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationStatistical Hypothesis Testing
Statistical Hypothesis Testing Dr. Phillip YAM 2012/2013 Spring Semester Reference: Chapter 7 of Tests of Statistical Hypotheses by Hogg and Tanis. Section 7.1 Tests about Proportions A statistical hypothesis
More informationINTERVAL ESTIMATION AND HYPOTHESES TESTING
INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,
More informationChapter 9: Hypothesis Testing Sections
Chapter 9: Hypothesis Testing Sections 9.1 Problems of Testing Hypotheses 9.2 Testing Simple Hypotheses 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.6 Comparing the Means of Two
More informationChapter 10. Chapter 10. Multinomial Experiments and. Multinomial Experiments and Contingency Tables. Contingency Tables.
Chapter 10 Multinomial Experiments and Contingency Tables 1 Chapter 10 Multinomial Experiments and Contingency Tables 10-1 1 Overview 10-2 2 Multinomial Experiments: of-fitfit 10-3 3 Contingency Tables:
More informationFormulas and Tables by Mario F. Triola
Copyright 010 Pearson Education, Inc. Ch. 3: Descriptive Statistics x f # x x f Mean 1x - x s - 1 n 1 x - 1 x s 1n - 1 s B variance s Ch. 4: Probability Mean (frequency table) Standard deviation P1A or
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More informationLecture Testing Hypotheses: The Neyman-Pearson Paradigm
Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12 Agenda Example:
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationCherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants
18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009
More information1 Hypothesis testing for a single mean
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationCHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:
CHAPTER 8 Test of Hypotheses Based on a Single Sample Hypothesis testing is the method that decide which of two contradictory claims about the parameter is correct. Here the parameters of interest are
More informationSolution: First note that the power function of the test is given as follows,
Problem 4.5.8: Assume the life of a tire given by X is distributed N(θ, 5000 ) Past experience indicates that θ = 30000. The manufacturere claims the tires made by a new process have mean θ > 30000. Is
More informationConfidence Intervals, Testing and ANOVA Summary
Confidence Intervals, Testing and ANOVA Summary 1 One Sample Tests 1.1 One Sample z test: Mean (σ known) Let X 1,, X n a r.s. from N(µ, σ) or n > 30. Let The test statistic is H 0 : µ = µ 0. z = x µ 0
More informationDefinition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.
Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed
More informationSTAT 135 Lab 7 Distributions derived from the normal distribution, and comparing independent samples.
STAT 135 Lab 7 Distributions derived from the normal distribution, and comparing independent samples. Rebecca Barter March 16, 2015 The χ 2 distribution The χ 2 distribution We have seen several instances
More informationTopic 19 Extensions on the Likelihood Ratio
Topic 19 Extensions on the Likelihood Ratio Two-Sided Tests 1 / 12 Outline Overview Normal Observations Power Analysis 2 / 12 Overview The likelihood ratio test is a popular choice for composite hypothesis
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationProbability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More information2.3 Analysis of Categorical Data
90 CHAPTER 2. ESTIMATION AND HYPOTHESIS TESTING 2.3 Analysis of Categorical Data 2.3.1 The Multinomial Probability Distribution A mulinomial random variable is a generalization of the binomial rv. It results
More informationLecture 7: Hypothesis Testing and ANOVA
Lecture 7: Hypothesis Testing and ANOVA Goals Overview of key elements of hypothesis testing Review of common one and two sample tests Introduction to ANOVA Hypothesis Testing The intent of hypothesis
More informationFormulas and Tables. for Elementary Statistics, Tenth Edition, by Mario F. Triola Copyright 2006 Pearson Education, Inc. ˆp E p ˆp E Proportion
Formulas and Tables for Elementary Statistics, Tenth Edition, by Mario F. Triola Copyright 2006 Pearson Education, Inc. Ch. 3: Descriptive Statistics x Sf. x x Sf Mean S(x 2 x) 2 s Å n 2 1 n(sx 2 ) 2 (Sx)
More informationLing 289 Contingency Table Statistics
Ling 289 Contingency Table Statistics Roger Levy and Christopher Manning This is a summary of the material that we ve covered on contingency tables. Contingency tables: introduction Odds ratios Counting,
More information4.5.1 The use of 2 log Λ when θ is scalar
4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis
More information11-2 Multinomial Experiment
Chapter 11 Multinomial Experiments and Contingency Tables 1 Chapter 11 Multinomial Experiments and Contingency Tables 11-11 Overview 11-2 Multinomial Experiments: Goodness-of-fitfit 11-3 Contingency Tables:
More informationComparison of Two Samples
2 Comparison of Two Samples 2.1 Introduction Problems of comparing two samples arise frequently in medicine, sociology, agriculture, engineering, and marketing. The data may have been generated by observation
More informationChapter 10. Hypothesis Testing (I)
Chapter 10. Hypothesis Testing (I) Hypothesis Testing, together with statistical estimation, are the two most frequently used statistical inference methods. It addresses a different type of practical problems
More informationHow do we compare the relative performance among competing models?
How do we compare the relative performance among competing models? 1 Comparing Data Mining Methods Frequent problem: we want to know which of the two learning techniques is better How to reliably say Model
More informationF79SM STATISTICAL METHODS
F79SM STATISTICAL METHODS SUMMARY NOTES 9 Hypothesis testing 9.1 Introduction As before we have a random sample x of size n of a population r.v. X with pdf/pf f(x;θ). The distribution we assign to X is
More informationSTAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015
STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationWhat is a Hypothesis?
What is a Hypothesis? A hypothesis is a claim (assumption) about a population parameter: population mean Example: The mean monthly cell phone bill in this city is μ = $42 population proportion Example:
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationTheory of Statistical Tests
Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationSTATISTICS SYLLABUS UNIT I
STATISTICS SYLLABUS UNIT I (Probability Theory) Definition Classical and axiomatic approaches.laws of total and compound probability, conditional probability, Bayes Theorem. Random variable and its distribution
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More informationHypothesis Testing Chap 10p460
Hypothesis Testing Chap 1p46 Elements of a statistical test p462 - Null hypothesis - Alternative hypothesis - Test Statistic - Rejection region Rejection Region p462 The rejection region (RR) specifies
More informationj=1 π j = 1. Let X j be the number
THE χ 2 TEST OF SIMPLE AND COMPOSITE HYPOTHESES 1. Multinomial distributions Suppose we have a multinomial (n,π 1,...,π k ) distribution, where π j is the probability of the jth of k possible outcomes
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More informationLECTURE 5 HYPOTHESIS TESTING
October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.
More informationGlossary for the Triola Statistics Series
Glossary for the Triola Statistics Series Absolute deviation The measure of variation equal to the sum of the deviations of each value from the mean, divided by the number of values Acceptance sampling
More informationTopic 10: Hypothesis Testing
Topic 10: Hypothesis Testing Course 003, 2016 Page 0 The Problem of Hypothesis Testing A statistical hypothesis is an assertion or conjecture about the probability distribution of one or more random variables.
More informationFormulas and Tables. for Essentials of Statistics, by Mario F. Triola 2002 by Addison-Wesley. ˆp E p ˆp E Proportion.
Formulas and Tables for Essentials of Statistics, by Mario F. Triola 2002 by Addison-Wesley. Ch. 2: Descriptive Statistics x Sf. x x Sf Mean S(x 2 x) 2 s Å n 2 1 n(sx 2 ) 2 (Sx) 2 s Å n(n 2 1) Mean (frequency
More informationHypothesis Testing: The Generalized Likelihood Ratio Test
Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random
More informationLecture 10: Generalized likelihood ratio test
Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual
More informationPerformance Evaluation and Comparison
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationBTRY 4090: Spring 2009 Theory of Statistics
BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)
More informationMath 152. Rumbos Fall Solutions to Assignment #12
Math 52. umbos Fall 2009 Solutions to Assignment #2. Suppose that you observe n iid Bernoulli(p) random variables, denoted by X, X 2,..., X n. Find the LT rejection region for the test of H o : p p o versus
More informationDefine characteristic function. State its properties. State and prove inversion theorem.
ASSIGNMENT - 1, MAY 013. Paper I PROBABILITY AND DISTRIBUTION THEORY (DMSTT 01) 1. (a) Give the Kolmogorov definition of probability. State and prove Borel cantelli lemma. Define : (i) distribution function
More informationThe Components of a Statistical Hypothesis Testing Problem
Statistical Inference: Recall from chapter 5 that statistical inference is the use of a subset of a population (the sample) to draw conclusions about the entire population. In chapter 5 we studied one
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More information2.6.3 Generalized likelihood ratio tests
26 HYPOTHESIS TESTING 113 263 Generalized likelihood ratio tests When a UMP test does not exist, we usually use a generalized likelihood ratio test to verify H 0 : θ Θ against H 1 : θ Θ\Θ It can be used
More informationS D / n t n 1 The paediatrician observes 3 =
Non-parametric tests Paired t-test A paediatrician measured the blood cholesterol of her patients and was worried to note that some had levels over 00mg/100ml To investigate whether dietary regulation
More informationThe University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80
The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple
More informationTopic 10: Hypothesis Testing
Topic 10: Hypothesis Testing Course 003, 2017 Page 0 The Problem of Hypothesis Testing A statistical hypothesis is an assertion or conjecture about the probability distribution of one or more random variables.
More informationFormulas and Tables for Elementary Statistics, Eighth Edition, by Mario F. Triola 2001 by Addison Wesley Longman Publishing Company, Inc.
Formulas and Tables for Elementary Statistics, Eighth Edition, by Mario F. Triola 2001 by Addison Wesley Longman Publishing Company, Inc. Ch. 2: Descriptive Statistics x Sf. x x Sf Mean S(x 2 x) 2 s 2
More informationHypothesis testing: theory and methods
Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable
More informationHYPOTHESIS TESTING: FREQUENTIST APPROACH.
HYPOTHESIS TESTING: FREQUENTIST APPROACH. These notes summarize the lectures on (the frequentist approach to) hypothesis testing. You should be familiar with the standard hypothesis testing from previous
More informationLecture 21: October 19
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 21: October 19 21.1 Likelihood Ratio Test (LRT) To test composite versus composite hypotheses the general method is to use
More informationReview of Statistics
Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and
More informationSTAT 135 Lab 5 Bootstrapping and Hypothesis Testing
STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,
More informationChapters 10. Hypothesis Testing
Chapters 10. Hypothesis Testing Some examples of hypothesis testing 1. Toss a coin 100 times and get 62 heads. Is this coin a fair coin? 2. Is the new treatment on blood pressure more effective than the
More informationSociology 6Z03 Review II
Sociology 6Z03 Review II John Fox McMaster University Fall 2016 John Fox (McMaster University) Sociology 6Z03 Review II Fall 2016 1 / 35 Outline: Review II Probability Part I Sampling Distributions Probability
More informationChapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments
Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments We consider two kinds of random variables: discrete and continuous random variables. For discrete random
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationMATH5745 Multivariate Methods Lecture 07
MATH5745 Multivariate Methods Lecture 07 Tests of hypothesis on covariance matrix March 16, 2018 MATH5745 Multivariate Methods Lecture 07 March 16, 2018 1 / 39 Test on covariance matrices: Introduction
More informationComparison of Two Population Means
Comparison of Two Population Means Esra Akdeniz March 15, 2015 Independent versus Dependent (paired) Samples We have independent samples if we perform an experiment in two unrelated populations. We have
More informationChapter 3. Comparing two populations
Chapter 3. Comparing two populations Contents Hypothesis for the difference between two population means: matched pairs Hypothesis for the difference between two population means: independent samples Two
More informationSpace Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses
Space Telescope Science Institute statistics mini-course October 2011 Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses James L Rosenberger Acknowledgements: Donald Richards, William
More informationHypothesis Testing One Sample Tests
STATISTICS Lecture no. 13 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 12. 1. 2010 Tests on Mean of a Normal distribution Tests on Variance of a Normal
More informationPart 1.) We know that the probability of any specific x only given p ij = p i p j is just multinomial(n, p) where p k1 k 2
Problem.) I will break this into two parts: () Proving w (m) = p( x (m) X i = x i, X j = x j, p ij = p i p j ). In other words, the probability of a specific table in T x given the row and column counts
More informationSTAT 830 Hypothesis Testing
STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These
More informationCHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities:
CHAPTER 9, 10 Hypothesis Testing Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities: The person is guilty. The person is innocent. To
More informationTopic 17: Simple Hypotheses
Topic 17: November, 2011 1 Overview and Terminology Statistical hypothesis testing is designed to address the question: Do the data provide sufficient evidence to conclude that we must depart from our
More informationhttp://www.math.uah.edu/stat/hypothesis/.xhtml 1 of 5 7/29/2009 3:14 PM Virtual Laboratories > 9. Hy pothesis Testing > 1 2 3 4 5 6 7 1. The Basic Statistical Model As usual, our starting point is a random
More informationSTAT 830 Hypothesis Testing
STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two
More information