ML Testing (Likelihood Ratio Testing) for non-gaussian models
|
|
- Posy Goodwin
- 5 years ago
- Views:
Transcription
1 ML Testing (Likelihood Ratio Testing) for non-gaussian models Surya Tokdar
2 ML test in a slightly different form Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) = {θ : l x (θ) max θ Θ l x (θ) c2 2 } ML test = reject H 0 if B c (x) Θ 0 = Same as: reject H 0 if Or equivalently: reject H 0 if max l x(θ) max l x (θ) c 2 /2 θ Θ θ Θ 0 Λ(x) = max θ Θ L x (θ) max θ Θ0 L x (θ) k 1 / 1
3 Size calculation Need to calculate max θ Θ0 P [X θ] (Λ(X ) > k) IID Will work for models of the form: X i g(x i θ), θ Θ And hypothesis: H 0 : A T θ = a 0 for a given p r matrix A and given scalar a 0 (typically a 0 = 0) Under regularity conditions: 2 log Λ(X ) d χ 2 (r) when IID X i g(x i θ 0 ) with θ 0 Θ 0. Follows from asymptotic normality of ˆθ MLE (X ) 2 / 1
4 Asymptotic normality of MLE Theorem. Assume the following (Cramér) conditions 1. θ θ log g(x θ) is twice continuously differentiable. 2. θ g( θ) is identifiable θ 3 log g(x i θ) < h(x), θ nbhd(θ 0 ), E [X θ0 ]h(x 1 ) < 4. ˆθ MLE (x) is the unique solution of l x (θ) = 0 for every x. IID If X i g(x i θ 0 ) then n(ˆθ MLE θ 0 ) d N (0, {I F 1 (θ 0 )} 1) where I F 1 (θ) = E [X i θ] 2 θ 2 log g(x 1 θ) = Fisher information at θ. 3 / 1
5 A sketch of a proof First order Taylor approximation: 0 = 1 lx (ˆθ MLE ) 1 lx (θ 0 ) + n l X (θ 0 ) T (ˆθ MLE θ 0 ) n n n Write T i = θ log g(x i θ). ETi = 0, VarT i = I F 1 (θ 0) 1 n l X (θ 0 ) = n( T 0) d N(0, I F 1 (θ 0)) [by CLT] Write W i = 2 log g(x θ 2 i θ) at θ = θ 0 EWi = I1 F (θ 0) l X (θ 0) n = W p I1 F (θ 0) Rest is Slutksy s theorem 4 / 1
6 Three large n approximations to Λ(x) Assume θ 0 Θ 0, i.e., A T θ 0 = a 0 Notation: ˆθ H0 = argmax θ Θ0 l x (θ) (must have A T ˆθH0 = a 0 ) Quadratic, Wald, Rao approximations F (X ) = (A T ˆθ MLE a 0 ) T {A T Ix 1 A} 1 (A T ˆθ MLE a 0 ) 2 log Λ(X ) W (X ) = n(a T ˆθ MLE a 0) T {A T I1 F (ˆθ MLE ) 1 A} 1 (A T ˆθ MLE a 0) S(X ) = 1 l n X (ˆθ H0 ) T {I1 F (ˆθ H0 )} 1 l X (ˆθ H0 ) 2 log Λ(X ) F (X ) p 0, etc. Each of 2 log Λ(X ), F (X ), W (X ), S(X ) d χ 2 (r) 5 / 1
7 Sketches of proof if you cared - part I 2 log Λ(X ) F (X ) follows from Quadratic approximation: l x (θ) l x (ˆθ MLE ) 1 2 (θ ˆθ MLE ) T I X (θ ˆθ MLE ) and the corresponding profile log-likelihood of η = A T θ: l x(η) l x (ˆθ MLE ) 1 2 (η AT ˆθMLE ) T {A T I 1 X A} 1 (η A T ˆθMLE ) and by noting Λ(x) = l x(a T ˆθ H0 ) l x(a T ˆθ MLE ) 2 log Λ(X ) W (X ) follows similarly, use I X ni F 1 (ˆθ MLE ). 6 / 1
8 Sketches of proof if you cared - part II 2 log Λ(X ) S(X ) follows because 0 = l x (ˆθ MLE ) l x (ˆθ H0 ) + l x (ˆθ H0 ) T (ˆθ MLE ˆθ H0 ) l x (ˆθ H0 ) ni F 1 (ˆθ H0 ) And so ˆθ MLE ˆθ H0 I F 1 (ˆθ H0 ) 1 l x (ˆθ H0 ) Plug this into the original quadratic approximation and use I X ni F 1 (θ 0) ni F 1 (ˆθ H0 ) because ˆθ MLE θ 0 7 / 1
9 Sketches of proof if you cared - part III It is easiest to prove W (X ) d χ 2 (r) By asymptotic normality of MLE and Slutsky s theorem W (X ) d Z T {A T I1 F (θ 0 ) 1 A} 1 Z where Z N r (0, {A T I1 F (θ 0) 1 A}) But Z N r (0, Σ) means Z T Σ 1 Z χ 2 (r). 8 / 1
10 Back to size calculation Recall: Each of 2 log Λ(X ), F (X ), W (X ), S(X ) d χ 2 (r) So each of the following test procedures has size α ML/LRT: Reject H0 if 2 log Λ(x) > q χ 2(1 α, r) Quadratic: Reject H 0 if F (x) > q χ 2(1 α, r) Wald: Reject H0 if W (x) > q χ 2(1 α, r) Rao: Reject H 0 if S(x) > q χ 2(1 α, r) Here q χ 2(u, r) = F 1 χ 2 (u, r) where F χ 2(x, r) = CDF of χ 2 (r) Also, one can calculate p-values as 1 F χ 2(2 log Λ(x), r) etc. 9 / 1
11 Which one to use? Depends on which is easier to compute To get W (x) you don t need ˆθ H0, only need ˆθ MLE. To get S(x) you only need ˆθ H0, don t need ˆθ MLE. In many cases F (X ) = W (X ) because I X = ni F 1 (ˆθ MLE ) This happens in most generalized linear models: IND Y i g(y i β, z i ) = h(z i, y i )e ξ(β)y i zi T β A(β,z i ) and Wald s tests are popular choices Multinomial model of category counts, S(X ) = Pearson s-χ 2 10 / 1
12 Example: low birthweight Natality data n = 500 records on US births in June Y i = 1 if i-th birth record has birthweight < 2500g, Y i = 0 otherwise. z i = (Cigarettes i, Black i ) records daily number of cigarettes and the race of the mother. Model log P(Y i = 1) 1 P(Y i = 1) = β 1 + β 2 Cigarettes i + β 3 Black i + β 4 Cigarettes i Black i. Can compute ˆβ MLE and I X (numerically). 11 / 1
13 MLE and curvature From glm() function in R ˆβ MLE = I 1 X = How would you test the null hypothesis that for a black mother, probability of low birthweight depends on the number of cigarettes? For a non-black mother? How would you test that this dependence is different for black and non-black mothers? 12 / 1
14 Categorical data & Multinomial models Data: Counts of categories formed by one or more attributes Tables could be one-way, two-way, etc., depending on how many attributes are used to decide categories. Hair color Eye color Blue Green Brown Black Total Blonde Red Brown Black Total / 1
15 Multinomial model Data are k category counts of n objects X = (X 1,, X k ), X j 0, X 1 + X X k = n. Model X Multinomial(n, p) where p = (p 1,, p k ) k k is k-dim simplex, contains k-dim prob vectors b = (b 1,, b k ), i.e., b j 0 j and j b j = 1) Multinomial(n, p) has pmf ( ) n f (x p) = p x 1 1 x 1 x px k k k for x = (x 1,, x k ) with x i 0 and i x i = n, f (x p) = 0 otherwise. 14 / 1
16 Multinomial model (contd.) This is an extension of the binomial distributions. In X we record the counts from n independent trials with k outcomes with probabilities given by p. Consequently X j Bin(n, p j ) for any single category 1 j k. Similarly for two categories j 1 j 2, X j1 + X j2 Bin(n, p j1 + p j2 ), and so on. 15 / 1
17 Multinomial model (contd.) For two-way tables, with k 1 rows and k 2 columns, we can write X and p either as k 1 k 2 -dim vectors, or more commonly as k 1 k 2 matrices. Even when they re written as matrices, we ll write X Multinomial(n, p) to mean that the multinomial distribution is placed on the vector forms of X and p. For multi-way table we ll use arrays of appropriate dimensions to represent X and p. 16 / 1
18 MLE To obtain the MLE of p based on data X = x, we can maximize the log-likelihood function in p with an additional Lagrange component to account for the constraint k j=1 p j = 1: l x (p, λ) = const + The solution in p equals: ˆp MLE = k k x j log p j + λ( p j 1) j=1 ( x1 n,, x ) k n j=1 17 / 1
19 Hypothesis testing: Mendel s peas Mendel, the founder of modern genetics, studied how physical characteristics are inherited in plants. His studies led him to propose the laws of segregation and independent assortment. We ll test this in a simple context. Under Mendel s laws, when pure round-yellow and pure green-wrinkled pea plants are cross-bred, the next generation of plant seeds should exhibit a 9:3:3:1 ratio of round-yellow, round-green, wrinkled-yellow and wrinkled-green combinations of shape and color. In a sample of 556 plants from the next generation the observed counts for these combinations are (315, 108, 101, 32). Does the data support Mendel s laws? 18 / 1
20 Formalization Data X = (X 1, X 2, X 3, X 4 ) giving the category counts of the four types of plants Model X Multinomial(n = 556, p), p 4 Test H 0 : p = ( 9 16, 3 16, 3 16, 1 16 ) 19 / 1
21 Hardy-Weinberg equilibrium The spotting on the wings of Scarlet tiger moths are controlled by a gene that comes in two varieties (alleles) whose combinations (moths have pairs of chromosomes) produce three varieties of spotting pattern: white spotted, little spotted and intermediate. If the moth population is in Hardy-Weinberg equilibrium (no current selection drift), then these varieties should be in the ratio a 2 : (1 a) 2 : 2a(1 a), where a (0, 1) denotes the abundance of the dominant white spotting allele. In a sample of 1612 moths, the three varieties were counted to be 1469, 5 and 138. Is the moth population in HW equilibrium? 20 / 1
22 Formalization Data X = (X 1, X 2, X 3 ) the category counts of the three spotting patterns, model X Multinomial(n = 1612, p), p 3, Test whether H 0 : p HW 3, HW 3 is a subset of 3 containing all vectors of the form (a 2, (1 a) 2, 2a(1 a)) for some a (0, 1). 21 / 1
23 Hair and eye color Are hair color and eye color two independent attributes? In this case, writing p in the matrix form p = ((p ij )), 1 i k 1 and 1 j k 2, we want to test if p ij = p (1) i p (2) j for some p (1) k1 and p (2) k2. It s elementary that if p ij factors as above, then p (1) i = k 2 j=1 p ij and p (2) j = k 1 i=1 p ij. Writing the row and column totals as p i and p j, the test of independence is often represented by H 0 : p ij = p i p j, i, j We ll use I k 1,k 2 to denote this set. 22 / 1
24 ML testing dim( k ) = k 1, as vector p k satisfy k i=1 p i = 1. Hypotheses H 0 : p Θ 0, Θ 0 is a sub-simplex of k of dimension q < k. Mendel s peas: Θ0 = ( 9 16, 3 16, 3 16, 1 16 ), q = 0 HW equilibrium: Θ0 = HW 3, q = 1 Independence: Θ 0 = I k 1,k 2, q = k 1 + k 2 2. Could use any of ML, quadratic, Wald or Rao tests 23 / 1
25 Pearson s χ 2 tests Rao s test statistics boils down to a very convenient form: S(X ) = k (X j nˆp H0,j) 2 j=1 nˆp H0,j = (Observed Expected) 2 Expected For categorical data, this was earlier discovered by Pearson who also ascertained its approximate χ 2 distribution. Because this is Rao statistic, we have S(X ) d χ 2 (k 1 q) 24 / 1
26 Example 1: point null For testing H 0 : p = p 0 against p p 0, where p 0 = (p 0,1,, p 0,k ) k is a fixed probability vector of interest (as in Mendel s peas example), then S(X ) = k (X j np 0,j ) 2 np 0,j j=1 which is asymptotically χ 2 (k 1 0) = χ 2 (k 1). Size-α Pearson s test rejects H 0 if S(X ) > q χ 2(1 α, k 1). 25 / 1
27 Example 2: parametric form HW test: H 0 : p 0 = (η 2, η(1 η), η(1 η), (1 η) 2 ), for some 0 < η < 1 To compute ˆp H0 it is equivalent to write the likelihood function in η and maximize: L X (η) = const. {η 2 } x 1 {η(1 η)} x 2 {η(1 η)} x 3 {(1 η) 2 } x 4 = const. η 2x 1+x 2 +x 3 (1 η) x 2+x 3 +2x 4 and so ˆη MLE = 2x 1+x 2 +x 3 2n. 26 / 1
28 Example 2: parametric form (contd.) Once we have ˆη MLE, we can construct ˆp H0 = (ˆη 2 MLE, ˆη MLE (1 ˆη MLE ), ˆη MLE (1 ˆη MLE ), (1 ˆη MLE ) 2 ) and evaluate S(X ). Because Θ 0 has dimension q = 1 (only a single number η needs to be known), the asymptotic distribution of S(X ) is χ 2 (k 2). The same calculations carry through for a more general parametric form: H 0 : p = (g 1 (η),, g k (η)) where η E is q-dim vector and g 1 (η),, g k (η) are functions such that for every η E, (g 1 (η),, g k (η)) k. 27 / 1
29 Example 3: independence Consider testing H 0 : p ij = p i p j, i, j To get ˆp H0, we write the likelihood function in terms of p i, 1 i k 1 and p j, 1 j k 2 : k 1 k 2 L x (p 1,, p k1, p 1,, p k2 ) = const. (p i p j ) x ij i=1 j=1 { k1 = const. i=1 p x i i }{ k2 where x i = k 2 j=1 x ij and x j = k 1 i=1 x ij are the margin counts of our two-way table. j=1 p x j j } 28 / 1
30 Example 3: independence (contd.) Because (p 1,, p k1 ) k1 maximizer is given by: and (p 1,, p k2 ) k2, the ˆp i = x i n, ˆp j = x j n, 1 i k 1, 1 j k 2 And so ˆp H0 has coordinates: ˆp H0,ij = x i x j n 2 giving k 1 k 2 S(X ) = i=1 j=1 (X ij X i X j n ) 2 X i X j n Because dim(θ 0 ) = q = k k 2 1, S(X ) d χ 2 (k 1 k 2 1 k k 2 + 1) = χ 2 ((k 1 1)(k 2 1)). 29 / 1
31 Hair-eye color Hair color Eye color Blue Green Brown Black Total Blonde 20 (13.9) 15 (13.6) 18 (28.4) 14 (11.0) 67 Red 11(8.5) 4 (8.3) 24 (17.4) 2 (6.7) 41 Brown 9 (15.4) 11 (15.1) 36 (31.4) 18 (12.2) 74 Black 8 (10.2) 17 (10.0) 20 (20.8) 4 (8.1) 49 Total S(x) = So the p-value is 1 F (4 1)(4 1) (30.9) = 1 F 9 (30.9) / 1
Theorem. Assume the following (Cramér) conditions 1. θ θ
ML test i a slightly differet form ML Testig (Likelihood Ratio Testig) for o-gaussia models Surya Tokdar Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) ={θ : l x (θ) max θ Θ l x (θ) c2
More informationPart 1.) We know that the probability of any specific x only given p ij = p i p j is just multinomial(n, p) where p k1 k 2
Problem.) I will break this into two parts: () Proving w (m) = p( x (m) X i = x i, X j = x j, p ij = p i p j ). In other words, the probability of a specific table in T x given the row and column counts
More informationLecture 10: Generalized likelihood ratio test
Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual
More informationStatistics 3858 : Contingency Tables
Statistics 3858 : Contingency Tables 1 Introduction Before proceeding with this topic the student should review generalized likelihood ratios ΛX) for multinomial distributions, its relation to Pearson
More informationLecture 21: October 19
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 21: October 19 21.1 Likelihood Ratio Test (LRT) To test composite versus composite hypotheses the general method is to use
More information2.3 Analysis of Categorical Data
90 CHAPTER 2. ESTIMATION AND HYPOTHESIS TESTING 2.3 Analysis of Categorical Data 2.3.1 The Multinomial Probability Distribution A mulinomial random variable is a generalization of the binomial rv. It results
More informationTUTORIAL 8 SOLUTIONS #
TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationComposite Hypotheses and Generalized Likelihood Ratio Tests
Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More informationDA Freedman Notes on the MLE Fall 2003
DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar
More informationClassical Inference for Gaussian Linear Models
Classical Inference for Gaussian Linear Models Surya Tokdar Gaussian linear model Data (z i, Y i ), i = 1,, n, dim(z i ) = p Model Y i = z T i Parameters: β R p, σ 2 > 0 Inference needed on η = a T β IID
More informationChapter 10. Prof. Tesler. Math 186 Winter χ 2 tests for goodness of fit and independence
Chapter 10 χ 2 tests for goodness of fit and independence Prof. Tesler Math 186 Winter 2018 Prof. Tesler Ch. 10: χ 2 goodness of fit tests Math 186 / Winter 2018 1 / 26 Multinomial test Consider a k-sided
More informationLecture 4 September 15
IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationStatistical Inference: Maximum Likelihood and Bayesian Approaches
Statistical Inference: Maximum Likelihood and Bayesian Approaches Surya Tokdar From model to inference So a statistical analysis begins by setting up a model {f (x θ) : θ Θ} for data X. Next we observe
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationGoodness of Fit Goodness of fit - 2 classes
Goodness of Fit Goodness of fit - 2 classes A B 78 22 Do these data correspond reasonably to the proportions 3:1? We previously discussed options for testing p A = 0.75! Exact p-value Exact confidence
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationChapter 10. Chapter 10. Multinomial Experiments and. Multinomial Experiments and Contingency Tables. Contingency Tables.
Chapter 10 Multinomial Experiments and Contingency Tables 1 Chapter 10 Multinomial Experiments and Contingency Tables 10-1 1 Overview 10-2 2 Multinomial Experiments: of-fitfit 10-3 3 Contingency Tables:
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationOne-Way Tables and Goodness of Fit
Stat 504, Lecture 5 1 One-Way Tables and Goodness of Fit Key concepts: One-way Frequency Table Pearson goodness-of-fit statistic Deviance statistic Pearson residuals Objectives: Learn how to compute the
More informationNormal distribution We have a random sample from N(m, υ). The sample mean is Ȳ and the corrected sum of squares is S yy. After some simplification,
Likelihood Let P (D H) be the probability an experiment produces data D, given hypothesis H. Usually H is regarded as fixed and D variable. Before the experiment, the data D are unknown, and the probability
More informationDiscrete Multivariate Statistics
Discrete Multivariate Statistics Univariate Discrete Random variables Let X be a discrete random variable which, in this module, will be assumed to take a finite number of t different values which are
More informationMultiple Sample Categorical Data
Multiple Sample Categorical Data paired and unpaired data, goodness-of-fit testing, testing for independence University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html
More informationMATH4427 Notebook 2 Fall Semester 2017/2018
MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationChapter 4. Theory of Tests. 4.1 Introduction
Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationChapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments
Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments We consider two kinds of random variables: discrete and continuous random variables. For discrete random
More informationMaximum likelihood estimation
Maximum likelihood estimation Guillaume Obozinski Ecole des Ponts - ParisTech Master MVA Maximum likelihood estimation 1/26 Outline 1 Statistical concepts 2 A short review of convex analysis and optimization
More informationSummary of Chapters 7-9
Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two
More information11-2 Multinomial Experiment
Chapter 11 Multinomial Experiments and Contingency Tables 1 Chapter 11 Multinomial Experiments and Contingency Tables 11-11 Overview 11-2 Multinomial Experiments: Goodness-of-fitfit 11-3 Contingency Tables:
More informationj=1 π j = 1. Let X j be the number
THE χ 2 TEST OF SIMPLE AND COMPOSITE HYPOTHESES 1. Multinomial distributions Suppose we have a multinomial (n,π 1,...,π k ) distribution, where π j is the probability of the jth of k possible outcomes
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationIntroduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007
Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0
More informationTesting Independence
Testing Independence Dipankar Bandyopadhyay Department of Biostatistics, Virginia Commonwealth University BIOS 625: Categorical Data & GLM 1/50 Testing Independence Previously, we looked at RR = OR = 1
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationLoglikelihood and Confidence Intervals
Stat 504, Lecture 2 1 Loglikelihood and Confidence Intervals The loglikelihood function is defined to be the natural logarithm of the likelihood function, l(θ ; x) = log L(θ ; x). For a variety of reasons,
More informationHEREDITY: Objective: I can describe what heredity is because I can identify traits and characteristics
Mendel and Heredity HEREDITY: SC.7.L.16.1 Understand and explain that every organism requires a set of instructions that specifies its traits, that this hereditary information. Objective: I can describe
More information1 Mendel and His Peas
CHAPTER 3 1 Mendel and His Peas SECTION Heredity BEFORE YOU READ After you read this section, you should be able to answer these questions: What is heredity? How did Gregor Mendel study heredity? National
More informationSTAT Chapter 13: Categorical Data. Recall we have studied binomial data, in which each trial falls into one of 2 categories (success/failure).
STAT 515 -- Chapter 13: Categorical Data Recall we have studied binomial data, in which each trial falls into one of 2 categories (success/failure). Many studies allow for more than 2 categories. Example
More informationLing 289 Contingency Table Statistics
Ling 289 Contingency Table Statistics Roger Levy and Christopher Manning This is a summary of the material that we ve covered on contingency tables. Contingency tables: introduction Odds ratios Counting,
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More informationCherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants
18.650 Statistics for Applications Chapter 5: Parametric hypothesis testing 1/37 Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009
More informationHypothesis testing: theory and methods
Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable
More informationECE 275A Homework 7 Solutions
ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationIntroduction to Maximum Likelihood Estimation
Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationLECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.
Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationCh 11.Introduction to Genetics.Biology.Landis
Nom Section 11 1 The Work of Gregor Mendel (pages 263 266) This section describes how Gregor Mendel studied the inheritance of traits in garden peas and what his conclusions were. Introduction (page 263)
More informationEcon 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE
Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Eric Zivot Winter 013 1 Wald, LR and LM statistics based on generalized method of moments estimation Let 1 be an iid sample
More informationSTAT 730 Chapter 4: Estimation
STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum
More informationGraduate Econometrics I: Maximum Likelihood I
Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood
More informationReview of Discrete Probability (contd.)
Stat 504, Lecture 2 1 Review of Discrete Probability (contd.) Overview of probability and inference Probability Data generating process Observed data Inference The basic problem we study in probability:
More informationCase-Control Association Testing. Case-Control Association Testing
Introduction Association mapping is now routinely being used to identify loci that are involved with complex traits. Technological advances have made it feasible to perform case-control association studies
More information4.5.1 The use of 2 log Λ when θ is scalar
4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis
More informationClass Copy! Return to teacher at the end of class! Mendel's Genetics
Class Copy! Return to teacher at the end of class! Mendel's Genetics For thousands of years farmers and herders have been selectively breeding their plants and animals to produce more useful hybrids. It
More informationSignificance Tests. Review Confidence Intervals. The Gauss Model. Genetics
15.0 Significance Tests Review Confidence Intervals The Gauss Model Genetics Significance Tests 1 15.1 CI Review The general formula for a two-sided C% confidence interval is: L, U = pe ± se cv (1 C)/2
More information10. Composite Hypothesis Testing. ECE 830, Spring 2014
10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters
More informationLecture 17: Likelihood ratio and asymptotic tests
Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f
More informationChi-squared (χ 2 ) (1.10.5) and F-tests (9.5.2) for the variance of a normal distribution ( )
Chi-squared (χ ) (1.10.5) and F-tests (9.5.) for the variance of a normal distribution χ tests for goodness of fit and indepdendence (3.5.4 3.5.5) Prof. Tesler Math 83 Fall 016 Prof. Tesler χ and F tests
More informationStatistics and econometrics
1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationMaximum Likelihood (ML) Estimation
Econometrics 2 Fall 2004 Maximum Likelihood (ML) Estimation Heino Bohn Nielsen 1of32 Outline of the Lecture (1) Introduction. (2) ML estimation defined. (3) ExampleI:Binomialtrials. (4) Example II: Linear
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationMaximum Likelihood Large Sample Theory
Maximum Likelihood Large Sample Theory MIT 18.443 Dr. Kempthorne Spring 2015 1 Outline 1 Large Sample Theory of Maximum Likelihood Estimates 2 Asymptotic Results: Overview Asymptotic Framework Data Model
More informationHeredity.. An Introduction Unit 5: Seventh Grade
Heredity.. An Introduction Unit 5: Seventh Grade Why don t you look like a rhinoceros? The answer seems simple --- neither of your parents is a rhinoceros (I assume). But there is more to this answer than
More informationGenetics_2011.notebook. May 13, Aim: What is heredity? Homework. Rd pp p.270 # 2,3,4. Feb 8 11:46 PM. Mar 25 1:15 PM.
Aim: What is heredity? LE1 3/25/11 Do Now: 1.Make a T Chart comparing and contrasting mitosis & meiosis. 2. Have your lab out to be collected Homework for Tuesday 3/29 Read pp. 267 270 p.270 # 1,3 Vocabulary:
More informationSTAT 536: Genetic Statistics
STAT 536: Genetic Statistics Tests for Hardy Weinberg Equilibrium Karin S. Dorman Department of Statistics Iowa State University September 7, 2006 Statistical Hypothesis Testing Identify a hypothesis,
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More information4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49
4 HYPOTHESIS TESTING 49 4 Hypothesis testing In sections 2 and 3 we considered the problem of estimating a single parameter of interest, θ. In this section we consider the related problem of testing whether
More informationδ -method and M-estimation
Econ 2110, fall 2016, Part IVb Asymptotic Theory: δ -method and M-estimation Maximilian Kasy Department of Economics, Harvard University 1 / 40 Example Suppose we estimate the average effect of class size
More informationOutline for today s lecture (Ch. 14, Part I)
Outline for today s lecture (Ch. 14, Part I) Ploidy vs. DNA content The basis of heredity ca. 1850s Mendel s Experiments and Theory Law of Segregation Law of Independent Assortment Introduction to Probability
More informationIntroduction to population genetics & evolution
Introduction to population genetics & evolution Course Organization Exam dates: Feb 19 March 1st Has everybody registered? Did you get the email with the exam schedule Summer seminar: Hot topics in Bioinformatics
More informationSTA 114: Statistics. Notes 21. Linear Regression
STA 114: Statistics Notes 1. Linear Regression Introduction A most widely used statistical analysis is regression, where one tries to explain a response variable Y by an explanatory variable X based on
More informationStatistics. Statistics
The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,
More information8. Genetic Diversity
8. Genetic Diversity Many ways to measure the diversity of a population: For any measure of diversity, we expect an estimate to be: when only one kind of object is present; low when >1 kind of objects
More informationReview. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 770: Categorical Data Analysis
Review Timothy Hanson Department of Statistics, University of South Carolina Stat 770: Categorical Data Analysis 1 / 22 Chapter 1: background Nominal, ordinal, interval data. Distributions: Poisson, binomial,
More informationGeneralized Linear Models Introduction
Generalized Linear Models Introduction Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Linear Models For many problems, standard linear regression approaches don t work. Sometimes,
More informationName Class Date. Pearson Education, Inc., publishing as Pearson Prentice Hall. 33
Chapter 11 Introduction to Genetics Chapter Vocabulary Review Matching On the lines provided, write the letter of the definition of each term. 1. genetics a. likelihood that something will happen 2. trait
More informationUnit 9: Inferences for Proportions and Count Data
Unit 9: Inferences for Proportions and Count Data Statistics 571: Statistical Methods Ramón V. León 1/15/008 Unit 9 - Stat 571 - Ramón V. León 1 Large Sample Confidence Interval for Proportion ( pˆ p)
More informationOPTIMALITY AND STABILITY OF SYMMETRIC EVOLUTIONARY GAMES WITH APPLICATIONS IN GENETIC SELECTION. (Communicated by Yang Kuang)
MATHEMATICAL BIOSCIENCES doi:10.3934/mbe.2015.12.503 AND ENGINEERING Volume 12, Number 3, June 2015 pp. 503 523 OPTIMALITY AND STABILITY OF SYMMETRIC EVOLUTIONARY GAMES WITH APPLICATIONS IN GENETIC SELECTION
More informationSome General Types of Tests
Some General Types of Tests We may not be able to find a UMP or UMPU test in a given situation. In that case, we may use test of some general class of tests that often have good asymptotic properties.
More informationReview of One-way Tables and SAS
Stat 504, Lecture 7 1 Review of One-way Tables and SAS In-class exercises: Ex1, Ex2, and Ex3 from http://v8doc.sas.com/sashtml/proc/z0146708.htm To calculate p-value for a X 2 or G 2 in SAS: http://v8doc.sas.com/sashtml/lgref/z0245929.htmz0845409
More informationLet us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided
Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or
More informationOne-Sample Numerical Data
One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationf(y θ) = g(t (y) θ)h(y)
EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationLecture 01: Introduction
Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction
More informationRecall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n
Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationBTRY 4090: Spring 2009 Theory of Statistics
BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)
More information