Optimal Tests of Hypotheses (Hogg Chapter Eight)

Size: px
Start display at page:

Download "Optimal Tests of Hypotheses (Hogg Chapter Eight)"

Transcription

1 Optimal Tests of Hypotheses Hogg hapter Eight STAT 406-0: Mathematical Statistics II Spring Semester 06 ontents Most Powerful Tests. Review of Hypothesis Testing The Neyman-Pearson Lemma Example: Gamma Distribution Uniformly Most Powerful Tests Example: One-Sided Hypothesis Example: Two-Sided Hypothesis Example: omposite Null Hypothesis The Karlin-Rubin Theorem Example: Exponential Family Likelihood Ratio Tests and UMP Unbiased Tests 7. Gamma Example ontinued Unbiased Tests Illustration for Gamma example omposite Hypothesis Testing with Priors 0 3. Example: Mean of a Normal Distribution ase : Uniform Prior opyright 06, John T. Whelan, and all that 3.. ase : Non-Uniform Prior Tuesday 6 April 06 Read Section 8. of Hogg Most Powerful Tests We turn now to another application of likelihood methods: as tests for hypotheses. First, a reminder of some of the notation and definitions associated with classical frequentist hypothesis testing.. Review of Hypothesis Testing In the classical framework, we define a test to reject or not reject a null hypothesis H 0 in the context of an alternative hypothesis H, based on a realization x of the n-dimensional random vector X which is often but not always a sample drawn from a univariate distribution. We refer to the support space for the whole sample as S, i.e., X S. We can define the test as partitioning the sample space into a critical region and its compelement.

2 If x S we reject H 0, while if x / i.e., x c we do not reject H 0. Strictly speaking, we should not refer to the outcome of the test as accepting one hypothesis or the other. We often think of this in terms of a parametrized distribution f X x; θ and define H 0 to specify a value θ = θ 0 Hogg calls this θ for the parameter and H to specify some other value θ = θ which Hogg calls θ. Since each hypothesis only specifies a single value for θ, we call these point hypotheses, and the parameter formalism is actually unneccessary; f X x; θ 0 = f X x H 0 and f X x; θ = f X x H can just be thought of as two different probsbility distributions relevant to H 0 and H, respectively. We define the size α of the critical region, also known as the significance or false-alarm probability of the test as the probability that we will reject H 0 if it is true: α = P X H 0 = f X x H 0 d n x. The power of the test is the probability that we ll reject H 0 if H is true: γ = P X H = f X x H d n x. We can also write this in terms of β = γ, which is known as the false dismissal probability. Given different tests with the same false-alarm probability α, we d naturally rather use the one with the lowest false-dismissal probability β = γ, i.e., the highest power γ. This is known as the most powerful test.. The Neyman-Pearson Lemma There is a theorem, usually known as the Neyman-Pearson lemma, that shows how the most powerful test to of one point hypothesis H 0 against another H can be constructed from the likelihood ratio Λx = f Xx H 0 f X x H = Lθ 0; x Lθ ; x.3 We define so that x if and only if Λx k where k is defined by P ΛX k H 0 = f X x H 0 d n x = α.4 which ensures that the critical region is of size α. I.e., we reject H 0 if and only if Λx k. The Neyman-Pearson lemma states that the power of this test γ = P ΛX k H = f X x H d n x.5 is greater than or equal to the power of any other test with the same significance. I.e., if A is some other critical region with size α, so that f X x H 0 d n x = α.6 the Neyman-Pearson lemma says that γ = f X x H d n x A A f X x H d n x.7 The demonstration of the Neyman-Pearson lemma involves breaking up the regions and A in terms of their overlap A. Evidentally, we can write = A c A A = c A A.8a.8b

3 The contribution to both α and γ from A cancel out of any comparison between and A. So the Neyman-Pearson lemma is equivalent to the condition that γ f X x H d n x A = A c f X x H d n x c A If we can prove that, we ve proved the lemma Now, by definition so while so A c c A f X x H 0 f X x H f X x H d n x f X x H d n x.9 k for x.0 f X x H 0 d n x. k A c f X x H 0 f X x H k for x c. f X x H d n x k c A f X x H 0 d n x.3 But since the tests defined by and A both have the same significance α, the integrals of f X x H 0 over the non-overlapping regions must be the same: α f X x H 0 d n x A = A c f X x H 0 d n x = c A f X x H 0 d n x.4 so the right-hand sides of.3 and. must be equal, which means f X x H d n x f X x H d n x.5 A c c A which, as we ve argued above, means that the power of the likelihood ratio test defined by is greater than or equal to that defined by A..3 Example: Gamma Distribution fx; θ = x5 θ 5 e x/θ, 0 < x <. Let H 0 : θ = and H : θ =. Likelihood is so likelihood ratio is Λ = Lθ; x = n i= x i 4 exp θ 5n n x i /θ i=.6 [ L; x L; x = 5n exp ] n x i = 5n e y/.7 i= where y = n i= x i. Note that the likelihood ratio depends only on the sufficient statistic y. A moment s thought will show that this will be true in general. If there are one or more sufficient statistics for the parameters which distinguish H 0 from H, the likelihood ratio must be a function of those alone. Note that in this case, Y is Gamman, of H 0 is true and Gamman, of H is true, so we can figure out the significance and power of the test by using the percentiles of the Gamma or equivalently chi-square distribution. 3

4 Thursday 8 April 06 Read Section 8. of Hogg.4 Uniformly Most Powerful Tests The Neyman-Pearson Lemma shows that when comparing point hypotheses H 0 and H, a test based on the likelihood ratio f X x H 0 /f X x H gives the most powerful test highest γ = P X H at a given significance α = P X H 0. We often wish to compare composite hypotheses, in which H and/or H 0 can correspond to parametrized families of distributions. In particular, if X is a sample drawn from a distribution fx; θ, we may have hypotheses H 0 : θ ω and H : θ Ω. If we first consider the case where H 0 is a point hypothesis θ = θ 0 and H is a composite hypothesis parametrized by θ, the power of a given test will depend on θ: γθ = P X ; θ = f X x; θd n x.8 If the same test is most powerful at a given α for every choice of θ, we say it is uniformly most powerful UMP. For example, recall the example from the last class, where X is a sample of size n from a Gamma5, θ distribution and H 0 is θ = θ 0 =. Now let H be the composite hypothesis θ >. We know from the Neyman-Pearson lemma that at each θ the most powerful test will be the one given by the likelihood ratio: : Λx; θ = Lθ 0; x Lθ; x where the constant kθ is defined by kθ.9 P ΛX; θ kθ H 0 = α.0.4. Example: One-Sided Hypothesis We saw last time that Lθ; x = n i= x4 i θ 5n e y/θ. where y is the sufficient statistic y = n i= x i, and Y is a Gamma5n, θ random variable. The likelihood ratio is Λθ; x = Lθ 5n [ 0; x θ Lθ; x = exp ] y θ 0 θ 0 θ = θ 5n exp θ. y θ As long as θ >, this is a decreasing function of y, so a test which rejects H 0 if Λ kθ will be equivalent to one which does so if Y a for some corresponding a. This can be defined so that the critical region is of size α: α = P Y a H 0 = γθ 0 = where Γs, x = a x y 5n Γ5nθ 5n 0 e y/θ 0 dy = Γ5n, a Γ5.3 t s e t dt.4 is the upper incomplete Gamma function. In any event, the value of a doesn t depend on θ, as long as θ >, so Y a is the UMP test of H 0 in light of the composite hypothesis H..4. Example: Two-Sided Hypothesis Note that there does not always exist a UMP test. Suppose instead the hypothesis H is θ θ 0 =. For θ >, the most 4

5 powerful test is to reject H 0 if Y a as defined above. For 0 < θ <, the likelihood ratio θ Λθ; x = θ 5n exp y.5 θ is a monotonically increasing function of y, so Λ kθ is equivalent to Y b, where b is defined by α = P Y b H 0 = γθ 0 = where b Γs Γs, x = y 5n Γ5nθ 5n 0 x e y/θ 0 dy = Γ5n Γ5n, b Γ5n.6 t s e t dt.7 is the lower incomplete Gamma function. Thus the most powerful test depends on the value of θ: rejecting H 0 when Y a is most powerful if θ > θ 0, while rejecting H when Y b is most powerful if θ < θ 0, To illustrate, we ll plot the power function for the two tests assuming α = 0.0 and n = 4. We ll get the percentiles and tail probabilities for the Gamma distribution using the SciPy stats package. For any distribution, the method sf is the survival function one minus the cdf, cdf is the cdf, and isf and ppf are the inverses of these functions. In []: from scipy.stats import gamma as gammadist In []: alpha = 0.0 In [3]: n = 4 In [4]: a = gammadist5*n.isfalpha In [5]: b = gammadist5*n.ppfalpha We check that the significance of the two tests is indeed 0%: In [6]: print gammadist5*n.sfa 0. In [7]: print gammadist5*n.cdfb 0. In [8]: theta = linspace0,3,000[:] The [:] is to omit the first element in the array, so that we re not actually trying to use θ = 0 as a value of the scale parameter. In [9]: gamma_a = gammadist5*n,scale=theta.sfa In [0]: gamma_b = gammadist5*n,scale=theta.cdfb In []: figure; In []: plottheta,gamma_a,'b-',label=r'$y\geq a$'; In [3]: plottheta,gamma_b,'r--',lw=,label=r'$y\leq b$'; In [4]: legendloc='center right'; In [5]: xlabelr'$\theta$'; In [6]: ylabelr'$\gamma\theta$'; In [7]: gridtrue In [8]: savefig'notes08_power.eps',bbox_inches='tight' Here are the two power curves: 5

6 γθ Y a Y b If we consider the test Y a, P X ; θ = a y 5n Γ5nθ 5n e y/θ dy = a/θ t 5n Γ5n e t dt.9 Since the integrand is positive definite and independent of θ, we see that we can maximize the integral by making θ as large as possible subject to H 0 so that the lower limit a/θ is as small as possible and the integral is maximized, i.e., θ = θ 0 =. So the sigificance of the test for null hypothesis H 0 : θ θ 0 = is the same as for H 0 : θ = θ 0 =, and the rest of the problem proceeds as before, i.e., Y a defines the UMP for H 0 : θ vs H : θ >..4.4 The Karlin-Rubin Theorem Note that they intersect at γ = 0., which makes sense since any test with significance α will satisfy γθ 0 = α. And as we can see, the power of the test Y a solid blue is higher for θ >, while the power of the test Y b dashed red is higher for θ <..4.3 Example: omposite Null Hypothesis Returning to the case where H is θ >, we can extend the discussion to comparison between two composite hypotheses by letting H 0 be θ. As a matter of convention, we define the significance α for a test with a composite null hypothesis as the worst-case false alarm rate: θ α = max P X ; θ.8 θ ω This is example is an illustration of a result known as the Karlin- Rubin theorem, which extends the Neyman-Pearson lemma to tests of one-sided composite hypotheses. If the likelihood function Lθ; x can be written in terms of a single sufficient statistic y = ux and has the property that, for any θ 0 and θ in the parameter space and satisfying θ 0 < θ, the likelihood ratio Λx; θ 0, θ = Lθ 0; θ Lθ ; θ.30 is a monotonically increasing or monotonically decreasing function of y, then there exists a uniformly most powerful test for H 0 : θ θ versus H : θ > θ. We call this situation a monotone likelihood ratio mlr..4.5 Example: Exponential Family The conditions for a monotone likelihood ratio seem a little restrictive, but there is a decent class of models which satisfy them. 6

7 If we consider a distribution which is a member of the regular exponential family, so fx; θ = exp ηθkx + Hx + qθ.3 then the likelihood for a sample of size n is Lθ; x = exp ηθ n Kx i + i= and the likelihood ratio is n Hx i + nqθ i=.3 Λx; θ 0, θ = Lθ 0; θ Lθ ; θ = exp [ηθ 0 ηθ ]y + n[qθ 0 qθ ].33 where the dependence on the data is via the sufficient statistic y = n i= Kx i. We see that: If ηθ is a monotonically increasing function of θ, then ηθ 0 ηθ < 0 for θ 0 < θ, and Λx; θ 0, θ is a monotonically decreasing function of y. If ηθ is a monotonically decreasing function of θ, then ηθ 0 ηθ > 0 for θ 0 < θ, and Λx; θ 0, θ is a monotonically increasing function of y. I.e., if ηθ is a monotone function of θ, we have a monotone likelihood ratio, and the Karlin-Rubin theorem tells us there s a UMP test to distinguish between one-sided hypotheses. Tuesday 3 May 06 Read Section 8.3 of Hogg Likelihood Ratio Tests and UMP Unbiased Tests. Gamma Example ontinued Recall that last time, we saw that, for a sample of size n drawn from a Gamma5, θ distribution, a test which rejected H 0 : θ = θ 0 = if n i= X i a was uniformly most powerful for comparing H 0 to the one-sided alternative hypothesis θ > θ 0 =, while one with rejected H 0 if n i= X i b was UMP if the alternative hypothesis was θ < θ 0 =, but if the alternative hypothesis was H = θ θ 0 =, there was no single test which was most powerful for every θ allowed under H. One method we ve considered before for comparing composite hypotheses is a likelihood ratio test, where we maximize the likelihood subject to each hypothesis. In this case, where the null hypothesis is still the point hypothesis θ = θ 0, this becomes specifically in this example, Λx = Lθ 0; x L θx, x Lθ; x =. n i= x4 i [Γ5] n θ 5n e y/θ. and the maximum likelihood solution θ is the solution to 0 = θ 5n ln θ y = 5n θ θ + y.3 θ 7

8 i.e., θ = y, which makes the likelihood ratio 5n Λx = L; x y 5n L y, x = e y+5n 5n 5n.4 Since this goes to zero as y goes to zero or infinity, a test which rejects H 0 if Λx k will be a two-sided test on Y, rejecting if Y c or Y c.5 where y = c and y = c correspond to equal values of the likelihood, i.e., c 5n e c = c 5n e c.6 which, together with the specified significance α determines c and c via α = c c f Y y; dy = Γ5n c c y 5n e y dy.7 Note that this likelihood ratio test is slightly different from the equal-tailed test defined by c sym 0 f Y y; dy = α = c sym f Y y; dy.8 which is slightly easier to construct. c sym and c sym can be written as the 00αth and 00 αth percentiles of the Gamma5n, distribution which applies to Y under the null hypothesis H 0. If the distribution function f Y y; and the likelihood ratio Λx were symmetric functions in y, of course, the likelihood ratio test would be equal-tailed. This is the case in several of the examples in Hogg. We can examine the power functions of the different options the two one-sided tests, the equal-tailed test, and the likelihood ratio test by continuing our plotting example from last time: In [9]: c_sym = gammadist5*n.isf0.5*alpha In [0]: c_sym = gammadist5*n.ppf0.5*alpha In []: gamma_sym =...: gammadist5*n,scale=theta.cdfc_sym...: + gammadist5*n,scale=theta.sfc_sym In []: plottheta,gamma_sym,'k:',lw=,...: label=r'$y\leq c_^{\rm sym}$ '...: + r'or $Y\geq c_^{\rm sym}$'; For the maximum likelihood test, we have to find the test which satisfies.6 from among the choices for which P Y c H 0 + P Y c H 0 = α: In [3]: alphaleft = linspace0,alpha,000[:-] In [4]: c = gammadist5*n.ppfalphaleft In [5]: c = gammadist5*n.isfalpha-alphaleft c and c are both 998-element arrays giving the lower and uppwe end of the range of y values for which we should not reject H 0. Rather than doing some sophisticated root-finding, we just brute-force it and pick the pair which has the lowest value of c 5n e c c 5n e c : In [6]: ind_ml = argminabsc**5*n*exp-c...: -c**5*n*exp-c In [7]: c_ml = c[ind_ml] In [8]: c_ml = c[ind_ml] 8

9 In [9]: gamma_ml =...: gammadist5*n,scale=theta.cdfc_ml...: + gammadist5*n,scale=theta.sfc_ml In [30]: plottheta,gamma_ml,'g-',...: label=r'$y\leq c_^{\rm ML}$ '...: + r'or $Y\geq c_^{\rm ML}$'; In [3]: legendloc='center right'; In [3]: savefig'notes08_ml.eps',bbox_inches='tight' γθ Y a Y b Y c sym or Y c sym Y c ML or Y c ML function that goes to unity as you go away from θ = θ 0 in either direction, as opposed to the one-sided tests, whose power function goes to zero on the wrong side of θ = θ 0.. Unbiased Tests One of the unsatisfactory properties of the one-sided hypothesis tests is that while they are the most powerful tests for some ranges of the parameter θ, for other ranges, the power is actually below the false alarm probability α. For instance, the test which rejects H 0 when Y a has γθ < α when θ < θ 0 =. This means the test is more likely to reject H 0 when it s true than when H is true, if it happens that θ <. A test with this undesirable property is called biased. onversely, an unbiased test is one for which γθ α for all θ allowed by H. It is often the case that, while no uniformly most powerful test exists, there is an unbiased test which, at all values of θ, is more powerful than any other unbiased test. This is known as the uniformly most powerful unbiased test. It can be shown that in many cases specifically the regular exponential family, the UMPU test is a two-sided test on the sufficient statistic Y, which rejects H 0 if Y c or Y c We see that the equal-tailed test is slightly different from the maximum likelihood test, thanks to the asymmetric distribution, but it s close. Both have the property of having a power θ.. Illustration for Gamma example Returning to the example of a sample from the Gamma5, θ distribution, consider a test which satisfies γθ α when θ θ 0 =. Since γθ 0 = α, that means the power function γθ must have a minimum at θ = θ 0. Recalling that 9

10 Y Gamma5n, θ, we have 0.0 γθ = c = Γ5n f Y y; θ dy = c Γ5n c /θ c /θ t 5n e t dt c c y 5n e y/θ dy θ 5n.9 where we have made the substitution t = y/θ in the last step. Requiring this to be a minimum at θ = θ 0 gives us 0 = γ θ 0 = c c 5n Γ5n θ0 which gives us, using θ 0 =, e c c θ 0 c 5n e c.0 γθ θ Y a Y b Y c sym or Y c sym Y c ML or Y c ML c 5n e c = c 5n e c. If we look back, we see this is just the condition.6 associated with the likelihood ratio test, so that test is in fact the UMP unbiased test for this model. We can verify this by zooming in on the power function plot above: In [33]: xlim[.95,.05]; In [34]: ylim[0.08,0.]; In [35]: savefig'notes08_mlzoom.eps',bbox_inches='tight' We see that the equal-tailed test has a power function which dips below α = 0.0 for a small range of θ values, so in fact that test is not unbiased. We see that the likelihood ratio test is indeed unbiased, and it turns out to tbe the UMPU test. Thursday 5 May 06 See Searle, 3 omposite Hypothesis Testing with Priors onsider now a slightly different situation. Suppose that the null hypothesis H 0 is a point hypothesis, but the alternative hypothesis H is a composite hypothesis which allows for a range 0

11 of values of the model parameters θ, but comes with a prior distribution f Θ θ H on those parameters. We include the possibility of a p-dimensional parameter space, but it may also be that p =. If we define a test which rejects H 0 when f X x H 0 f X x H = Lθ 0 ; x Lθ; x fθ θ H d p θ k 3. the Neyman-Pearson lemma tells us this is the most powerful test of H 0 versus H. This is most powerful in the sense of maximizing the power function γh = P X H = f X x H d n x = Lθ; x f Θ θ H d p θ d n x = γθ f Θ θ H d p θ A few points to note: 3. The test statistic in 3. is just the Bayes factor which we ve already motivated using Bayes s theorem in the form to write P H i x = f Xx H i P H i f X x P H 0 x P H x = f Xx H 0 P H 0 f X x H P H If there is a uniformly most powerful test which covers any θ in the support of f Θ θ H, this will also be the most powerful test of H 0 against H for any prior distribution. A possible objection is that the frequentist hypothesis testing formalism only applies to the outcomes of repeated experiments, and isn t supposed to know about any prior distribution. The preprint by Searle referenced above actually refers to the outcome of a Monte arlo experiment, where θ is also randomly generated along with the realization of X, so the relevant joint distribution is f XΘ x, θ H = f X x; θ f Θ θ H. In that context, the Bayes factor constructed using the same prior as the Monte arlo simulation gives the most powerful test, when attention is restricted to tests which define using only X and not Θ. 3. Example: Mean of a Normal Distribution Suppose that the data vector is X = X X but rather than being a sample from a given distribution, they are independent random variables X Nθ, and X Nθ,. Let the null hypothesis be H 0 : θ = 0 = θ and the alternative hypothesis be H : θ θ 0 0 Lθ; x = π exp. The likelihood function is x θ x θ 3.5 The maximum likelihood solution is θ = x and θ = x, so the likelihood ratio test would reject H 0 if L0, x Λx = L θ, x = exp x + x k 3.6 which means, for a test of significance α, rejecting H 0 if X + X χ,α 3.7 On the other hand, if we have a prior f Θ θ, θ H the Neyman-Pearson lemma tells us the optimal test statistic is the Bayes factor exp x +x f Θθ, θ H exp x θ x θ 3.8 dθ dθ

12 3.. ase : Uniform Prior Suppose that the prior f Θ θ, θ H is uniform in θ and θ. In that case the denominator is proportional to exp x θ x θ dθ dθ = constant 3.9 and the Bayes factor is proportional to the maximum-likelihood ratio exp x + x 3.0 so using x + x as a test statistic does give the optimal test in light of the prior. 3.. ase : Non-Uniform Prior It could be that the prior on θ and θ is more complicated. For instance, suppose f Θ θ, θ θ θ. Then the denominator of the Bayes factor is proportional to θθ exp x θ x θ dθ dθ + x + x 3. This is an improper prior, so technically the normalization factor will be zero and Bayes factor will be infinite, but we re only interested in the x dependence of the Bayes factor anyway. If we want to be careful, we can use a prior that is e.g., Gaussian with some large width, much greater than any of the other values in the problem. E.g., for the gravitational wave signal from a rotating or orbiting system, the amplitudes of the left- and right-circularly polarized parts of the signal are proportional to ±cos ι where ι is an inclination angle, which is generally unknown with a uniform prior on cos ι [, ], which produces a non-trivial prior distribution for the two amplitudes. See Whelan et al, lassical and Quantum Gravity 3, ; which means the Bayes factor is proportional to exp x +x + x + x and the optimal test rejects H 0 if 3. X + X + ln + X + ln + X c 3.3 We can show the difference between the two families of critical regions on a contour plot: In []: x = linspace-4,4,00 In []: xgrid, xgrid = meshgridx,x In [3]: Fgrid = xgrid** + xgrid** In [4]: Bgrid = Fgrid...: + *log+xgrid** + *log+xgrid** In [5]: xlevels = arange6 In [6]: Blevels = xlevels** + *log+xlevels** In [7]: Flevels = xlevels** In [8]: figurefigsize=5,5; In [9]: contourxgrid,xgrid,fgrid,flevels,colors='r',...: linewidths=,linestyles='dashed'; In [0]: contourxgrid,xgrid,bgrid,blevels,colors='b'; In []: gridtrue

13 In []: xlabelr'$x_$'; In [3]: ylabelr'$x_$'; In [4]: savefig'notes08_contours.eps',...: bbox_inches='tight' 4 3 to estimate the significance of the Bayes-factor test numerically. See figure 3 of Whelan et al 04 for an example of this. Tuesday 0 May 06 Review for Final Exam The exam is comprehensive, but with relatively more emphasis on chapter eight and the second half of chapter seven. Please come with questions and topics you d like to go over. Thursday May 06 Review for Final Exam The exam is comprehensive, but with relatively more emphasis on chapter eight and the second half of chapter seven. Please come with questions and topics you d like to go over. x x Note that we drew the contours at arbitrary levels; to draw them at the same significance for the two tests, we d probably need 3

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Lecture 21. Hypothesis Testing II

Lecture 21. Hypothesis Testing II Lecture 21. Hypothesis Testing II December 7, 2011 In the previous lecture, we dened a few key concepts of hypothesis testing and introduced the framework for parametric hypothesis testing. In the parametric

More information

Hypothesis Testing - Frequentist

Hypothesis Testing - Frequentist Frequentist Hypothesis Testing - Frequentist Compare two hypotheses to see which one better explains the data. Or, alternatively, what is the best way to separate events into two classes, those originating

More information

Theory of Statistical Tests

Theory of Statistical Tests Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H

More information

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Hypothesis Testing Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution. Suppose the family of population distributions is indexed

More information

10. Composite Hypothesis Testing. ECE 830, Spring 2014

10. Composite Hypothesis Testing. ECE 830, Spring 2014 10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES 557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES Example Suppose that X,..., X n N, ). To test H 0 : 0 H : the most powerful test at level α is based on the statistic λx) f π) X x ) n/ exp

More information

Hypothesis Testing: The Generalized Likelihood Ratio Test

Hypothesis Testing: The Generalized Likelihood Ratio Test Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random

More information

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider

More information

Interval Estimation. Chapter 9

Interval Estimation. Chapter 9 Chapter 9 Interval Estimation 9.1 Introduction Definition 9.1.1 An interval estimate of a real-values parameter θ is any pair of functions, L(x 1,..., x n ) and U(x 1,..., x n ), of a sample that satisfy

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses. Stat 300A Theory of Statistics Homework 7: Solutions Nikos Ignatiadis Due on November 28, 208 Solutions should be complete and concisely written. Please, use a separate sheet or set of sheets for each

More information

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then

More information

Methods of evaluating tests

Methods of evaluating tests Methods of evaluating tests Let X,, 1 Xn be i.i.d. Bernoulli( p ). Then 5 j= 1 j ( 5, ) T = X Binomial p. We test 1 H : p vs. 1 1 H : p>. We saw that a LRT is 1 if t k* φ ( x ) =. otherwise (t is the observed

More information

λ(x + 1)f g (x) > θ 0

λ(x + 1)f g (x) > θ 0 Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where

More information

Chapter 4. Theory of Tests. 4.1 Introduction

Chapter 4. Theory of Tests. 4.1 Introduction Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

MAT 271E Probability and Statistics

MAT 271E Probability and Statistics MAT 71E Probability and Statistics Spring 013 Instructor : Class Meets : Office Hours : Textbook : Supp. Text : İlker Bayram EEB 1103 ibayram@itu.edu.tr 13.30 1.30, Wednesday EEB 5303 10.00 1.00, Wednesday

More information

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Valen E. Johnson Texas A&M University February 27, 2014 Valen E. Johnson Texas A&M University Uniformly most powerful Bayes

More information

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering

More information

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4 STA 73: Inference Notes. Neyman-Pearsonian Classical Hypothesis Testing B&D 4 1 Testing as a rule Fisher s quantification of extremeness of observed evidence clearly lacked rigorous mathematical interpretation.

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

STAT 514 Solutions to Assignment #6

STAT 514 Solutions to Assignment #6 STAT 514 Solutions to Assignment #6 Question 1: Suppose that X 1,..., X n are a simple random sample from a Weibull distribution with density function f θ x) = θcx c 1 exp{ θx c }I{x > 0} for some fixed

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

HYPOTHESIS TESTING: FREQUENTIST APPROACH. HYPOTHESIS TESTING: FREQUENTIST APPROACH. These notes summarize the lectures on (the frequentist approach to) hypothesis testing. You should be familiar with the standard hypothesis testing from previous

More information

Statistics Masters Comprehensive Exam March 21, 2003

Statistics Masters Comprehensive Exam March 21, 2003 Statistics Masters Comprehensive Exam March 21, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

Primer on statistics:

Primer on statistics: Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood

More information

E. Santovetti lesson 4 Maximum likelihood Interval estimation

E. Santovetti lesson 4 Maximum likelihood Interval estimation E. Santovetti lesson 4 Maximum likelihood Interval estimation 1 Extended Maximum Likelihood Sometimes the number of total events measurements of the experiment n is not fixed, but, for example, is a Poisson

More information

Lecture 2: Basic Concepts of Statistical Decision Theory

Lecture 2: Basic Concepts of Statistical Decision Theory EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture

More information

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes

More information

Lecture 12 November 3

Lecture 12 November 3 STATS 300A: Theory of Statistics Fall 2015 Lecture 12 November 3 Lecturer: Lester Mackey Scribe: Jae Hyuck Park, Christian Fong Warning: These notes may contain factual and/or typographic errors. 12.1

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple

More information

Statistical Inference

Statistical Inference Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Spring, 2006 1. DeGroot 1973 In (DeGroot 1973), Morrie DeGroot considers testing the

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics MAS 713 Chapter 8 Previous lecture: 1 Bayesian Inference 2 Decision theory 3 Bayesian Vs. Frequentist 4 Loss functions 5 Conjugate priors Any questions? Mathematical Statistics

More information

Lecture 16 November Application of MoUM to our 2-sided testing problem

Lecture 16 November Application of MoUM to our 2-sided testing problem STATS 300A: Theory of Statistics Fall 2015 Lecture 16 November 17 Lecturer: Lester Mackey Scribe: Reginald Long, Colin Wei Warning: These notes may contain factual and/or typographic errors. 16.1 Recap

More information

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson

More information

Composite Hypotheses and Generalized Likelihood Ratio Tests

Composite Hypotheses and Generalized Likelihood Ratio Tests Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve

More information

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2007 Problems 4: Solution sketches Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the

More information

Topic 15: Simple Hypotheses

Topic 15: Simple Hypotheses Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is

More information

Answers to the 8th problem set. f(x θ = θ 0 ) L(θ 0 )

Answers to the 8th problem set. f(x θ = θ 0 ) L(θ 0 ) Answers to the 8th problem set The likelihood ratio with which we worked in this problem set is: Λ(x) = f(x θ = θ 1 ) L(θ 1 ) =. f(x θ = θ 0 ) L(θ 0 ) With a lower-case x, this defines a function. With

More information

Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, Discreteness versus Hypothesis Tests

Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, Discreteness versus Hypothesis Tests Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, 2016 1 Discreteness versus Hypothesis Tests You cannot do an exact level α test for any α when the data are discrete.

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004

Hypothesis Testing. BS2 Statistical Inference, Lecture 11 Michaelmas Term Steffen Lauritzen, University of Oxford; November 15, 2004 Hypothesis Testing BS2 Statistical Inference, Lecture 11 Michaelmas Term 2004 Steffen Lauritzen, University of Oxford; November 15, 2004 Hypothesis testing We consider a family of densities F = {f(x; θ),

More information

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS 1a. Under the null hypothesis X has the binomial (100,.5) distribution with E(X) = 50 and SE(X) = 5. So P ( X 50 > 10) is (approximately) two tails

More information

LECTURE NOTES 57. Lecture 9

LECTURE NOTES 57. Lecture 9 LECTURE NOTES 57 Lecture 9 17. Hypothesis testing A special type of decision problem is hypothesis testing. We partition the parameter space into H [ A with H \ A = ;. Wewrite H 2 H A 2 A. A decision problem

More information

4.5.1 The use of 2 log Λ when θ is scalar

4.5.1 The use of 2 log Λ when θ is scalar 4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters D. Richard Brown III Worcester Polytechnic Institute 26-February-2009 Worcester Polytechnic Institute D. Richard Brown III 26-February-2009

More information

6.867 Machine Learning

6.867 Machine Learning 6.867 Machine Learning Problem set 1 Due Thursday, September 19, in class What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

MAT 271E Probability and Statistics

MAT 271E Probability and Statistics MAT 7E Probability and Statistics Spring 6 Instructor : Class Meets : Office Hours : Textbook : İlker Bayram EEB 3 ibayram@itu.edu.tr 3.3 6.3, Wednesday EEB 6.., Monday D. B. Bertsekas, J. N. Tsitsiklis,

More information

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood

More information

Lecture 26: Likelihood ratio tests

Lecture 26: Likelihood ratio tests Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for

More information

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49 4 HYPOTHESIS TESTING 49 4 Hypothesis testing In sections 2 and 3 we considered the problem of estimating a single parameter of interest, θ. In this section we consider the related problem of testing whether

More information

Bayesian Statistics (Hogg Chapter Eleven)

Bayesian Statistics (Hogg Chapter Eleven) Bayesian Statistics (Hogg Chapter Eleven) STAT 406-0: Mathematical Statistics II Spring Semester 206 Contents 0 Preliminaries 2 0. Administrata.................... 2 0.2 Outline........................

More information

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12 Agenda Example:

More information

Line Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions.

Line Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions. Line Broadening Spectral lines are not arbitrarily sharp. There are a variety of mechanisms that give them finite width, and some of those mechanisms contain significant information. We ll consider a few

More information

5.1 Uniformly Most Accurate Families of Confidence

5.1 Uniformly Most Accurate Families of Confidence Chapter 5 Confidence Estimation Let (X, B X, P ), P {P θ θ Θ} be sample space and {Θ, B Θ, µ} be measure space with some σ finite measure µ. 5.1 Uniformly Most Accurate Families of Confidence Sets Def.

More information

More Empirical Process Theory

More Empirical Process Theory More Empirical Process heory 4.384 ime Series Analysis, Fall 2008 Recitation by Paul Schrimpf Supplementary to lectures given by Anna Mikusheva October 24, 2008 Recitation 8 More Empirical Process heory

More information

Hypothesis testing: theory and methods

Hypothesis testing: theory and methods Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable

More information

14.30 Introduction to Statistical Methods in Economics Spring 2009

14.30 Introduction to Statistical Methods in Economics Spring 2009 MIT OpenCourseWare http://ocw.mit.edu.30 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. .30

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Partitioning the Parameter Space. Topic 18 Composite Hypotheses Topic 18 Composite Hypotheses Partitioning the Parameter Space 1 / 10 Outline Partitioning the Parameter Space 2 / 10 Partitioning the Parameter Space Simple hypotheses limit us to a decision between one

More information

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013 UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013 STAB57H3 Introduction to Statistics Duration: 3 hours Last Name: First Name: Student number:

More information

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 8 Basics Hypotheses H 0

More information

Stat 710: Mathematical Statistics Lecture 27

Stat 710: Mathematical Statistics Lecture 27 Stat 710: Mathematical Statistics Lecture 27 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 27 April 3, 2009 1 / 10 Lecture 27:

More information

Statistical Theory MT 2006 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine

More information

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)? ECE 830 / CS 76 Spring 06 Instructors: R. Willett & R. Nowak Lecture 3: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we

More information

6.867 Machine Learning

6.867 Machine Learning 6.867 Machine Learning Problem set 1 Solutions Thursday, September 19 What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.

More information

Distribution-Free Procedures (Devore Chapter Fifteen)

Distribution-Free Procedures (Devore Chapter Fifteen) Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?

More information

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D. probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

ST495: Survival Analysis: Hypothesis testing and confidence intervals

ST495: Survival Analysis: Hypothesis testing and confidence intervals ST495: Survival Analysis: Hypothesis testing and confidence intervals Eric B. Laber Department of Statistics, North Carolina State University April 3, 2014 I remember that one fateful day when Coach took

More information

TUTORIAL 8 SOLUTIONS #

TUTORIAL 8 SOLUTIONS # TUTORIAL 8 SOLUTIONS #9.11.21 Suppose that a single observation X is taken from a uniform density on [0,θ], and consider testing H 0 : θ = 1 versus H 1 : θ =2. (a) Find a test that has significance level

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

3 Continuous Random Variables

3 Continuous Random Variables Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

http://www.math.uah.edu/stat/hypothesis/.xhtml 1 of 5 7/29/2009 3:14 PM Virtual Laboratories > 9. Hy pothesis Testing > 1 2 3 4 5 6 7 1. The Basic Statistical Model As usual, our starting point is a random

More information

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30 Problem Set MAS 6J/1.16J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain

More information

Charles Geyer University of Minnesota. joint work with. Glen Meeden University of Minnesota.

Charles Geyer University of Minnesota. joint work with. Glen Meeden University of Minnesota. Fuzzy Confidence Intervals and P -values Charles Geyer University of Minnesota joint work with Glen Meeden University of Minnesota http://www.stat.umn.edu/geyer/fuzz 1 Ordinary Confidence Intervals OK

More information

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques

Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques Vol:7, No:0, 203 Derivation of Monotone Likelihood Ratio Using Two Sided Uniformly Normal Distribution Techniques D. A. Farinde International Science Index, Mathematical and Computational Sciences Vol:7,

More information

8 Basics of Hypothesis Testing

8 Basics of Hypothesis Testing 8 Basics of Hypothesis Testing 4 Problems Problem : The stochastic signal S is either 0 or E with equal probability, for a known value E > 0. Consider an observation X = x of the stochastic variable X

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Detection Theory. Composite tests

Detection Theory. Composite tests Composite tests Chapter 5: Correction Thu I claimed that the above, which is the most general case, was captured by the below Thu Chapter 5: Correction Thu I claimed that the above, which is the most general

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information